/linux-4.1.27/arch/m68k/include/asm/ |
D | bitops.h | 31 static inline void bset_reg_set_bit(int nr, volatile unsigned long *vaddr) in bset_reg_set_bit() argument 33 char *p = (char *)vaddr + (nr ^ 31) / 8; in bset_reg_set_bit() 41 static inline void bset_mem_set_bit(int nr, volatile unsigned long *vaddr) in bset_mem_set_bit() argument 43 char *p = (char *)vaddr + (nr ^ 31) / 8; in bset_mem_set_bit() 50 static inline void bfset_mem_set_bit(int nr, volatile unsigned long *vaddr) in bfset_mem_set_bit() argument 54 : "d" (nr ^ 31), "o" (*vaddr) in bfset_mem_set_bit() 59 #define set_bit(nr, vaddr) bset_reg_set_bit(nr, vaddr) argument 61 #define set_bit(nr, vaddr) bset_mem_set_bit(nr, vaddr) argument 63 #define set_bit(nr, vaddr) (__builtin_constant_p(nr) ? \ argument 64 bset_mem_set_bit(nr, vaddr) : \ [all …]
|
D | page_mm.h | 9 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument 58 #define clear_user_page(addr, vaddr, page) \ argument 62 #define copy_user_page(to, from, vaddr, page) \ argument 73 static inline unsigned long ___pa(void *vaddr) in ___pa() argument 80 : "0" (vaddr), "i" (m68k_fixup_memoffset)); in ___pa() 83 #define __pa(vaddr) ___pa((void *)(long)(vaddr)) argument 86 void *vaddr; in __va() local 90 : "=r" (vaddr) in __va() 92 return vaddr; in __va()
|
D | page_no.h | 9 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument 15 #define clear_user_page(page, vaddr, pg) clear_page(page) argument 16 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument 18 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument 19 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr) 22 #define __pa(vaddr) ((unsigned long)(vaddr)) argument
|
D | cacheflush_mm.h | 155 extern void cache_push_v(unsigned long vaddr, int len); 222 static inline void __flush_page_to_ram(void *vaddr) in __flush_page_to_ram() argument 226 addr = ((unsigned long) vaddr) & ~(PAGE_SIZE - 1); in __flush_page_to_ram() 239 : : "a" (__pa(vaddr))); in __flush_page_to_ram() 261 struct page *page, unsigned long vaddr, in copy_to_user_page() argument 264 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page() 266 flush_icache_user_range(vma, page, vaddr, len); in copy_to_user_page() 269 struct page *page, unsigned long vaddr, in copy_from_user_page() argument 272 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_from_user_page()
|
/linux-4.1.27/arch/arm/mm/ |
D | cache-xsc3l2.c | 100 unsigned long vaddr; in xsc3_l2_inv_range() local 107 vaddr = -1; /* to force the first mapping */ in xsc3_l2_inv_range() 113 vaddr = l2_map_va(start & ~(CACHE_LINE_SIZE - 1), vaddr); in xsc3_l2_inv_range() 114 xsc3_l2_clean_mva(vaddr); in xsc3_l2_inv_range() 115 xsc3_l2_inv_mva(vaddr); in xsc3_l2_inv_range() 123 vaddr = l2_map_va(start, vaddr); in xsc3_l2_inv_range() 124 xsc3_l2_inv_mva(vaddr); in xsc3_l2_inv_range() 132 vaddr = l2_map_va(start, vaddr); in xsc3_l2_inv_range() 133 xsc3_l2_clean_mva(vaddr); in xsc3_l2_inv_range() 134 xsc3_l2_inv_mva(vaddr); in xsc3_l2_inv_range() [all …]
|
D | highmem.c | 23 unsigned long vaddr = __fix_to_virt(idx); in set_fixmap_pte() local 24 pte_t *ptep = pte_offset_kernel(pmd_off_k(vaddr), vaddr); in set_fixmap_pte() 27 local_flush_tlb_kernel_page(vaddr); in set_fixmap_pte() 30 static inline pte_t get_fixmap_pte(unsigned long vaddr) in get_fixmap_pte() argument 32 pte_t *ptep = pte_offset_kernel(pmd_off_k(vaddr), vaddr); in get_fixmap_pte() 58 unsigned long vaddr; in kmap_atomic() local 82 vaddr = __fix_to_virt(idx); in kmap_atomic() 88 BUG_ON(!pte_none(get_fixmap_pte(vaddr))); in kmap_atomic() 97 return (void *)vaddr; in kmap_atomic() 103 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local [all …]
|
D | copypage-v6.c | 34 struct page *from, unsigned long vaddr, struct vm_area_struct *vma) in v6_copy_user_highpage_nonaliasing() argument 49 static void v6_clear_user_highpage_nonaliasing(struct page *page, unsigned long vaddr) in v6_clear_user_highpage_nonaliasing() argument 73 struct page *from, unsigned long vaddr, struct vm_area_struct *vma) in v6_copy_user_highpage_aliasing() argument 75 unsigned int offset = CACHE_COLOUR(vaddr); in v6_copy_user_highpage_aliasing() 106 static void v6_clear_user_highpage_aliasing(struct page *page, unsigned long vaddr) in v6_clear_user_highpage_aliasing() argument 108 unsigned long to = COPYPAGE_V6_TO + (CACHE_COLOUR(vaddr) << PAGE_SHIFT); in v6_clear_user_highpage_aliasing()
|
D | copypage-feroceon.c | 71 unsigned long vaddr, struct vm_area_struct *vma) in feroceon_copy_user_highpage() argument 77 flush_cache_page(vma, vaddr, page_to_pfn(from)); in feroceon_copy_user_highpage() 83 void feroceon_clear_user_highpage(struct page *page, unsigned long vaddr) in feroceon_clear_user_highpage() argument
|
D | copypage-xsc3.c | 74 unsigned long vaddr, struct vm_area_struct *vma) in xsc3_mc_copy_user_highpage() argument 80 flush_cache_page(vma, vaddr, page_to_pfn(from)); in xsc3_mc_copy_user_highpage() 91 void xsc3_mc_clear_user_highpage(struct page *page, unsigned long vaddr) in xsc3_mc_clear_user_highpage() argument
|
D | copypage-v4wb.c | 51 unsigned long vaddr, struct vm_area_struct *vma) in v4wb_copy_user_highpage() argument 57 flush_cache_page(vma, vaddr, page_to_pfn(from)); in v4wb_copy_user_highpage() 68 void v4wb_clear_user_highpage(struct page *page, unsigned long vaddr) in v4wb_clear_user_highpage() argument
|
D | flush.c | 26 static void flush_pfn_alias(unsigned long pfn, unsigned long vaddr) in flush_pfn_alias() argument 28 unsigned long to = FLUSH_ALIAS_START + (CACHE_COLOUR(vaddr) << PAGE_SHIFT); in flush_pfn_alias() 40 static void flush_icache_alias(unsigned long pfn, unsigned long vaddr, unsigned long len) in flush_icache_alias() argument 42 unsigned long va = FLUSH_ALIAS_START + (CACHE_COLOUR(vaddr) << PAGE_SHIFT); in flush_icache_alias() 43 unsigned long offset = vaddr & (PAGE_SIZE - 1); in flush_icache_alias() 103 #define flush_pfn_alias(pfn,vaddr) do { } while (0) argument 104 #define flush_icache_alias(pfn,vaddr,len) do { } while (0) argument
|
/linux-4.1.27/arch/parisc/kernel/ |
D | pci-dma.c | 83 unsigned long vaddr, in map_pte_uncached() argument 87 unsigned long orig_vaddr = vaddr; in map_pte_uncached() 89 vaddr &= ~PMD_MASK; in map_pte_uncached() 90 end = vaddr + size; in map_pte_uncached() 102 vaddr += PAGE_SIZE; in map_pte_uncached() 106 } while (vaddr < end); in map_pte_uncached() 110 static inline int map_pmd_uncached(pmd_t * pmd, unsigned long vaddr, in map_pmd_uncached() argument 114 unsigned long orig_vaddr = vaddr; in map_pmd_uncached() 116 vaddr &= ~PGDIR_MASK; in map_pmd_uncached() 117 end = vaddr + size; in map_pmd_uncached() [all …]
|
/linux-4.1.27/arch/sh/mm/ |
D | kmap.c | 18 #define kmap_get_fixmap_pte(vaddr) \ argument 19 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), (vaddr)) 25 unsigned long vaddr; in kmap_coherent_init() local 28 vaddr = __fix_to_virt(FIX_CMAP_BEGIN); in kmap_coherent_init() 29 kmap_coherent_pte = kmap_get_fixmap_pte(vaddr); in kmap_coherent_init() 35 unsigned long vaddr; in kmap_coherent() local 45 vaddr = __fix_to_virt(idx); in kmap_coherent() 50 return (void *)vaddr; in kmap_coherent() 56 unsigned long vaddr = (unsigned long)kvaddr & PAGE_MASK; in kunmap_coherent() local 57 enum fixed_addresses idx = __virt_to_fix(vaddr); in kunmap_coherent() [all …]
|
D | init.c | 153 unsigned long vaddr, pte_t *lastpte) in page_table_kmap_check() argument 166 unsigned long vaddr; in page_table_range_init() local 168 vaddr = start; in page_table_range_init() 169 i = __pgd_offset(vaddr); in page_table_range_init() 170 j = __pud_offset(vaddr); in page_table_range_init() 171 k = __pmd_offset(vaddr); in page_table_range_init() 174 for ( ; (i < PTRS_PER_PGD) && (vaddr != end); pgd++, i++) { in page_table_range_init() 176 for ( ; (j < PTRS_PER_PUD) && (vaddr != end); pud++, j++) { in page_table_range_init() 181 for (; (k < PTRS_PER_PMD) && (vaddr != end); pmd++, k++) { in page_table_range_init() 183 pmd, vaddr, pte); in page_table_range_init() [all …]
|
D | cache.c | 59 unsigned long vaddr, void *dst, const void *src, in copy_to_user_page() argument 64 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page() 74 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page() 78 unsigned long vaddr, void *dst, const void *src, in copy_from_user_page() argument 83 void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_from_user_page() 94 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument 102 vfrom = kmap_coherent(from, vaddr); in copy_user_highpage() 111 if (pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK) || in copy_user_highpage() 121 void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument 127 if (pages_do_alias((unsigned long)kaddr, vaddr & PAGE_MASK)) in clear_user_highpage()
|
D | pmb.c | 137 static bool pmb_mapping_exists(unsigned long vaddr, phys_addr_t phys, in pmb_mapping_exists() argument 156 if ((vaddr < pmbe->vpn) || (vaddr >= (pmbe->vpn + pmbe->size))) in pmb_mapping_exists() 335 int pmb_bolt_mapping(unsigned long vaddr, phys_addr_t phys, in pmb_bolt_mapping() argument 345 if (!pmb_addr_valid(vaddr, size)) in pmb_bolt_mapping() 347 if (pmb_mapping_exists(vaddr, phys, size)) in pmb_bolt_mapping() 350 orig_addr = vaddr; in pmb_bolt_mapping() 353 flush_tlb_kernel_range(vaddr, vaddr + size); in pmb_bolt_mapping() 363 pmbe = pmb_alloc(vaddr, phys, pmb_flags | in pmb_bolt_mapping() 377 vaddr += pmbe->size; in pmb_bolt_mapping() 413 unsigned long vaddr; in pmb_remap_caller() local [all …]
|
D | cache-sh4.c | 215 void *vaddr; in sh4_flush_cache_page() local 236 vaddr = NULL; in sh4_flush_cache_page() 246 vaddr = kmap_coherent(page, address); in sh4_flush_cache_page() 248 vaddr = kmap_atomic(page); in sh4_flush_cache_page() 250 address = (unsigned long)vaddr; in sh4_flush_cache_page() 259 if (vaddr) { in sh4_flush_cache_page() 261 kunmap_coherent(vaddr); in sh4_flush_cache_page() 263 kunmap_atomic(vaddr); in sh4_flush_cache_page()
|
D | consistent.c | 68 void *vaddr, dma_addr_t dma_handle, in dma_generic_free_coherent() argument 78 iounmap(vaddr); in dma_generic_free_coherent() 81 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument 87 (void *)CAC_ADDR((unsigned long)vaddr) : vaddr; in dma_cache_sync()
|
D | ioremap.c | 108 unsigned long vaddr = (unsigned long __force)addr; in __iounmap() local 114 if (iomapping_nontranslatable(vaddr)) in __iounmap() 129 p = remove_vm_area((void *)(vaddr & PAGE_MASK)); in __iounmap()
|
/linux-4.1.27/arch/microblaze/mm/ |
D | consistent.c | 64 unsigned long order, vaddr; in consistent_alloc() local 82 vaddr = __get_free_pages(gfp, order); in consistent_alloc() 83 if (!vaddr) in consistent_alloc() 90 flush_dcache_range(virt_to_phys((void *)vaddr), in consistent_alloc() 91 virt_to_phys((void *)vaddr) + size); in consistent_alloc() 94 ret = (void *)vaddr; in consistent_alloc() 113 free_pages(vaddr, order); in consistent_alloc() 120 *dma_handle = pa = __virt_to_phys(vaddr); in consistent_alloc() 129 page = virt_to_page(vaddr); in consistent_alloc() 151 free_pages(vaddr, order); in consistent_alloc() [all …]
|
D | highmem.c | 37 unsigned long vaddr; in kmap_atomic_prot() local 48 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot() 52 set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot)); in kmap_atomic_prot() 53 local_flush_tlb_page(NULL, vaddr); in kmap_atomic_prot() 55 return (void *) vaddr; in kmap_atomic_prot() 61 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 64 if (vaddr < __fix_to_virt(FIX_KMAP_END)) { in __kunmap_atomic() 75 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in __kunmap_atomic() 81 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic() 82 local_flush_tlb_page(NULL, vaddr); in __kunmap_atomic()
|
/linux-4.1.27/arch/mips/mm/ |
D | highmem.c | 47 unsigned long vaddr; in kmap_atomic() local 57 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic() 62 local_flush_tlb_one((unsigned long)vaddr); in kmap_atomic() 64 return (void*) vaddr; in kmap_atomic() 70 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 73 if (vaddr < FIXADDR_START) { // FIXME in __kunmap_atomic() 83 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in __kunmap_atomic() 89 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic() 90 local_flush_tlb_one(vaddr); in __kunmap_atomic() 104 unsigned long vaddr; in kmap_atomic_pfn() local [all …]
|
D | pgtable-32.c | 35 unsigned long vaddr; in pagetable_init() local 54 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; in pagetable_init() 55 fixrange_init(vaddr, vaddr + FIXADDR_SIZE, pgd_base); in pagetable_init() 61 vaddr = PKMAP_BASE; in pagetable_init() 62 fixrange_init(vaddr, vaddr + PAGE_SIZE*LAST_PKMAP, pgd_base); in pagetable_init() 64 pgd = swapper_pg_dir + __pgd_offset(vaddr); in pagetable_init() 65 pud = pud_offset(pgd, vaddr); in pagetable_init() 66 pmd = pmd_offset(pud, vaddr); in pagetable_init() 67 pte = pte_offset_kernel(pmd, vaddr); in pagetable_init()
|
D | init.c | 86 unsigned long vaddr, flags, entrylo; in __kmap_pgprot() local 96 vaddr = __fix_to_virt(FIX_CMAP_END - idx); in __kmap_pgprot() 106 write_c0_entryhi(vaddr & (PAGE_MASK << 1)); in __kmap_pgprot() 123 return (void*) vaddr; in __kmap_pgprot() 158 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument 165 vfrom = kmap_coherent(from, vaddr); in copy_user_highpage() 174 pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK)) in copy_user_highpage() 182 struct page *page, unsigned long vaddr, void *dst, const void *src, in copy_to_user_page() argument 187 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page() 196 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page() [all …]
|
D | pgtable-64.c | 97 unsigned long vaddr; in pagetable_init() local 109 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; in pagetable_init() 110 fixrange_init(vaddr, vaddr + FIXADDR_SIZE, pgd_base); in pagetable_init()
|
/linux-4.1.27/drivers/media/v4l2-core/ |
D | videobuf2-vmalloc.c | 25 void *vaddr; member 48 buf->vaddr = vmalloc_user(buf->size); in vb2_vmalloc_alloc() 54 if (!buf->vaddr) { in vb2_vmalloc_alloc() 69 vfree(buf->vaddr); in vb2_vmalloc_put() 74 static void *vb2_vmalloc_get_userptr(void *alloc_ctx, unsigned long vaddr, in vb2_vmalloc_get_userptr() argument 89 offset = vaddr & ~PAGE_MASK; in vb2_vmalloc_get_userptr() 93 vma = find_vma(current->mm, vaddr); in vb2_vmalloc_get_userptr() 95 if (vb2_get_contig_userptr(vaddr, size, &vma, &physp)) in vb2_vmalloc_get_userptr() 98 buf->vaddr = (__force void *)ioremap_nocache(physp, size); in vb2_vmalloc_get_userptr() 99 if (!buf->vaddr) in vb2_vmalloc_get_userptr() [all …]
|
D | videobuf-vmalloc.c | 105 __func__, i, mem->vaddr); in videobuf_vm_close() 107 vfree(mem->vaddr); in videobuf_vm_close() 108 mem->vaddr = NULL; in videobuf_vm_close() 173 if (!mem->vaddr) { in __videobuf_iolock() 192 mem->vaddr = vmalloc_user(pages); in __videobuf_iolock() 193 if (!mem->vaddr) { in __videobuf_iolock() 198 mem->vaddr, pages); in __videobuf_iolock() 257 mem->vaddr = vmalloc_user(pages); in __videobuf_mmap_mapper() 258 if (!mem->vaddr) { in __videobuf_mmap_mapper() 262 dprintk(1, "vmalloc is at addr %p (%d pages)\n", mem->vaddr, pages); in __videobuf_mmap_mapper() [all …]
|
D | videobuf2-dma-sg.c | 39 void *vaddr; member 120 buf->vaddr = NULL; in vb2_dma_sg_alloc() 192 if (buf->vaddr) in vb2_dma_sg_put() 193 vm_unmap_ram(buf->vaddr, buf->num_pages); in vb2_dma_sg_put() 232 static void *vb2_dma_sg_get_userptr(void *alloc_ctx, unsigned long vaddr, in vb2_dma_sg_get_userptr() argument 250 buf->vaddr = NULL; in vb2_dma_sg_get_userptr() 253 buf->offset = vaddr & ~PAGE_MASK; in vb2_dma_sg_get_userptr() 257 first = (vaddr & PAGE_MASK) >> PAGE_SHIFT; in vb2_dma_sg_get_userptr() 258 last = ((vaddr + size - 1) & PAGE_MASK) >> PAGE_SHIFT; in vb2_dma_sg_get_userptr() 266 vma = find_vma(current->mm, vaddr); in vb2_dma_sg_get_userptr() [all …]
|
D | videobuf-dma-contig.c | 28 void *vaddr; member 45 mem->vaddr = dma_alloc_coherent(dev, mem->size, in __videobuf_dc_alloc() 48 if (!mem->vaddr) { in __videobuf_dc_alloc() 53 dev_dbg(dev, "dma mapped data is at %p (%ld)\n", mem->vaddr, mem->size); in __videobuf_dc_alloc() 61 dma_free_coherent(dev, mem->size, mem->vaddr, mem->dma_handle); in __videobuf_dc_free() 63 mem->vaddr = NULL; in __videobuf_dc_free() 117 i, mem->vaddr); in videobuf_vm_close() 120 mem->vaddr = NULL; in videobuf_vm_close() 233 return mem->vaddr; in __videobuf_to_vaddr() 250 if (!mem->vaddr) { in __videobuf_iolock() [all …]
|
D | videobuf2-dma-contig.c | 30 void *vaddr; member 101 if (!buf->vaddr && buf->db_attach) in vb2_dc_vaddr() 102 buf->vaddr = dma_buf_vmap(buf->db_attach->dmabuf); in vb2_dc_vaddr() 104 return buf->vaddr; in vb2_dc_vaddr() 153 dma_free_coherent(buf->dev, buf->size, buf->vaddr, buf->dma_addr); in vb2_dc_put() 169 buf->vaddr = dma_alloc_coherent(dev, size, &buf->dma_addr, in vb2_dc_alloc() 171 if (!buf->vaddr) { in vb2_dc_alloc() 207 ret = dma_mmap_coherent(buf->dev, vma, buf->vaddr, in vb2_dc_mmap() 351 return buf->vaddr + pgnum * PAGE_SIZE; in vb2_dc_dmabuf_ops_kmap() 358 return buf->vaddr; in vb2_dc_dmabuf_ops_vmap() [all …]
|
/linux-4.1.27/arch/sparc/mm/ |
D | io-unit.c | 94 static unsigned long iounit_get_area(struct iounit_struct *iounit, unsigned long vaddr, int size) in iounit_get_area() argument 100 npages = ((vaddr & ~PAGE_MASK) + size + (PAGE_SIZE-1)) >> PAGE_SHIFT; in iounit_get_area() 109 IOD(("iounit_get_area(%08lx,%d[%d])=", vaddr, size, npages)); in iounit_get_area() 124 panic("iounit_get_area: Couldn't find free iopte slots for (%08lx,%d)\n", vaddr, size); in iounit_get_area() 132 iopte = MKIOPTE(__pa(vaddr & PAGE_MASK)); in iounit_get_area() 133 vaddr = IOUNIT_DMA_BASE + (scan << PAGE_SHIFT) + (vaddr & ~PAGE_MASK); in iounit_get_area() 138 IOD(("%08lx\n", vaddr)); in iounit_get_area() 139 return vaddr; in iounit_get_area() 142 static __u32 iounit_get_scsi_one(struct device *dev, char *vaddr, unsigned long len) in iounit_get_scsi_one() argument 148 ret = iounit_get_area(iounit, (unsigned long)vaddr, len); in iounit_get_scsi_one() [all …]
|
D | highmem.c | 53 unsigned long vaddr; in kmap_atomic() local 63 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic() 67 __flush_cache_one(vaddr); in kmap_atomic() 78 __flush_tlb_one(vaddr); in kmap_atomic() 83 return (void*) vaddr; in kmap_atomic() 89 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 92 if (vaddr < FIXADDR_START) { // FIXME in __kunmap_atomic() 104 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN+idx)); in __kunmap_atomic() 108 __flush_cache_one(vaddr); in __kunmap_atomic() 117 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic() [all …]
|
D | tlb.c | 69 static void tlb_batch_add_one(struct mm_struct *mm, unsigned long vaddr, in tlb_batch_add_one() argument 75 vaddr &= PAGE_MASK; in tlb_batch_add_one() 77 vaddr |= 0x1UL; in tlb_batch_add_one() 87 flush_tsb_user_page(mm, vaddr); in tlb_batch_add_one() 88 global_flush_tlb_page(mm, vaddr); in tlb_batch_add_one() 95 tb->vaddrs[nr] = vaddr; in tlb_batch_add_one() 104 void tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, in tlb_batch_add() argument 126 if ((paddr ^ vaddr) & (1 << 13)) in tlb_batch_add() 132 tlb_batch_add_one(mm, vaddr, pte_exec(orig)); in tlb_batch_add() 136 static void tlb_batch_pmd_scan(struct mm_struct *mm, unsigned long vaddr, in tlb_batch_pmd_scan() argument [all …]
|
D | leon_mm.c | 35 unsigned long leon_swprobe(unsigned long vaddr, unsigned long *paddr) in leon_swprobe() argument 88 ptr += ((((vaddr) >> LEON_PGD_SH) & LEON_PGD_M) * 4); in leon_swprobe() 111 ptr += (((vaddr >> LEON_PMD_SH) & LEON_PMD_M) * 4); in leon_swprobe() 139 ptr += (((vaddr >> LEON_PTE_SH) & LEON_PTE_M) * 4); in leon_swprobe() 160 (vaddr & ~(-1 << LEON_PTE_SH)) | ((pte & ~0xff) << 4); in leon_swprobe() 164 (vaddr & ~(-1 << LEON_PMD_SH)) | ((pte & ~0xff) << 4); in leon_swprobe() 168 (vaddr & ~(-1 << LEON_PGD_SH)) | ((pte & ~0xff) << 4); in leon_swprobe() 172 paddr_calc = vaddr; in leon_swprobe()
|
D | iommu.c | 207 static u32 iommu_get_scsi_one(struct device *dev, char *vaddr, unsigned int len) in iommu_get_scsi_one() argument 214 off = (unsigned long)vaddr & ~PAGE_MASK; in iommu_get_scsi_one() 216 page = virt_to_page((unsigned long)vaddr & PAGE_MASK); in iommu_get_scsi_one() 221 static __u32 iommu_get_scsi_one_gflush(struct device *dev, char *vaddr, unsigned long len) in iommu_get_scsi_one_gflush() argument 224 return iommu_get_scsi_one(dev, vaddr, len); in iommu_get_scsi_one_gflush() 227 static __u32 iommu_get_scsi_one_pflush(struct device *dev, char *vaddr, unsigned long len) in iommu_get_scsi_one_pflush() argument 229 unsigned long page = ((unsigned long) vaddr) & PAGE_MASK; in iommu_get_scsi_one_pflush() 231 while(page < ((unsigned long)(vaddr + len))) { in iommu_get_scsi_one_pflush() 235 return iommu_get_scsi_one(dev, vaddr, len); in iommu_get_scsi_one_pflush() 299 static void iommu_release_scsi_one(struct device *dev, __u32 vaddr, unsigned long len) in iommu_release_scsi_one() argument [all …]
|
/linux-4.1.27/arch/arm/mach-ixp4xx/include/mach/ |
D | io.h | 98 const u8 *vaddr, int count) in __indirect_writesb() argument 101 writeb(*vaddr++, bus_addr); in __indirect_writesb() 121 const u16 *vaddr, int count) in __indirect_writesw() argument 124 writew(*vaddr++, bus_addr); in __indirect_writesw() 140 const u32 *vaddr, int count) in __indirect_writesl() argument 143 writel(*vaddr++, bus_addr); in __indirect_writesl() 163 u8 *vaddr, u32 count) in __indirect_readsb() argument 166 *vaddr++ = readb(bus_addr); in __indirect_readsb() 186 u16 *vaddr, u32 count) in __indirect_readsw() argument 189 *vaddr++ = readw(bus_addr); in __indirect_readsw() [all …]
|
/linux-4.1.27/arch/metag/mm/ |
D | highmem.c | 43 unsigned long vaddr; in kmap_atomic() local 53 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic() 59 return (void *)vaddr; in kmap_atomic() 65 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 78 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic() 79 flush_tlb_kernel_range(vaddr, vaddr + PAGE_SIZE); in __kunmap_atomic() 95 unsigned long vaddr; in kmap_atomic_pfn() local 102 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_pfn() 107 flush_tlb_kernel_range(vaddr, vaddr + PAGE_SIZE); in kmap_atomic_pfn() 109 return (void *)vaddr; in kmap_atomic_pfn() [all …]
|
D | init.c | 270 unsigned long vaddr; in allocate_pgtables() local 272 vaddr = start; in allocate_pgtables() 273 i = pgd_index(vaddr); in allocate_pgtables() 274 j = pmd_index(vaddr); in allocate_pgtables() 277 for ( ; (i < PTRS_PER_PGD) && (vaddr != end); pgd++, i++) { in allocate_pgtables() 279 for (; (j < PTRS_PER_PMD) && (vaddr != end); pmd++, j++) { in allocate_pgtables() 280 vaddr += PMD_SIZE; in allocate_pgtables() 294 unsigned long vaddr, end; in fixedrange_init() local 303 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; in fixedrange_init() 305 allocate_pgtables(vaddr, end); in fixedrange_init() [all …]
|
D | mmu-meta2.c | 18 unsigned long mmu_read_first_level_page(unsigned long vaddr) in mmu_read_first_level_page() argument 25 if (is_global_space(vaddr)) in mmu_read_first_level_page() 26 vaddr &= ~0x80000000; in mmu_read_first_level_page() 28 offset = vaddr >> PGDIR_SHIFT; in mmu_read_first_level_page() 53 unsigned long mmu_read_second_level_page(unsigned long vaddr) in mmu_read_second_level_page() argument 55 return __builtin_meta2_cacherd((void *)(vaddr & PAGE_MASK)); in mmu_read_second_level_page()
|
D | mmu-meta1.c | 111 unsigned long mmu_read_first_level_page(unsigned long vaddr) in mmu_read_first_level_page() argument 113 return metag_in32(pgd_entry_addr(vaddr)); in mmu_read_first_level_page() 116 unsigned long mmu_read_second_level_page(unsigned long vaddr) in mmu_read_second_level_page() argument 118 return metag_in32(pgtable_entry_addr(vaddr)); in mmu_read_second_level_page()
|
/linux-4.1.27/arch/m68k/sun3/ |
D | dvma.c | 23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) in dvma_page() argument 36 if(ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] != pte) { in dvma_page() 37 sun3_put_pte(vaddr, pte); in dvma_page() 38 ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] = pte; in dvma_page() 41 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page() 50 unsigned long vaddr; in dvma_map_iommu() local 52 vaddr = dvma_btov(baddr); in dvma_map_iommu() 54 end = vaddr + len; in dvma_map_iommu() 56 while(vaddr < end) { in dvma_map_iommu() 57 dvma_page(kaddr, vaddr); in dvma_map_iommu() [all …]
|
D | mmu_emu.c | 117 void print_pte_vaddr (unsigned long vaddr) in print_pte_vaddr() argument 119 printk (" vaddr=%lx [%02lx]", vaddr, sun3_get_segmap (vaddr)); in print_pte_vaddr() 120 print_pte (__pte (sun3_get_pte (vaddr))); in print_pte_vaddr() 282 inline void mmu_emu_map_pmeg (int context, int vaddr) in mmu_emu_map_pmeg() argument 288 vaddr &= ~SUN3_PMEG_MASK; in mmu_emu_map_pmeg() 297 curr_pmeg, context, vaddr); in mmu_emu_map_pmeg() 309 if(vaddr >= PAGE_OFFSET) { in mmu_emu_map_pmeg() 315 sun3_put_segmap (vaddr, curr_pmeg); in mmu_emu_map_pmeg() 325 sun3_put_segmap (vaddr, curr_pmeg); in mmu_emu_map_pmeg() 328 pmeg_vaddr[curr_pmeg] = vaddr; in mmu_emu_map_pmeg() [all …]
|
/linux-4.1.27/drivers/lguest/ |
D | page_tables.c | 83 static pgd_t *spgd_addr(struct lg_cpu *cpu, u32 i, unsigned long vaddr) in spgd_addr() argument 85 unsigned int index = pgd_index(vaddr); in spgd_addr() 97 static pmd_t *spmd_addr(struct lg_cpu *cpu, pgd_t spgd, unsigned long vaddr) in spmd_addr() argument 99 unsigned int index = pmd_index(vaddr); in spmd_addr() 115 static pte_t *spte_addr(struct lg_cpu *cpu, pgd_t spgd, unsigned long vaddr) in spte_addr() argument 118 pmd_t *pmd = spmd_addr(cpu, spgd, vaddr); in spte_addr() 129 return &page[pte_index(vaddr)]; in spte_addr() 136 static unsigned long gpgd_addr(struct lg_cpu *cpu, unsigned long vaddr) in gpgd_addr() argument 138 unsigned int index = vaddr >> (PGDIR_SHIFT); in gpgd_addr() 144 static unsigned long gpmd_addr(pgd_t gpgd, unsigned long vaddr) in gpmd_addr() argument [all …]
|
/linux-4.1.27/arch/tile/mm/ |
D | highmem.c | 20 #define kmap_get_pte(vaddr) \ argument 21 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)),\ 22 (vaddr)), (vaddr)) 200 unsigned long vaddr; in kmap_atomic_prot() local 215 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot() 216 pte = kmap_get_pte(vaddr); in kmap_atomic_prot() 220 kmap_atomic_register(page, type, vaddr, pte, mk_pte(page, prot)); in kmap_atomic_prot() 222 return (void *)vaddr; in kmap_atomic_prot() 235 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 237 if (vaddr >= __fix_to_virt(FIX_KMAP_END) && in __kunmap_atomic() [all …]
|
D | init.c | 180 unsigned long vaddr; in page_table_range_init() local 183 for (vaddr = start; vaddr < end; vaddr += PMD_SIZE) { in page_table_range_init() 184 pmd_t *pmd = get_pmd(pgd, vaddr); in page_table_range_init() 643 unsigned long vaddr; in permanent_kmaps_init() local 645 vaddr = PKMAP_BASE; in permanent_kmaps_init() 646 page_table_range_init(vaddr, vaddr + PAGE_SIZE*LAST_PKMAP, pgd_base); in permanent_kmaps_init() 648 pgd = swapper_pg_dir + pgd_index(vaddr); in permanent_kmaps_init() 649 pud = pud_offset(pgd, vaddr); in permanent_kmaps_init() 650 pmd = pmd_offset(pud, vaddr); in permanent_kmaps_init() 651 pte = pte_offset_kernel(pmd, vaddr); in permanent_kmaps_init()
|
/linux-4.1.27/arch/m68k/sun3x/ |
D | dvma.c | 83 unsigned long vaddr, int len) in dvma_map_cpu() argument 90 vaddr &= PAGE_MASK; in dvma_map_cpu() 92 end = PAGE_ALIGN(vaddr + len); in dvma_map_cpu() 96 kaddr, vaddr); in dvma_map_cpu() 98 pgd = pgd_offset_k(vaddr); in dvma_map_cpu() 104 if((pmd = pmd_alloc(&init_mm, pgd, vaddr)) == NULL) { in dvma_map_cpu() 109 if((end & PGDIR_MASK) > (vaddr & PGDIR_MASK)) in dvma_map_cpu() 110 end2 = (vaddr + (PGDIR_SIZE-1)) & PGDIR_MASK; in dvma_map_cpu() 118 if((pte = pte_alloc_kernel(pmd, vaddr)) == NULL) { in dvma_map_cpu() 123 if((end2 & PMD_MASK) > (vaddr & PMD_MASK)) in dvma_map_cpu() [all …]
|
/linux-4.1.27/mm/ |
D | highmem.c | 154 struct page *kmap_to_page(void *vaddr) in kmap_to_page() argument 156 unsigned long addr = (unsigned long)vaddr; in kmap_to_page() 219 unsigned long vaddr; in map_new_virtual() local 261 vaddr = PKMAP_ADDR(last_pkmap_nr); in map_new_virtual() 262 set_pte_at(&init_mm, vaddr, in map_new_virtual() 266 set_page_address(page, (void *)vaddr); in map_new_virtual() 268 return vaddr; in map_new_virtual() 281 unsigned long vaddr; in kmap_high() local 288 vaddr = (unsigned long)page_address(page); in kmap_high() 289 if (!vaddr) in kmap_high() [all …]
|
D | dmapool.c | 58 void *vaddr; member 217 *(int *)(page->vaddr + offset) = next; in pool_initialise_page() 229 page->vaddr = dma_alloc_coherent(pool->dev, pool->allocation, in pool_alloc_page() 231 if (page->vaddr) { in pool_alloc_page() 233 memset(page->vaddr, POOL_POISON_FREED, pool->allocation); in pool_alloc_page() 255 memset(page->vaddr, POOL_POISON_FREED, pool->allocation); in pool_free_page() 257 dma_free_coherent(pool->dev, pool->allocation, page->vaddr, dma); in pool_free_page() 292 pool->name, page->vaddr); in dma_pool_destroy() 296 pool->name, page->vaddr); in dma_pool_destroy() 347 page->offset = *(int *)(page->vaddr + offset); in dma_pool_alloc() [all …]
|
/linux-4.1.27/arch/sparc/include/asm/ |
D | viking.h | 212 static inline unsigned long viking_hwprobe(unsigned long vaddr) in viking_hwprobe() argument 216 vaddr &= PAGE_MASK; in viking_hwprobe() 220 : "r" (vaddr | 0x400), "i" (ASI_M_FLUSH_PROBE)); in viking_hwprobe() 227 : "r" (vaddr | 0x200), "i" (ASI_M_FLUSH_PROBE)); in viking_hwprobe() 229 vaddr &= ~SRMMU_PGDIR_MASK; in viking_hwprobe() 230 vaddr >>= PAGE_SHIFT; in viking_hwprobe() 231 return val | (vaddr << 8); in viking_hwprobe() 237 : "r" (vaddr | 0x100), "i" (ASI_M_FLUSH_PROBE)); in viking_hwprobe() 239 vaddr &= ~SRMMU_REAL_PMD_MASK; in viking_hwprobe() 240 vaddr >>= PAGE_SHIFT; in viking_hwprobe() [all …]
|
D | tlbflush_64.h | 19 void flush_tsb_user_page(struct mm_struct *mm, unsigned long vaddr); 48 void __flush_tlb_page(unsigned long context, unsigned long vaddr); 53 static inline void global_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr) in global_flush_tlb_page() argument 55 __flush_tlb_page(CTX_HWBITS(mm->context), vaddr); in global_flush_tlb_page() 61 void smp_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr); 63 #define global_flush_tlb_page(mm, vaddr) \ argument 64 smp_flush_tlb_page(mm, vaddr)
|
D | cacheflush_64.h | 57 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 59 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 61 flush_ptrace_access(vma, page, vaddr, src, len, 0); \ 64 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument 66 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 68 flush_ptrace_access(vma, page, vaddr, dst, len, 1); \
|
D | cacheflush_32.h | 21 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 23 flush_cache_page(vma, vaddr, page_to_pfn(page));\ 26 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument 28 flush_cache_page(vma, vaddr, page_to_pfn(page));\
|
D | dma.h | 110 #define mmu_get_scsi_one(dev,vaddr,len) \ argument 111 sparc32_dma_ops->get_scsi_one(dev, vaddr, len) 114 #define mmu_release_scsi_one(dev,vaddr,len) \ argument 115 sparc32_dma_ops->release_scsi_one(dev, vaddr,len)
|
D | oplib_64.h | 155 unsigned long vaddr); 159 unsigned long vaddr); 175 unsigned long vaddr, unsigned long paddr); 176 void prom_unmap(unsigned long size, unsigned long vaddr);
|
/linux-4.1.27/arch/powerpc/mm/ |
D | highmem.c | 34 unsigned long vaddr; in kmap_atomic_prot() local 44 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot() 48 __set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot), 1); in kmap_atomic_prot() 49 local_flush_tlb_page(NULL, vaddr); in kmap_atomic_prot() 51 return (void*) vaddr; in kmap_atomic_prot() 57 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 60 if (vaddr < __fix_to_virt(FIX_KMAP_END)) { in __kunmap_atomic() 72 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in __kunmap_atomic() 78 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic() 79 local_flush_tlb_page(NULL, vaddr); in __kunmap_atomic()
|
D | hash_utils_64.c | 185 unsigned long vaddr, paddr; in htab_bolt_mapping() local 197 for (vaddr = vstart, paddr = pstart; vaddr < vend; in htab_bolt_mapping() 198 vaddr += step, paddr += step) { in htab_bolt_mapping() 200 unsigned long vsid = get_kernel_vsid(vaddr, ssize); in htab_bolt_mapping() 201 unsigned long vpn = hpt_vpn(vaddr, vsid, ssize); in htab_bolt_mapping() 210 if (overlaps_kernel_text(vaddr, vaddr + step)) in htab_bolt_mapping() 214 if (overlaps_kvm_tmp(vaddr, vaddr + step)) in htab_bolt_mapping() 228 overlaps_interrupt_vector_text(vaddr, vaddr + step)) in htab_bolt_mapping() 252 unsigned long vaddr; in htab_remove_mapping() local 264 for (vaddr = vstart; vaddr < vend; vaddr += step) in htab_remove_mapping() [all …]
|
D | dma-noncoherent.c | 218 unsigned long vaddr = c->vm_start; in __dma_alloc_coherent() local 230 map_page(vaddr, page_to_phys(page), in __dma_alloc_coherent() 233 vaddr += PAGE_SIZE; in __dma_alloc_coherent() 257 void __dma_free_coherent(size_t size, void *vaddr) in __dma_free_coherent() argument 266 c = ppc_vm_region_find(&consistent_head, (unsigned long)vaddr); in __dma_free_coherent() 309 __func__, vaddr); in __dma_free_coherent() 317 void __dma_sync(void *vaddr, size_t size, int direction) in __dma_sync() argument 319 unsigned long start = (unsigned long)vaddr; in __dma_sync()
|
D | mem.c | 72 static inline pte_t *virt_to_kpte(unsigned long vaddr) in virt_to_kpte() argument 74 return pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), in virt_to_kpte() 75 vaddr), vaddr), vaddr); in virt_to_kpte() 432 void clear_user_page(void *page, unsigned long vaddr, struct page *pg) in clear_user_page() argument 445 void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, in copy_user_page() argument
|
/linux-4.1.27/arch/x86/mm/ |
D | init_32.c | 108 pmd_t * __init populate_extra_pmd(unsigned long vaddr) in populate_extra_pmd() argument 110 int pgd_idx = pgd_index(vaddr); in populate_extra_pmd() 111 int pmd_idx = pmd_index(vaddr); in populate_extra_pmd() 116 pte_t * __init populate_extra_pte(unsigned long vaddr) in populate_extra_pte() argument 118 int pte_idx = pte_index(vaddr); in populate_extra_pte() 121 pmd = populate_extra_pmd(vaddr); in populate_extra_pte() 133 unsigned long vaddr; in page_table_range_init_count() local 138 vaddr = start; in page_table_range_init_count() 139 pgd_idx = pgd_index(vaddr); in page_table_range_init_count() 140 pmd_idx = pmd_index(vaddr); in page_table_range_init_count() [all …]
|
D | highmem_32.c | 35 unsigned long vaddr; in kmap_atomic_prot() local 46 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot() 51 return (void *)vaddr; in kmap_atomic_prot() 73 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 75 if (vaddr >= __fix_to_virt(FIX_KMAP_END) && in __kunmap_atomic() 76 vaddr <= __fix_to_virt(FIX_KMAP_BEGIN)) { in __kunmap_atomic() 83 WARN_ON_ONCE(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in __kunmap_atomic() 91 kpte_clear_flush(kmap_pte-idx, vaddr); in __kunmap_atomic() 97 BUG_ON(vaddr < PAGE_OFFSET); in __kunmap_atomic() 98 BUG_ON(vaddr >= (unsigned long)high_memory); in __kunmap_atomic() [all …]
|
D | pgtable_32.c | 27 void set_pte_vaddr(unsigned long vaddr, pte_t pteval) in set_pte_vaddr() argument 34 pgd = swapper_pg_dir + pgd_index(vaddr); in set_pte_vaddr() 39 pud = pud_offset(pgd, vaddr); in set_pte_vaddr() 44 pmd = pmd_offset(pud, vaddr); in set_pte_vaddr() 49 pte = pte_offset_kernel(pmd, vaddr); in set_pte_vaddr() 51 set_pte_at(&init_mm, vaddr, pte, pteval); in set_pte_vaddr() 53 pte_clear(&init_mm, vaddr, pte); in set_pte_vaddr() 59 __flush_tlb_one(vaddr); in set_pte_vaddr()
|
D | iomap_32.c | 59 unsigned long vaddr; in kmap_atomic_prot_pfn() local 66 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot_pfn() 70 return (void *)vaddr; in kmap_atomic_prot_pfn() 97 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in iounmap_atomic() local 99 if (vaddr >= __fix_to_virt(FIX_KMAP_END) && in iounmap_atomic() 100 vaddr <= __fix_to_virt(FIX_KMAP_BEGIN)) { in iounmap_atomic() 107 WARN_ON_ONCE(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in iounmap_atomic() 115 kpte_clear_flush(kmap_pte-idx, vaddr); in iounmap_atomic()
|
D | ioremap.c | 31 int ioremap_change_attr(unsigned long vaddr, unsigned long size, in ioremap_change_attr() argument 40 err = _set_memory_uc(vaddr, nrpages); in ioremap_change_attr() 43 err = _set_memory_wc(vaddr, nrpages); in ioremap_change_attr() 46 err = _set_memory_wb(vaddr, nrpages); in ioremap_change_attr() 85 unsigned long offset, vaddr; in __ioremap_caller() local 186 vaddr = (unsigned long) area->addr; in __ioremap_caller() 191 if (ioremap_page_range(vaddr, vaddr + size, phys_addr, prot)) in __ioremap_caller() 194 ret_addr = (void __iomem *) (vaddr + offset); in __ioremap_caller() 356 unsigned long vaddr; in xlate_dev_mem_ptr() local 362 vaddr = (unsigned long)ioremap_cache(start, PAGE_SIZE); in xlate_dev_mem_ptr() [all …]
|
D | init_64.c | 233 static pud_t *fill_pud(pgd_t *pgd, unsigned long vaddr) in fill_pud() argument 242 return pud_offset(pgd, vaddr); in fill_pud() 245 static pmd_t *fill_pmd(pud_t *pud, unsigned long vaddr) in fill_pmd() argument 254 return pmd_offset(pud, vaddr); in fill_pmd() 257 static pte_t *fill_pte(pmd_t *pmd, unsigned long vaddr) in fill_pte() argument 265 return pte_offset_kernel(pmd, vaddr); in fill_pte() 268 void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte) in set_pte_vaddr_pud() argument 274 pud = pud_page + pud_index(vaddr); in set_pte_vaddr_pud() 275 pmd = fill_pmd(pud, vaddr); in set_pte_vaddr_pud() 276 pte = fill_pte(pmd, vaddr); in set_pte_vaddr_pud() [all …]
|
D | pageattr.c | 32 unsigned long *vaddr; member 130 void clflush_cache_range(void *vaddr, unsigned int size) in clflush_cache_range() argument 132 void *vend = vaddr + size - 1; in clflush_cache_range() 136 for (; vaddr < vend; vaddr += boot_cpu_data.x86_clflush_size) in clflush_cache_range() 137 clflushopt(vaddr); in clflush_cache_range() 1090 static int __cpa_process_fault(struct cpa_data *cpa, unsigned long vaddr, in __cpa_process_fault() argument 1094 return populate_pgd(cpa, vaddr); in __cpa_process_fault() 1109 if (within(vaddr, PAGE_OFFSET, in __cpa_process_fault() 1112 cpa->pfn = __pa(vaddr) >> PAGE_SHIFT; in __cpa_process_fault() 1116 "vaddr = %lx cpa->vaddr = %lx\n", vaddr, in __cpa_process_fault() [all …]
|
/linux-4.1.27/drivers/base/ |
D | dma-mapping.c | 22 void *vaddr; member 30 dma_free_coherent(dev, this->size, this->vaddr, this->dma_handle); in dmam_coherent_release() 37 dma_free_noncoherent(dev, this->size, this->vaddr, this->dma_handle); in dmam_noncoherent_release() 44 if (this->vaddr == match->vaddr) { in dmam_match() 69 void *vaddr; in dmam_alloc_coherent() local 75 vaddr = dma_alloc_coherent(dev, size, dma_handle, gfp); in dmam_alloc_coherent() 76 if (!vaddr) { in dmam_alloc_coherent() 81 dr->vaddr = vaddr; in dmam_alloc_coherent() 87 return vaddr; in dmam_alloc_coherent() 100 void dmam_free_coherent(struct device *dev, size_t size, void *vaddr, in dmam_free_coherent() argument [all …]
|
D | dma-coherent.c | 213 int dma_release_from_coherent(struct device *dev, int order, void *vaddr) in dma_release_from_coherent() argument 217 if (mem && vaddr >= mem->virt_base && vaddr < in dma_release_from_coherent() 219 int page = (vaddr - mem->virt_base) >> PAGE_SHIFT; in dma_release_from_coherent() 247 void *vaddr, size_t size, int *ret) in dma_mmap_from_coherent() argument 251 if (mem && vaddr >= mem->virt_base && vaddr + size <= in dma_mmap_from_coherent() 254 int start = (vaddr - mem->virt_base) >> PAGE_SHIFT; in dma_mmap_from_coherent()
|
/linux-4.1.27/drivers/block/ |
D | cciss.h | 66 void __iomem *vaddr; member 226 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command() 227 readl(h->vaddr + SA5_SCRATCHPAD_OFFSET); in SA5_submit_command() 243 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 244 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 249 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 250 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 263 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask() 264 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask() 269 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask() [all …]
|
D | smart1,2.h | 44 writel(c->busaddr, h->vaddr + S42XX_REQUEST_PORT_OFFSET); in smart4_submit_command() 56 writel(0, h->vaddr + S42XX_REPLY_INTR_MASK_OFFSET); in smart4_intr_mask() 60 h->vaddr + S42XX_REPLY_INTR_MASK_OFFSET); in smart4_intr_mask() 72 return (!readl(h->vaddr + S42XX_REQUEST_PORT_OFFSET)); in smart4_fifo_full() 82 = readl(h->vaddr + S42XX_REPLY_PORT_OFFSET); in smart4_completed() 90 writel(0, h->vaddr + S42XX_REPLY_PORT_OFFSET); in smart4_completed() 103 readl(h->vaddr + S42XX_INTR_STATUS); in smart4_intr_pending() 123 writel(c->busaddr, h->vaddr + COMMAND_FIFO); in smart2_submit_command() 128 writel(val, h->vaddr + INTR_MASK); in smart2_intr_mask() 133 return readl(h->vaddr + COMMAND_FIFO); in smart2_fifo_full() [all …]
|
/linux-4.1.27/arch/frv/mm/ |
D | cache-page.c | 25 void *vaddr; in flush_dcache_page() local 29 vaddr = kmap_atomic_primary(page); in flush_dcache_page() 31 frv_dcache_writeback((unsigned long) vaddr, (unsigned long) vaddr + PAGE_SIZE); in flush_dcache_page() 33 kunmap_atomic_primary(vaddr); in flush_dcache_page() 53 void *vaddr; in flush_icache_user_range() local 57 vaddr = kmap_atomic_primary(page); in flush_icache_user_range() 59 start = (start & ~PAGE_MASK) | (unsigned long) vaddr; in flush_icache_user_range() 62 kunmap_atomic_primary(vaddr); in flush_icache_user_range()
|
D | dma-alloc.c | 141 void consistent_free(void *vaddr) in consistent_free() argument 145 vfree(vaddr); in consistent_free() 151 void consistent_sync(void *vaddr, size_t size, int direction) in consistent_sync() argument 153 unsigned long start = (unsigned long) vaddr; in consistent_sync()
|
/linux-4.1.27/arch/um/kernel/ |
D | mem.c | 94 unsigned long vaddr; in fixrange_init() local 96 vaddr = start; in fixrange_init() 97 i = pgd_index(vaddr); in fixrange_init() 98 j = pmd_index(vaddr); in fixrange_init() 101 for ( ; (i < PTRS_PER_PGD) && (vaddr < end); pgd++, i++) { in fixrange_init() 102 pud = pud_offset(pgd, vaddr); in fixrange_init() 105 pmd = pmd_offset(pud, vaddr); in fixrange_init() 106 for (; (j < PTRS_PER_PMD) && (vaddr < end); pmd++, j++) { in fixrange_init() 108 vaddr += PMD_SIZE; in fixrange_init() 123 unsigned long v, vaddr = FIXADDR_USER_START; in fixaddr_user_init() local [all …]
|
/linux-4.1.27/arch/xtensa/mm/ |
D | cache.c | 64 unsigned long vaddr) in kmap_invalidate_coherent() argument 66 if (!DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) { in kmap_invalidate_coherent() 84 unsigned long vaddr, unsigned long *paddr) in coherent_kvaddr() argument 86 if (PageHighMem(page) || !DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) { in coherent_kvaddr() 88 return (void *)(base + (vaddr & DCACHE_ALIAS_MASK)); in coherent_kvaddr() 95 void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument 98 void *kvaddr = coherent_kvaddr(page, TLBTEMP_BASE_1, vaddr, &paddr); in clear_user_highpage() 101 kmap_invalidate_coherent(page, vaddr); in clear_user_highpage() 108 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument 111 void *dst_vaddr = coherent_kvaddr(dst, TLBTEMP_BASE_1, vaddr, in copy_user_highpage() [all …]
|
D | mmu.c | 21 static void * __init init_pmd(unsigned long vaddr, unsigned long n_pages) in init_pmd() argument 23 pgd_t *pgd = pgd_offset_k(vaddr); in init_pmd() 24 pmd_t *pmd = pmd_offset(pgd, vaddr); in init_pmd() 31 __func__, vaddr, n_pages); in init_pmd()
|
D | highmem.c | 43 unsigned long vaddr; in kmap_atomic() local 51 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic() 57 return (void *)vaddr; in kmap_atomic()
|
/linux-4.1.27/arch/avr32/include/asm/ |
D | cacheflush.h | 26 static inline void invalidate_dcache_line(void *vaddr) in invalidate_dcache_line() argument 30 : "r"(vaddr), "n"(CACHE_OP_DCACHE_INVALIDATE) in invalidate_dcache_line() 38 static inline void clean_dcache_line(void *vaddr) in clean_dcache_line() argument 42 : "r"(vaddr), "n"(CACHE_OP_DCACHE_CLEAN) in clean_dcache_line() 50 static inline void flush_dcache_line(void *vaddr) in flush_dcache_line() argument 54 : "r"(vaddr), "n"(CACHE_OP_DCACHE_CLEAN_INVAL) in flush_dcache_line() 62 static inline void invalidate_icache_line(void *vaddr) in invalidate_icache_line() argument 66 : "r"(vaddr), "n"(CACHE_OP_ICACHE_INVALIDATE) in invalidate_icache_line() 122 unsigned long vaddr, void *dst, const void *src, 126 struct page *page, unsigned long vaddr, void *dst, in copy_from_user_page() argument
|
/linux-4.1.27/arch/xtensa/include/asm/ |
D | fixmap.h | 66 static inline unsigned long virt_to_fix(const unsigned long vaddr) in virt_to_fix() argument 68 BUG_ON(vaddr >= FIXADDR_TOP || vaddr < FIXADDR_START); in virt_to_fix() 69 return __virt_to_fix(vaddr); in virt_to_fix() 74 #define kmap_get_fixmap_pte(vaddr) \ argument 76 pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), \ 77 (vaddr) \
|
D | page.h | 149 extern void clear_page_alias(void *vaddr, unsigned long paddr); 154 void clear_user_highpage(struct page *page, unsigned long vaddr); 157 unsigned long vaddr, struct vm_area_struct *vma); 159 # define clear_user_page(page, vaddr, pg) clear_page(page) argument 160 # define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
|
/linux-4.1.27/arch/frv/include/asm/ |
D | virtconvert.h | 24 #define phys_to_virt(vaddr) ((void *) ((unsigned long)(vaddr) + PAGE_OFFSET)) argument 25 #define virt_to_phys(vaddr) ((unsigned long) (vaddr) - PAGE_OFFSET) argument 29 #define phys_to_virt(vaddr) ((void *) (vaddr)) argument 30 #define virt_to_phys(vaddr) ((unsigned long) (vaddr)) argument
|
D | page.h | 11 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument 17 #define clear_user_page(pgaddr, vaddr, page) memset((pgaddr), 0, PAGE_SIZE) argument 18 #define copy_user_page(vto, vfrom, vaddr, topg) memcpy((vto), (vfrom), PAGE_SIZE) argument 45 #define __pa(vaddr) virt_to_phys((void *) (unsigned long) (vaddr)) argument
|
D | pci.h | 29 extern void consistent_free(void *vaddr); 30 extern void consistent_sync(void *vaddr, size_t size, int direction); 39 void *vaddr, dma_addr_t dma_handle);
|
D | cacheflush.h | 96 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 99 flush_icache_user_range((vma), (page), (vaddr), (len)); \ 102 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
/linux-4.1.27/drivers/misc/sgi-gru/ |
D | grufault.c | 61 struct vm_area_struct *gru_find_vma(unsigned long vaddr) in gru_find_vma() argument 65 vma = find_vma(current->mm, vaddr); in gru_find_vma() 66 if (vma && vma->vm_start <= vaddr && vma->vm_ops == &gru_vm_ops) in gru_find_vma() 79 static struct gru_thread_state *gru_find_lock_gts(unsigned long vaddr) in gru_find_lock_gts() argument 86 vma = gru_find_vma(vaddr); in gru_find_lock_gts() 88 gts = gru_find_thread_state(vma, TSID(vaddr, vma)); in gru_find_lock_gts() 96 static struct gru_thread_state *gru_alloc_locked_gts(unsigned long vaddr) in gru_alloc_locked_gts() argument 103 vma = gru_find_vma(vaddr); in gru_alloc_locked_gts() 107 gts = gru_alloc_thread_state(vma, TSID(vaddr, vma)); in gru_alloc_locked_gts() 191 unsigned long vaddr, int write, in non_atomic_pte_lookup() argument [all …]
|
D | gruhandles.c | 152 unsigned long vaddr, unsigned long vaddrmask, in tgh_invalidate() argument 156 tgh->vaddr = vaddr; in tgh_invalidate() 170 unsigned long vaddr, int asid, int dirty, in tfh_write_only() argument 174 tfh->fillvaddr = vaddr; in tfh_write_only() 186 unsigned long vaddr, int asid, int dirty, in tfh_write_restart() argument 190 tfh->fillvaddr = vaddr; in tfh_write_restart()
|
/linux-4.1.27/arch/arm/kernel/ |
D | crash_dump.c | 37 void *vaddr; in copy_oldmem_page() local 42 vaddr = ioremap(__pfn_to_phys(pfn), PAGE_SIZE); in copy_oldmem_page() 43 if (!vaddr) in copy_oldmem_page() 47 if (copy_to_user(buf, vaddr + offset, csize)) { in copy_oldmem_page() 48 iounmap(vaddr); in copy_oldmem_page() 52 memcpy(buf, vaddr + offset, csize); in copy_oldmem_page() 55 iounmap(vaddr); in copy_oldmem_page()
|
/linux-4.1.27/arch/x86/kernel/ |
D | crash_dump_64.c | 29 void *vaddr; in copy_oldmem_page() local 34 vaddr = ioremap_cache(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page() 35 if (!vaddr) in copy_oldmem_page() 39 if (copy_to_user(buf, vaddr + offset, csize)) { in copy_oldmem_page() 40 iounmap(vaddr); in copy_oldmem_page() 44 memcpy(buf, vaddr + offset, csize); in copy_oldmem_page() 47 iounmap(vaddr); in copy_oldmem_page()
|
D | crash_dump_32.c | 53 void *vaddr; in copy_oldmem_page() local 61 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page() 64 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_page() 65 kunmap_atomic(vaddr); in copy_oldmem_page() 70 kunmap_atomic(vaddr); in copy_oldmem_page() 73 copy_page(kdump_buf_page, vaddr); in copy_oldmem_page() 74 kunmap_atomic(vaddr); in copy_oldmem_page()
|
D | pci-swiotlb.c | 21 void *vaddr; in x86_swiotlb_alloc_coherent() local 23 vaddr = dma_generic_alloc_coherent(hwdev, size, dma_handle, flags, in x86_swiotlb_alloc_coherent() 25 if (vaddr) in x86_swiotlb_alloc_coherent() 26 return vaddr; in x86_swiotlb_alloc_coherent() 32 void *vaddr, dma_addr_t dma_addr, in x86_swiotlb_free_coherent() argument 36 swiotlb_free_coherent(dev, size, vaddr, dma_addr); in x86_swiotlb_free_coherent() 38 dma_generic_free_coherent(dev, size, vaddr, dma_addr, attrs); in x86_swiotlb_free_coherent()
|
D | tboot.c | 123 static int map_tboot_page(unsigned long vaddr, unsigned long pfn, in map_tboot_page() argument 131 pgd = pgd_offset(&tboot_mm, vaddr); in map_tboot_page() 132 pud = pud_alloc(&tboot_mm, pgd, vaddr); in map_tboot_page() 135 pmd = pmd_alloc(&tboot_mm, pud, vaddr); in map_tboot_page() 138 pte = pte_alloc_map(&tboot_mm, NULL, pmd, vaddr); in map_tboot_page() 141 set_pte_at(&tboot_mm, vaddr, pte, pfn_pte(pfn, prot)); in map_tboot_page() 146 static int map_tboot_pages(unsigned long vaddr, unsigned long start_pfn, in map_tboot_pages() argument 154 for (; nr > 0; nr--, vaddr += PAGE_SIZE, start_pfn++) { in map_tboot_pages() 155 if (map_tboot_page(vaddr, start_pfn, PAGE_KERNEL_EXEC)) in map_tboot_pages()
|
D | machine_kexec_32.c | 104 unsigned long vaddr, unsigned long paddr) in machine_kexec_page_table_set_one() argument 108 pgd += pgd_index(vaddr); in machine_kexec_page_table_set_one() 113 pud = pud_offset(pgd, vaddr); in machine_kexec_page_table_set_one() 114 pmd = pmd_offset(pud, vaddr); in machine_kexec_page_table_set_one() 117 pte = pte_offset_kernel(pmd, vaddr); in machine_kexec_page_table_set_one()
|
/linux-4.1.27/arch/arc/mm/ |
D | cache_arc700.c | 202 static inline void __cache_line_loop(unsigned long paddr, unsigned long vaddr, in __cache_line_loop() argument 232 vaddr &= CACHE_LINE_MASK; in __cache_line_loop() 239 paddr |= (vaddr >> PAGE_SHIFT) & 0x1F; in __cache_line_loop() 254 write_aux_reg(aux_cmd, vaddr); in __cache_line_loop() 255 vaddr += L1_CACHE_BYTES; in __cache_line_loop() 326 static inline void __dc_line_op(unsigned long paddr, unsigned long vaddr, in __dc_line_op() argument 336 __cache_line_loop(paddr, vaddr, sz, cacheop); in __dc_line_op() 346 #define __dc_line_op(paddr, vaddr, sz, cacheop) argument 407 __ic_line_inv_vaddr_local(unsigned long paddr, unsigned long vaddr, in __ic_line_inv_vaddr_local() argument 413 __cache_line_loop(paddr, vaddr, sz, OP_INV_IC); in __ic_line_inv_vaddr_local() [all …]
|
D | ioremap.c | 44 void __iomem *vaddr; in ioremap_prot() local 73 vaddr = (void __iomem *)area->addr; in ioremap_prot() 74 if (ioremap_page_range((unsigned long)vaddr, in ioremap_prot() 75 (unsigned long)vaddr + size, paddr, prot)) { in ioremap_prot() 76 vunmap((void __force *)vaddr); in ioremap_prot() 79 return (void __iomem *)(off + (char __iomem *)vaddr); in ioremap_prot()
|
/linux-4.1.27/arch/parisc/include/asm/ |
D | cacheflush.h | 51 static inline void flush_kernel_vmap_range(void *vaddr, int size) in flush_kernel_vmap_range() argument 53 unsigned long start = (unsigned long)vaddr; in flush_kernel_vmap_range() 57 static inline void invalidate_kernel_vmap_range(void *vaddr, int size) in invalidate_kernel_vmap_range() argument 59 unsigned long start = (unsigned long)vaddr; in invalidate_kernel_vmap_range() 60 void *cursor = vaddr; in invalidate_kernel_vmap_range() 62 for ( ; cursor < vaddr + size; cursor += PAGE_SIZE) { in invalidate_kernel_vmap_range() 92 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 94 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 99 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument 101 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ [all …]
|
D | dma-mapping.h | 13 void (*free_consistent)(struct device *dev, size_t size, void *vaddr, dma_addr_t iova); 68 void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument 70 hppa_dma_ops->free_consistent(dev, size, vaddr, dma_handle); in dma_free_coherent() 75 void *vaddr, dma_addr_t dma_handle) in dma_free_noncoherent() argument 77 hppa_dma_ops->free_consistent(dev, size, vaddr, dma_handle); in dma_free_noncoherent() 191 dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument 195 flush_kernel_dcache_range((unsigned long)vaddr, size); in dma_cache_sync()
|
/linux-4.1.27/arch/m68k/mm/ |
D | cache.c | 14 static unsigned long virt_to_phys_slow(unsigned long vaddr) in virt_to_phys_slow() argument 36 : "0" (vaddr)); in virt_to_phys_slow() 46 : "a" (vaddr)); in virt_to_phys_slow() 49 return (mmusr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow() 57 : "a" (vaddr), "d" (get_fs().seg)); in virt_to_phys_slow() 63 return (*descaddr & 0xfe000000) | (vaddr & 0x01ffffff); in virt_to_phys_slow() 65 return (*descaddr & 0xfffc0000) | (vaddr & 0x0003ffff); in virt_to_phys_slow() 67 return (*descaddr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow()
|
/linux-4.1.27/arch/sh/kernel/ |
D | crash_dump.c | 28 void *vaddr; in copy_oldmem_page() local 33 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page() 36 if (copy_to_user(buf, (vaddr + offset), csize)) { in copy_oldmem_page() 37 iounmap(vaddr); in copy_oldmem_page() 41 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_page() 43 iounmap(vaddr); in copy_oldmem_page()
|
/linux-4.1.27/arch/mn10300/include/asm/ |
D | highmem.h | 75 unsigned long vaddr; in kmap_atomic() local 84 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic() 90 local_flush_tlb_one(vaddr); in kmap_atomic() 92 return (void *)vaddr; in kmap_atomic() 95 static inline void __kunmap_atomic(unsigned long vaddr) in __kunmap_atomic() argument 99 if (vaddr < FIXADDR_START) { /* FIXME */ in __kunmap_atomic() 111 if (vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)) in __kunmap_atomic() 119 local_flush_tlb_one(vaddr); in __kunmap_atomic()
|
/linux-4.1.27/arch/powerpc/kernel/ |
D | crash_dump.c | 72 static size_t copy_oldmem_vaddr(void *vaddr, char *buf, size_t csize, in copy_oldmem_vaddr() argument 76 if (copy_to_user((char __user *)buf, (vaddr + offset), csize)) in copy_oldmem_vaddr() 79 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_vaddr() 100 void *vaddr; in copy_oldmem_page() local 110 vaddr = __va(paddr); in copy_oldmem_page() 111 csize = copy_oldmem_vaddr(vaddr, buf, csize, offset, userbuf); in copy_oldmem_page() 113 vaddr = __ioremap(paddr, PAGE_SIZE, 0); in copy_oldmem_page() 114 csize = copy_oldmem_vaddr(vaddr, buf, csize, offset, userbuf); in copy_oldmem_page() 115 iounmap(vaddr); in copy_oldmem_page()
|
D | io-workarounds.c | 28 static struct iowa_bus *iowa_pci_find(unsigned long vaddr, unsigned long paddr) in iowa_pci_find() argument 38 if (vaddr) { in iowa_pci_find() 41 if ((vaddr >= vstart) && (vaddr <= vend)) in iowa_pci_find() 68 unsigned long vaddr, paddr; in iowa_mem_find_bus() local 71 vaddr = (unsigned long)PCI_FIX_ADDR(addr); in iowa_mem_find_bus() 72 if (vaddr < PHB_IO_BASE || vaddr >= PHB_IO_END) in iowa_mem_find_bus() 78 ptep = __find_linux_pte_or_hugepte(init_mm.pgd, vaddr, in iowa_mem_find_bus() 86 bus = iowa_pci_find(vaddr, paddr); in iowa_mem_find_bus() 104 unsigned long vaddr = (unsigned long)pci_io_base + port; in iowa_pio_find_bus() local 105 return iowa_pci_find(vaddr, 0); in iowa_pio_find_bus()
|
D | iommu.c | 460 unsigned long vaddr, npages, entry, slen; in ppc_iommu_map_sg() local 469 vaddr = (unsigned long) sg_virt(s); in ppc_iommu_map_sg() 470 npages = iommu_num_pages(vaddr, slen, IOMMU_PAGE_SIZE(tbl)); in ppc_iommu_map_sg() 473 (vaddr & ~PAGE_MASK) == 0) in ppc_iommu_map_sg() 478 DBG(" - vaddr: %lx, size: %lx\n", vaddr, slen); in ppc_iommu_map_sg() 484 "vaddr %lx npages %lu\n", tbl, vaddr, in ppc_iommu_map_sg() 499 vaddr & IOMMU_PAGE_MASK(tbl), in ppc_iommu_map_sg() 559 unsigned long vaddr, npages; in ppc_iommu_map_sg() local 561 vaddr = s->dma_address & IOMMU_PAGE_MASK(tbl); in ppc_iommu_map_sg() 564 __iommu_free(tbl, vaddr, npages); in ppc_iommu_map_sg() [all …]
|
/linux-4.1.27/drivers/scsi/ |
D | hpsa.h | 128 void __iomem *vaddr; member 359 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command() 360 (void) readl(h->vaddr + SA5_SCRATCHPAD_OFFSET); in SA5_submit_command() 366 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command_no_read() 372 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command_ioaccel2() 384 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 385 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 389 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 390 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 398 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_performant_intr_mask() [all …]
|
D | sun3_scsi.c | 402 unsigned char *vaddr; in sun3scsi_dma_finish() local 404 vaddr = (unsigned char *)dvma_vmetov(sun3_dma_orig_addr); in sun3scsi_dma_finish() 406 vaddr += (sun3_dma_orig_count - fifo); in sun3scsi_dma_finish() 407 vaddr--; in sun3scsi_dma_finish() 411 *vaddr = (dregs->bpack_lo & 0xff00) >> 8; in sun3scsi_dma_finish() 412 vaddr--; in sun3scsi_dma_finish() 415 *vaddr = (dregs->bpack_hi & 0x00ff); in sun3scsi_dma_finish() 416 vaddr--; in sun3scsi_dma_finish() 419 *vaddr = (dregs->bpack_hi & 0xff00) >> 8; in sun3scsi_dma_finish() 448 unsigned char *vaddr; in sun3scsi_dma_finish() local [all …]
|
D | NCR_Q720.c | 79 int irq, int slot, __u32 paddr, void __iomem *vaddr) in NCR_Q720_probe_one() argument 84 __u8 scsr1 = readb(vaddr + NCR_Q720_SCSR_OFFSET + 1); in NCR_Q720_probe_one() 85 __u8 differential = readb(vaddr + NCR_Q720_SCSR_OFFSET) & 0x20; in NCR_Q720_probe_one() 94 writeb(scsr1, vaddr + NCR_Q720_SCSR_OFFSET + 1); in NCR_Q720_probe_one() 96 version = readb(vaddr + 0x18) >> 4; in NCR_Q720_probe_one() 106 device.slot.base_v = vaddr; in NCR_Q720_probe_one() 118 scsr1 = readb(vaddr + NCR_Q720_SCSR_OFFSET + 1); in NCR_Q720_probe_one() 121 writeb(scsr1, vaddr + NCR_Q720_SCSR_OFFSET + 1); in NCR_Q720_probe_one()
|
/linux-4.1.27/include/linux/ |
D | highmem.h | 23 static inline void flush_kernel_vmap_range(void *vaddr, int size) in flush_kernel_vmap_range() argument 26 static inline void invalidate_kernel_vmap_range(void *vaddr, int size) in invalidate_kernel_vmap_range() argument 132 static inline void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument 135 clear_user_page(addr, vaddr, page); in clear_user_highpage() 158 unsigned long vaddr) in __alloc_zeroed_user_highpage() argument 161 vma, vaddr); in __alloc_zeroed_user_highpage() 164 clear_user_highpage(page, vaddr); in __alloc_zeroed_user_highpage() 180 unsigned long vaddr) in alloc_zeroed_user_highpage_movable() argument 182 return __alloc_zeroed_user_highpage(__GFP_MOVABLE, vma, vaddr); in alloc_zeroed_user_highpage_movable() 225 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument [all …]
|
D | io-mapping.h | 97 io_mapping_unmap_atomic(void __iomem *vaddr) in io_mapping_unmap_atomic() argument 99 iounmap_atomic(vaddr); in io_mapping_unmap_atomic() 114 io_mapping_unmap(void __iomem *vaddr) in io_mapping_unmap() argument 116 iounmap(vaddr); in io_mapping_unmap() 149 io_mapping_unmap_atomic(void __iomem *vaddr) in io_mapping_unmap_atomic() argument 162 io_mapping_unmap(void __iomem *vaddr) in io_mapping_unmap() argument
|
D | uprobes.h | 79 unsigned long vaddr; member 101 extern int set_swbp(struct arch_uprobe *aup, struct mm_struct *mm, unsigned long vaddr); 102 extern int set_orig_insn(struct arch_uprobe *aup, struct mm_struct *mm, unsigned long vaddr); 107 extern int uprobe_write_opcode(struct mm_struct *mm, unsigned long vaddr, uprobe_opcode_t); 132 extern void arch_uprobe_copy_ixol(struct page *page, unsigned long vaddr,
|
/linux-4.1.27/arch/mips/kernel/ |
D | crash_dump.c | 29 void *vaddr; in copy_oldmem_page() local 34 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page() 37 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_page() 38 kunmap_atomic(vaddr); in copy_oldmem_page() 45 copy_page(kdump_buf_page, vaddr); in copy_oldmem_page() 46 kunmap_atomic(vaddr); in copy_oldmem_page()
|
D | mips-r2-to-r6-emul.c | 907 unsigned long vaddr; in mipsr2_decoder() local 1200 vaddr = regs->regs[MIPSInst_RS(inst)] + MIPSInst_SIMM(inst); in mipsr2_decoder() 1201 if (!access_ok(VERIFY_READ, vaddr, 4)) { in mipsr2_decoder() 1202 current->thread.cp0_baduaddr = vaddr; in mipsr2_decoder() 1261 "+&r"(vaddr), "+&r"(err) in mipsr2_decoder() 1273 vaddr = regs->regs[MIPSInst_RS(inst)] + MIPSInst_SIMM(inst); in mipsr2_decoder() 1274 if (!access_ok(VERIFY_READ, vaddr, 4)) { in mipsr2_decoder() 1275 current->thread.cp0_baduaddr = vaddr; in mipsr2_decoder() 1336 "+&r"(vaddr), "+&r"(err) in mipsr2_decoder() 1347 vaddr = regs->regs[MIPSInst_RS(inst)] + MIPSInst_SIMM(inst); in mipsr2_decoder() [all …]
|
/linux-4.1.27/arch/hexagon/include/asm/ |
D | fixmap.h | 31 #define kmap_get_fixmap_pte(vaddr) \ argument 32 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), \ 33 (vaddr)), (vaddr)), (vaddr))
|
/linux-4.1.27/arch/arm/include/asm/ |
D | page.h | 114 void (*cpu_clear_user_highpage)(struct page *page, unsigned long vaddr); 116 unsigned long vaddr, struct vm_area_struct *vma); 130 extern void __cpu_clear_user_highpage(struct page *page, unsigned long vaddr); 132 unsigned long vaddr, struct vm_area_struct *vma); 135 #define clear_user_highpage(page,vaddr) \ argument 136 __cpu_clear_user_highpage(page, vaddr) 139 #define copy_user_highpage(to,from,vaddr,vma) \ argument 140 __cpu_copy_user_highpage(to, from, vaddr, vma)
|
D | page-nommu.h | 20 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument 26 #define clear_user_page(page, vaddr, pg) clear_page(page) argument 27 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
|
/linux-4.1.27/arch/arm64/include/asm/ |
D | dma-mapping.h | 117 void *vaddr; in dma_alloc_attrs() local 119 if (dma_alloc_from_coherent(dev, size, dma_handle, &vaddr)) in dma_alloc_attrs() 120 return vaddr; in dma_alloc_attrs() 122 vaddr = ops->alloc(dev, size, dma_handle, flags, attrs); in dma_alloc_attrs() 123 debug_dma_alloc_coherent(dev, size, *dma_handle, vaddr); in dma_alloc_attrs() 124 return vaddr; in dma_alloc_attrs() 128 void *vaddr, dma_addr_t dev_addr, in dma_free_attrs() argument 133 if (dma_release_from_coherent(dev, get_order(size), vaddr)) in dma_free_attrs() 136 debug_dma_free_coherent(dev, size, vaddr, dev_addr); in dma_free_attrs() 137 ops->free(dev, size, vaddr, dev_addr, attrs); in dma_free_attrs()
|
D | page.h | 59 #define clear_user_page(addr,vaddr,pg) __cpu_clear_user_page(addr, vaddr) argument 60 #define copy_user_page(to,from,vaddr,pg) __cpu_copy_user_page(to, from, vaddr) argument
|
/linux-4.1.27/arch/frv/mb93090-mb00/ |
D | pci-dma.c | 34 void dma_free_coherent(struct device *hwdev, size_t size, void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument 36 consistent_free(vaddr); in dma_free_coherent() 57 void *vaddr; in dma_map_sg() local 65 vaddr = kmap_atomic_primary(sg_page(&sg[i])); in dma_map_sg() 67 frv_dcache_writeback((unsigned long) vaddr, in dma_map_sg() 68 (unsigned long) vaddr + PAGE_SIZE); in dma_map_sg() 72 kunmap_atomic_primary(vaddr); in dma_map_sg()
|
/linux-4.1.27/arch/score/include/asm/ |
D | page.h | 36 #define clear_user_page(pgaddr, vaddr, page) memset((pgaddr), 0, PAGE_SIZE) argument 37 #define copy_user_page(vto, vfrom, vaddr, topg) \ argument 67 #define virt_to_pfn(vaddr) (phys_to_pfn((__pa(vaddr)))) argument 70 #define virt_to_page(vaddr) (pfn_to_page(virt_to_pfn(vaddr))) argument 83 #define virt_addr_valid(vaddr) (pfn_valid(virt_to_pfn(vaddr))) argument
|
D | cacheflush.h | 38 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument 41 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 45 flush_cache_page(vma, vaddr, page_to_pfn(page));\
|
D | tlbflush.h | 23 extern void local_flush_tlb_one(unsigned long vaddr); 32 #define flush_tlb_one(vaddr) local_flush_tlb_one(vaddr) argument
|
/linux-4.1.27/arch/nios2/include/asm/ |
D | dma-mapping.h | 17 static inline void __dma_sync_for_device(void *vaddr, size_t size, in __dma_sync_for_device() argument 22 invalidate_dcache_range((unsigned long)vaddr, in __dma_sync_for_device() 23 (unsigned long)(vaddr + size)); in __dma_sync_for_device() 31 flush_dcache_range((unsigned long)vaddr, in __dma_sync_for_device() 32 (unsigned long)(vaddr + size)); in __dma_sync_for_device() 39 static inline void __dma_sync_for_cpu(void *vaddr, size_t size, in __dma_sync_for_cpu() argument 45 invalidate_dcache_range((unsigned long)vaddr, in __dma_sync_for_cpu() 46 (unsigned long)(vaddr + size)); in __dma_sync_for_cpu() 62 void *vaddr, dma_addr_t dma_handle); 125 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument
|
D | page.h | 53 extern void clear_user_page(void *addr, unsigned long vaddr, struct page *page); 54 extern void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, 93 # define virt_to_page(vaddr) pfn_to_page(PFN_DOWN(virt_to_phys(vaddr))) argument 94 # define virt_addr_valid(vaddr) pfn_valid(PFN_DOWN(virt_to_phys(vaddr))) argument
|
D | io.h | 55 #define phys_to_virt(vaddr) \ argument 56 ((void *)((unsigned long)(vaddr) | CONFIG_NIOS2_KERNEL_REGION_BASE)) 58 #define virt_to_phys(vaddr) \ argument 59 ((unsigned long)((unsigned long)(vaddr) & ~0xE0000000))
|
/linux-4.1.27/arch/s390/include/asm/ |
D | idals.h | 29 idal_is_needed(void *vaddr, unsigned int length) in idal_is_needed() argument 31 return ((__pa(vaddr) + length - 1) >> 31) != 0; in idal_is_needed() 38 static inline unsigned int idal_nr_words(void *vaddr, unsigned int length) in idal_nr_words() argument 40 return ((__pa(vaddr) & (IDA_BLOCK_SIZE-1)) + length + in idal_nr_words() 48 void *vaddr, unsigned int length) in idal_create_words() argument 53 paddr = __pa(vaddr); in idal_create_words() 70 set_normalized_cda(struct ccw1 * ccw, void *vaddr) in set_normalized_cda() argument 77 nridaws = idal_nr_words(vaddr, ccw->count); in set_normalized_cda() 83 idal_create_words(idal, vaddr, ccw->count); in set_normalized_cda() 85 vaddr = idal; in set_normalized_cda() [all …]
|
D | page.h | 59 #define clear_user_page(page, vaddr, pg) clear_page(page) argument 60 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument 62 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument 63 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
|
/linux-4.1.27/kernel/events/ |
D | uprobes.c | 118 unsigned long vaddr; /* Page(s) of instruction slots */ member 144 static loff_t vaddr_to_offset(struct vm_area_struct *vma, unsigned long vaddr) in vaddr_to_offset() argument 146 return ((loff_t)vma->vm_pgoff << PAGE_SHIFT) + (vaddr - vma->vm_start); in vaddr_to_offset() 241 static void copy_from_page(struct page *page, unsigned long vaddr, void *dst, int len) in copy_from_page() argument 244 memcpy(dst, kaddr + (vaddr & ~PAGE_MASK), len); in copy_from_page() 248 static void copy_to_page(struct page *page, unsigned long vaddr, const void *src, int len) in copy_to_page() argument 251 memcpy(kaddr + (vaddr & ~PAGE_MASK), src, len); in copy_to_page() 255 static int verify_opcode(struct page *page, unsigned long vaddr, uprobe_opcode_t *new_opcode) in verify_opcode() argument 269 copy_from_page(page, vaddr, &old_opcode, UPROBE_SWBP_INSN_SIZE); in verify_opcode() 300 int uprobe_write_opcode(struct mm_struct *mm, unsigned long vaddr, in uprobe_write_opcode() argument [all …]
|
/linux-4.1.27/arch/metag/include/asm/ |
D | fixmap.h | 56 #define kmap_get_fixmap_pte(vaddr) \ argument 58 pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), \ 59 (vaddr) \
|
/linux-4.1.27/arch/mips/include/asm/ |
D | fixmap.h | 72 #define kmap_get_fixmap_pte(vaddr) \ argument 73 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), (vaddr))
|
D | cacheflush.h | 101 struct page *page, unsigned long vaddr, void *dst, const void *src, 105 struct page *page, unsigned long vaddr, void *dst, const void *src, 135 extern void (*__flush_kernel_vmap_range)(unsigned long vaddr, int size); 137 static inline void flush_kernel_vmap_range(void *vaddr, int size) in flush_kernel_vmap_range() argument 140 __flush_kernel_vmap_range((unsigned long) vaddr, size); in flush_kernel_vmap_range() 143 static inline void invalidate_kernel_vmap_range(void *vaddr, int size) in invalidate_kernel_vmap_range() argument 146 __flush_kernel_vmap_range((unsigned long) vaddr, size); in invalidate_kernel_vmap_range()
|
D | tlbflush.h | 23 extern void local_flush_tlb_one(unsigned long vaddr); 33 extern void flush_tlb_one(unsigned long vaddr); 43 #define flush_tlb_one(vaddr) local_flush_tlb_one(vaddr) argument
|
D | dma-mapping.h | 65 extern void dma_cache_sync(struct device *dev, void *vaddr, size_t size, 87 void *vaddr, dma_addr_t dma_handle, in dma_free_attrs() argument 92 ops->free(dev, size, vaddr, dma_handle, attrs); in dma_free_attrs() 94 debug_dma_free_coherent(dev, size, vaddr, dma_handle); in dma_free_attrs() 102 void *vaddr, dma_addr_t dma_handle);
|
/linux-4.1.27/arch/mips/boot/ |
D | elf2ecoff.c | 58 unsigned long vaddr; member 105 if (base->vaddr + base->len != new->vaddr) { in combine() 107 base->len = new->vaddr - base->vaddr; in combine() 282 text.vaddr = data.vaddr = bss.vaddr = 0; in main() 362 ndata.vaddr = ph[i].p_vaddr; in main() 364 nbss.vaddr = ph[i].p_vaddr + ph[i].p_filesz; in main() 372 ntxt.vaddr = ph[i].p_vaddr; in main() 392 if (text.vaddr > data.vaddr || data.vaddr > bss.vaddr || in main() 393 text.vaddr + text.len > data.vaddr in main() 394 || data.vaddr + data.len > bss.vaddr) { in main() [all …]
|
/linux-4.1.27/arch/powerpc/platforms/44x/ |
D | canyonlands.c | 69 void __iomem *vaddr; in ppc460ex_canyonlands_fixup() local 94 vaddr = of_iomap(np, 0); in ppc460ex_canyonlands_fixup() 97 if (!vaddr) { in ppc460ex_canyonlands_fixup() 117 setbits32((vaddr + GPIO0_OSRH), 0x42000000); in ppc460ex_canyonlands_fixup() 118 setbits32((vaddr + GPIO0_TSRH), 0x42000000); in ppc460ex_canyonlands_fixup() 120 iounmap(vaddr); in ppc460ex_canyonlands_fixup()
|
/linux-4.1.27/arch/sh/boards/mach-sdk7786/ |
D | sram.c | 25 void __iomem *vaddr; in fpga_sram_init() local 53 vaddr = ioremap(phys, SZ_2K); in fpga_sram_init() 54 if (unlikely(!vaddr)) { in fpga_sram_init() 63 ret = gen_pool_add(sram_pool, (unsigned long)vaddr, SZ_2K, -1); in fpga_sram_init() 66 iounmap(vaddr); in fpga_sram_init()
|
/linux-4.1.27/arch/mips/dec/ |
D | kn01-berr.c | 83 long asid, entryhi, vaddr; in dec_kn01_be_backend() local 105 vaddr = regs->regs[insn.i_format.rs] + in dec_kn01_be_backend() 108 vaddr = (long)pc; in dec_kn01_be_backend() 109 if (KSEGX(vaddr) == CKSEG0 || KSEGX(vaddr) == CKSEG1) in dec_kn01_be_backend() 110 address = CPHYSADDR(vaddr); in dec_kn01_be_backend() 115 entryhi |= vaddr & ~(PAGE_SIZE - 1); in dec_kn01_be_backend() 123 offset = vaddr & (PAGE_SIZE - 1); in dec_kn01_be_backend()
|
/linux-4.1.27/drivers/xen/xenbus/ |
D | xenbus_client.c | 73 void **vaddr); 74 int (*unmap)(struct xenbus_device *dev, void *vaddr); 375 int xenbus_grant_ring(struct xenbus_device *dev, void *vaddr, in xenbus_grant_ring() argument 382 unsigned long addr = (unsigned long)vaddr + in xenbus_grant_ring() 465 unsigned int nr_grefs, void **vaddr) in xenbus_map_ring_valloc() argument 467 return ring_ops->map(dev, gnt_refs, nr_grefs, vaddr); in xenbus_map_ring_valloc() 539 void **vaddr) in xenbus_map_ring_valloc_pv() argument 549 *vaddr = NULL; in xenbus_map_ring_valloc_pv() 581 *vaddr = area->addr; in xenbus_map_ring_valloc_pv() 597 void **vaddr) in xenbus_map_ring_valloc_hvm() argument [all …]
|
/linux-4.1.27/arch/arc/include/asm/ |
D | cacheflush.h | 35 void __sync_icache_dcache(unsigned long paddr, unsigned long vaddr, int len); 36 void __inv_icache_page(unsigned long paddr, unsigned long vaddr); 37 void ___flush_dcache_page(unsigned long paddr, unsigned long vaddr); 110 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 114 __sync_icache_dcache((unsigned long)(dst), vaddr, len); \ 117 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
D | page.h | 16 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument 91 #define __pa(vaddr) ((unsigned long)vaddr) argument
|
/linux-4.1.27/sound/pci/asihpi/ |
D | hpios.c | 50 p_mem_area->vaddr = in hpios_locked_mem_alloc() 54 if (p_mem_area->vaddr) { in hpios_locked_mem_alloc() 57 p_mem_area->vaddr); in hpios_locked_mem_alloc() 73 p_mem_area->vaddr, p_mem_area->dma_handle); in hpios_locked_mem_free() 77 p_mem_area->vaddr); in hpios_locked_mem_free()
|
/linux-4.1.27/arch/ia64/kernel/ |
D | crash_dump.c | 37 void *vaddr; in copy_oldmem_page() local 41 vaddr = __va(pfn<<PAGE_SHIFT); in copy_oldmem_page() 43 if (copy_to_user(buf, (vaddr + offset), csize)) { in copy_oldmem_page() 47 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_page()
|
/linux-4.1.27/arch/metag/kernel/ |
D | dma.c | 225 unsigned long vaddr = c->vm_start; in dma_alloc_coherent() local 226 pte_t *pte = consistent_pte + CONSISTENT_OFFSET(vaddr); in dma_alloc_coherent() 241 set_pte_at(&init_mm, vaddr, in dma_alloc_coherent() 247 vaddr += PAGE_SIZE; in dma_alloc_coherent() 272 void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument 282 c = metag_vm_region_find(&consistent_head, (unsigned long)vaddr); in dma_free_coherent() 329 __func__, vaddr); in dma_free_coherent() 426 void dma_sync_for_device(void *vaddr, size_t size, int dma_direction) in dma_sync_for_device() argument 444 flush_dcache_region(vaddr, size); in dma_sync_for_device() 452 writeback_dcache_region(vaddr, size); in dma_sync_for_device() [all …]
|
D | tcm.c | 51 unsigned long vaddr; in tcm_alloc() local 58 vaddr = gen_pool_alloc(pool->pool, len); in tcm_alloc() 59 if (!vaddr) in tcm_alloc() 62 return vaddr; in tcm_alloc()
|
/linux-4.1.27/arch/xtensa/kernel/ |
D | pci-dma.c | 68 void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument 70 unsigned long addr = (unsigned long)vaddr + in dma_free_coherent() 81 void consistent_sync(void *vaddr, size_t size, int direction) in consistent_sync() argument 87 __invalidate_dcache_range((unsigned long)vaddr, in consistent_sync() 93 __flush_invalidate_dcache_range((unsigned long)vaddr, in consistent_sync()
|
/linux-4.1.27/arch/sparc/prom/ |
D | misc_64.c | 203 unsigned long tte_data, unsigned long vaddr) in tlb_load() argument 212 args[5] = vaddr; in tlb_load() 224 unsigned long vaddr) in prom_itlb_load() argument 226 return tlb_load("SUNW,itlb-load", index, tte_data, vaddr); in prom_itlb_load() 231 unsigned long vaddr) in prom_dtlb_load() argument 233 return tlb_load("SUNW,dtlb-load", index, tte_data, vaddr); in prom_dtlb_load() 237 unsigned long vaddr, unsigned long paddr) in prom_map() argument 249 args[7] = vaddr; in prom_map() 262 void prom_unmap(unsigned long size, unsigned long vaddr) in prom_unmap() argument 272 args[6] = vaddr; in prom_unmap()
|
/linux-4.1.27/drivers/infiniband/hw/ipath/ |
D | ipath_keys.c | 143 isge->vaddr = (void *) sge->addr; in ipath_lkey_ok() 176 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in ipath_lkey_ok() 200 u32 len, u64 vaddr, u32 rkey, int acc) in ipath_rkey_ok() argument 223 sge->vaddr = (void *) vaddr; in ipath_rkey_ok() 239 off = vaddr - mr->iova; in ipath_rkey_ok() 240 if (unlikely(vaddr < mr->iova || off + len > mr->length || in ipath_rkey_ok() 258 sge->vaddr = mr->map[m]->segs[n].vaddr + off; in ipath_rkey_ok()
|
/linux-4.1.27/arch/microblaze/include/asm/ |
D | page.h | 81 # define clear_user_page(pgaddr, vaddr, page) memset((pgaddr), 0, PAGE_SIZE) argument 82 # define copy_user_page(vto, vfrom, vaddr, topg) \ argument 147 # define virt_to_pfn(vaddr) (phys_to_pfn((__pa(vaddr)))) argument 157 # define virt_to_page(vaddr) (pfn_to_page(virt_to_pfn(vaddr))) argument 175 #define virt_addr_valid(vaddr) (pfn_valid(virt_to_pfn(vaddr))) argument
|
D | highmem.h | 81 unsigned long idx, vaddr = (unsigned long) ptr; in kmap_atomic_to_page() local 84 if (vaddr < FIXADDR_START) in kmap_atomic_to_page() 87 idx = virt_to_fix(vaddr); in kmap_atomic_to_page()
|
/linux-4.1.27/arch/openrisc/include/asm/ |
D | fixmap.h | 81 static inline unsigned long virt_to_fix(const unsigned long vaddr) in virt_to_fix() argument 83 BUG_ON(vaddr >= FIXADDR_TOP || vaddr < FIXADDR_START); in virt_to_fix() 84 return __virt_to_fix(vaddr); in virt_to_fix()
|
D | page.h | 43 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument 49 #define clear_user_page(page, vaddr, pg) clear_page(page) argument 50 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
|
/linux-4.1.27/drivers/staging/android/ion/ |
D | ion_test.c | 68 void *vaddr = vmap(&page, 1, VM_MAP, pgprot); in ion_handle_test_dma() local 72 if (!vaddr) { in ion_handle_test_dma() 78 ret = copy_from_user(vaddr + offset, ptr, to_copy); in ion_handle_test_dma() 80 ret = copy_to_user(ptr, vaddr + offset, to_copy); in ion_handle_test_dma() 82 vunmap(vaddr); in ion_handle_test_dma() 118 void *vaddr = dma_buf_kmap(dma_buf, page_offset); in ion_handle_test_kernel() local 120 if (!vaddr) in ion_handle_test_kernel() 126 ret = copy_from_user(vaddr + copy_offset, ptr, to_copy); in ion_handle_test_kernel() 128 ret = copy_to_user(ptr, vaddr + copy_offset, to_copy); in ion_handle_test_kernel() 130 dma_buf_kunmap(dma_buf, page_offset, vaddr); in ion_handle_test_kernel()
|
/linux-4.1.27/arch/mn10300/mm/ |
D | pgtable.c | 35 void set_pmd_pfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags) in set_pmd_pfn() argument 41 if (vaddr & (PMD_SIZE-1)) { /* vaddr is misaligned */ in set_pmd_pfn() 49 pgd = swapper_pg_dir + pgd_index(vaddr); in set_pmd_pfn() 54 pud = pud_offset(pgd, vaddr); in set_pmd_pfn() 55 pmd = pmd_offset(pud, vaddr); in set_pmd_pfn() 61 local_flush_tlb_one(vaddr); in set_pmd_pfn()
|
/linux-4.1.27/arch/nios2/mm/ |
D | tlb.c | 77 unsigned long vaddr = CONFIG_NIOS2_IO_REGION_BASE + in flush_tlb_one_pid() local 81 vaddr, way, (pid_misc >> TLBMISC_PID_SHIFT)); in flush_tlb_one_pid() 83 WRCTL(CTL_PTEADDR, (vaddr >> 12) << 2); in flush_tlb_one_pid() 139 unsigned long vaddr = CONFIG_NIOS2_IO_REGION_BASE + in flush_tlb_one() local 144 vaddr, way, (pid_misc >> TLBMISC_PID_SHIFT)); in flush_tlb_one() 148 WRCTL(CTL_PTEADDR, (vaddr >> 12) << 2); in flush_tlb_one() 247 unsigned long vaddr = CONFIG_NIOS2_IO_REGION_BASE; in flush_tlb_all() local 260 WRCTL(CTL_PTEADDR, ((vaddr) >> PAGE_SHIFT) << 2); in flush_tlb_all() 263 vaddr += 1UL << 12; in flush_tlb_all()
|
D | cacheflush.c | 246 void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, in copy_user_page() argument 249 __flush_dcache(vaddr, vaddr + PAGE_SIZE); in copy_user_page() 250 __flush_icache(vaddr, vaddr + PAGE_SIZE); in copy_user_page() 256 void clear_user_page(void *addr, unsigned long vaddr, struct page *page) in clear_user_page() argument 258 __flush_dcache(vaddr, vaddr + PAGE_SIZE); in clear_user_page() 259 __flush_icache(vaddr, vaddr + PAGE_SIZE); in clear_user_page()
|
/linux-4.1.27/arch/hexagon/mm/ |
D | vm_tlb.c | 49 void flush_tlb_one(unsigned long vaddr) in flush_tlb_one() argument 51 __vmclrmap((void *)vaddr, PAGE_SIZE); in flush_tlb_one() 78 void flush_tlb_page(struct vm_area_struct *vma, unsigned long vaddr) in flush_tlb_page() argument 83 __vmclrmap((void *)vaddr, PAGE_SIZE); in flush_tlb_page()
|
/linux-4.1.27/arch/s390/kvm/ |
D | gaccess.c | 543 union vaddress vaddr = {.addr = gva}; in guest_translate() local 559 if (vaddr.rfx01 > asce.tl) in guest_translate() 561 ptr += vaddr.rfx * 8; in guest_translate() 564 if (vaddr.rfx) in guest_translate() 566 if (vaddr.rsx01 > asce.tl) in guest_translate() 568 ptr += vaddr.rsx * 8; in guest_translate() 571 if (vaddr.rfx || vaddr.rsx) in guest_translate() 573 if (vaddr.rtx01 > asce.tl) in guest_translate() 575 ptr += vaddr.rtx * 8; in guest_translate() 578 if (vaddr.rfx || vaddr.rsx || vaddr.rtx) in guest_translate() [all …]
|
/linux-4.1.27/drivers/net/wireless/ath/ath6kl/ |
D | target.h | 333 #define AR6003_VTOP(vaddr) ((vaddr) & 0x001fffff) argument 334 #define AR6004_VTOP(vaddr) (vaddr) argument 336 #define TARG_VTOP(target_type, vaddr) \ argument 337 (((target_type) == TARGET_TYPE_AR6003) ? AR6003_VTOP(vaddr) : \ 338 (((target_type) == TARGET_TYPE_AR6004) ? AR6004_VTOP(vaddr) : 0))
|
/linux-4.1.27/arch/sh/include/asm/ |
D | dma-mapping.h | 39 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, 79 void *vaddr, dma_addr_t dma_handle, in dma_free_attrs() argument 84 if (dma_release_from_coherent(dev, get_order(size), vaddr)) in dma_free_attrs() 87 debug_dma_free_coherent(dev, size, vaddr, dma_handle); in dma_free_attrs() 89 ops->free(dev, size, vaddr, dma_handle, attrs); in dma_free_attrs() 97 void *vaddr, dma_addr_t dma_handle,
|
/linux-4.1.27/arch/powerpc/platforms/512x/ |
D | mpc512x_shared.c | 226 void __iomem *vaddr; in mpc512x_init_diu() local 250 vaddr = ioremap(desc, sizeof(struct diu_ad)); in mpc512x_init_diu() 251 if (!vaddr) { in mpc512x_init_diu() 255 memcpy(&diu_shared_fb.ad0, vaddr, sizeof(struct diu_ad)); in mpc512x_init_diu() 261 pix_fmt = in_le32(vaddr); in mpc512x_init_diu() 263 diu_shared_fb.fb_phys = in_le32(vaddr + 4); in mpc512x_init_diu() 266 iounmap(vaddr); in mpc512x_init_diu() 269 vaddr = ioremap(desc, sizeof(diu_shared_fb.gamma)); in mpc512x_init_diu() 270 if (!vaddr) { in mpc512x_init_diu() 275 memcpy(&diu_shared_fb.gamma, vaddr, sizeof(diu_shared_fb.gamma)); in mpc512x_init_diu() [all …]
|
/linux-4.1.27/arch/arm/probes/uprobes/ |
D | core.c | 33 unsigned long vaddr) in set_swbp() argument 35 return uprobe_write_opcode(mm, vaddr, in set_swbp() 116 void arch_uprobe_copy_ixol(struct page *page, unsigned long vaddr, in arch_uprobe_copy_ixol() argument 120 void *dst = xol_page_kaddr + (vaddr & ~PAGE_MASK); in arch_uprobe_copy_ixol() 128 flush_uprobe_xol_access(page, vaddr, dst, len); in arch_uprobe_copy_ixol() 157 regs->ARM_pc = utask->vaddr + 4; in arch_uprobe_post_xol() 178 instruction_pointer_set(regs, utask->vaddr); in arch_uprobe_abort_xol()
|
/linux-4.1.27/arch/x86/include/asm/ |
D | page.h | 24 static inline void clear_user_page(void *page, unsigned long vaddr, in clear_user_page() argument 30 static inline void copy_user_page(void *to, void *from, unsigned long vaddr, in copy_user_page() argument 36 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument 37 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
|
D | dma-mapping.h | 68 void *vaddr, dma_addr_t dma_addr, 97 dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument 161 void *vaddr, dma_addr_t bus, in dma_free_attrs() argument 168 if (dma_release_from_coherent(dev, get_order(size), vaddr)) in dma_free_attrs() 171 debug_dma_free_coherent(dev, size, vaddr, bus); in dma_free_attrs() 173 ops->free(dev, size, vaddr, bus, attrs); in dma_free_attrs()
|
D | pgtable_32.h | 60 #define kpte_clear_flush(ptep, vaddr) \ argument 62 pte_clear(&init_mm, (vaddr), (ptep)); \ 63 __flush_tlb_one((vaddr)); \
|
/linux-4.1.27/drivers/acpi/apei/ |
D | ghes.c | 153 unsigned long vaddr; in ghes_ioremap_pfn_nmi() local 155 vaddr = (unsigned long)GHES_IOREMAP_NMI_PAGE(ghes_ioremap_area->addr); in ghes_ioremap_pfn_nmi() 156 ioremap_page_range(vaddr, vaddr + PAGE_SIZE, in ghes_ioremap_pfn_nmi() 159 return (void __iomem *)vaddr; in ghes_ioremap_pfn_nmi() 164 unsigned long vaddr; in ghes_ioremap_pfn_irq() local 166 vaddr = (unsigned long)GHES_IOREMAP_IRQ_PAGE(ghes_ioremap_area->addr); in ghes_ioremap_pfn_irq() 167 ioremap_page_range(vaddr, vaddr + PAGE_SIZE, in ghes_ioremap_pfn_irq() 170 return (void __iomem *)vaddr; in ghes_ioremap_pfn_irq() 175 unsigned long vaddr = (unsigned long __force)vaddr_ptr; in ghes_iounmap_nmi() local 178 BUG_ON(vaddr != (unsigned long)GHES_IOREMAP_NMI_PAGE(base)); in ghes_iounmap_nmi() [all …]
|
/linux-4.1.27/arch/alpha/mm/ |
D | init.c | 193 unsigned long vaddr; in callback_init() local 205 vaddr = (unsigned long)console_remap_vm.addr; in callback_init() 211 crb->map[i].va = vaddr; in callback_init() 216 if (pmd != pmd_offset(pgd, vaddr)) { in callback_init() 218 pmd = pmd_offset(pgd, vaddr); in callback_init() 222 set_pte(pte_offset_kernel(pmd, vaddr), in callback_init() 225 vaddr += PAGE_SIZE; in callback_init()
|
/linux-4.1.27/arch/x86/xen/ |
D | grant-table.c | 121 void *vaddr; in xlated_setup_gnttab_pages() local 147 vaddr = vmap(pages, nr_grant_frames, 0, PAGE_KERNEL); in xlated_setup_gnttab_pages() 148 if (!vaddr) { in xlated_setup_gnttab_pages() 160 xen_auto_xlat_grant_frames.vaddr = vaddr; in xlated_setup_gnttab_pages()
|
D | mmu.c | 126 unsigned long arbitrary_virt_to_mfn(void *vaddr) in arbitrary_virt_to_mfn() argument 128 xmaddr_t maddr = arbitrary_virt_to_machine(vaddr); in arbitrary_virt_to_mfn() 133 xmaddr_t arbitrary_virt_to_machine(void *vaddr) in arbitrary_virt_to_machine() argument 135 unsigned long address = (unsigned long)vaddr; in arbitrary_virt_to_machine() 144 if (virt_addr_valid(vaddr)) in arbitrary_virt_to_machine() 145 return virt_to_machine(vaddr); in arbitrary_virt_to_machine() 156 void make_lowmem_page_readonly(void *vaddr) in make_lowmem_page_readonly() argument 159 unsigned long address = (unsigned long)vaddr; in make_lowmem_page_readonly() 172 void make_lowmem_page_readwrite(void *vaddr) in make_lowmem_page_readwrite() argument 175 unsigned long address = (unsigned long)vaddr; in make_lowmem_page_readwrite() [all …]
|
/linux-4.1.27/drivers/i2c/busses/ |
D | i2c-ibm_iic.c | 85 volatile struct iic_regs __iomem *iic = dev->vaddr; in dump_iic_regs() 129 out_8(&dev->vaddr->intmsk, enable ? INTRMSK_EIMTC : 0); in iic_interrupt_mode() 137 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_dev_init() 182 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_dev_reset() 243 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_smbus_quick() 329 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_handler() 347 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_xfer_result() 380 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_abort_xfer() 412 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_wait_for_tc() 463 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_xfer_bytes() [all …]
|
/linux-4.1.27/drivers/video/fbdev/matrox/ |
D | matroxfb_base.h | 128 void __iomem* vaddr; member 132 return readb(va.vaddr + offs); in mga_readb() 136 writeb(value, va.vaddr + offs); in mga_writeb() 140 writew(value, va.vaddr + offs); in mga_writew() 144 return readl(va.vaddr + offs); in mga_readl() 148 writel(value, va.vaddr + offs); in mga_writel() 160 iowrite32_rep(va.vaddr, src, len >> 2); in mga_memcpy_toio() 162 u_int32_t __iomem* addr = va.vaddr; in mga_memcpy_toio() 183 va->vaddr += offs; in vaddr_add() 187 return va.vaddr; in vaddr_va() [all …]
|
/linux-4.1.27/include/asm-generic/ |
D | fixmap.h | 35 static inline unsigned long virt_to_fix(const unsigned long vaddr) in virt_to_fix() argument 37 BUG_ON(vaddr >= FIXADDR_TOP || vaddr < FIXADDR_START); in virt_to_fix() 38 return __virt_to_fix(vaddr); in virt_to_fix()
|
D | dma-coherent.h | 11 int dma_release_from_coherent(struct device *dev, int order, void *vaddr); 28 #define dma_release_from_coherent(dev, order, vaddr) (0) argument 29 #define dma_mmap_from_coherent(dev, vma, vaddr, order, ret) (0) argument
|
D | cacheflush.h | 26 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 29 flush_icache_user_range(vma, page, vaddr, len); \ 31 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
/linux-4.1.27/drivers/media/platform/coda/ |
D | coda-jpeg.c | 172 coda_memcpy_parabuf(ctx->parabuf.vaddr, huff + i); in coda_jpeg_write_tables() 176 coda_memcpy_parabuf(ctx->parabuf.vaddr, qmat + i); in coda_jpeg_write_tables() 183 void *vaddr = vb2_plane_vaddr(vb, 0); in coda_jpeg_check_buffer() local 184 u16 soi = be16_to_cpup((__be16 *)vaddr); in coda_jpeg_check_buffer() 185 u16 eoi = be16_to_cpup((__be16 *)(vaddr + in coda_jpeg_check_buffer()
|
/linux-4.1.27/drivers/net/wireless/orinoco/ |
D | airport.c | 29 void __iomem *vaddr; member 101 if (card->vaddr) in airport_detach() 102 iounmap(card->vaddr); in airport_detach() 103 card->vaddr = NULL; in airport_detach() 184 card->vaddr = ioremap(phys_addr, AIRPORT_IO_LEN); in airport_attach() 185 if (!card->vaddr) { in airport_attach() 190 hermes_struct_init(hw, card->vaddr, HERMES_16BIT_REGSPACING); in airport_attach()
|
/linux-4.1.27/arch/cris/include/asm/ |
D | page.h | 15 #define clear_user_page(page, vaddr, pg) clear_page(page) argument 16 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument 18 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument 19 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
|
/linux-4.1.27/drivers/vfio/ |
D | vfio_iommu_type1.c | 75 unsigned long vaddr; /* Process virtual addr */ member 231 static int vaddr_get_pfn(unsigned long vaddr, int prot, unsigned long *pfn) in vaddr_get_pfn() argument 237 if (get_user_pages_fast(vaddr, 1, !!(prot & IOMMU_WRITE), page) == 1) { in vaddr_get_pfn() 244 vma = find_vma_intersection(current->mm, vaddr, vaddr + 1); in vaddr_get_pfn() 247 *pfn = ((vaddr - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; in vaddr_get_pfn() 262 static long vfio_pin_pages(unsigned long vaddr, long npage, in vfio_pin_pages() argument 273 ret = vaddr_get_pfn(vaddr, prot, pfn_base); in vfio_pin_pages() 293 for (i = 1, vaddr += PAGE_SIZE; i < npage; i++, vaddr += PAGE_SIZE) { in vfio_pin_pages() 296 ret = vaddr_get_pfn(vaddr, prot, &pfn); in vfio_pin_pages() 552 unsigned long vaddr = map->vaddr; in vfio_dma_do_map() local [all …]
|
/linux-4.1.27/drivers/gpu/drm/ |
D | drm_gem_cma_helper.c | 110 cma_obj->vaddr = dma_alloc_writecombine(drm->dev, size, in drm_gem_cma_create() 112 if (!cma_obj->vaddr) { in drm_gem_cma_create() 192 if (cma_obj->vaddr) { in drm_gem_cma_free_object() 194 cma_obj->vaddr, cma_obj->paddr); in drm_gem_cma_free_object() 331 cma_obj->vaddr, cma_obj->paddr, in drm_gem_cma_mmap_obj() 393 off, &cma_obj->paddr, cma_obj->vaddr, obj->size); in drm_gem_cma_describe() 422 ret = dma_get_sgtable(obj->dev->dev, sgt, cma_obj->vaddr, in drm_gem_cma_prime_get_sg_table() 524 return cma_obj->vaddr; in drm_gem_cma_prime_vmap() 539 void drm_gem_cma_prime_vunmap(struct drm_gem_object *obj, void *vaddr) in drm_gem_cma_prime_vunmap() argument
|
D | drm_pci.c | 60 dmah->vaddr = dma_alloc_coherent(&dev->pdev->dev, size, &dmah->busaddr, GFP_KERNEL | __GFP_COMP); in drm_pci_alloc() 62 if (dmah->vaddr == NULL) { in drm_pci_alloc() 67 memset(dmah->vaddr, 0, size); in drm_pci_alloc() 71 for (addr = (unsigned long)dmah->vaddr, sz = size; in drm_pci_alloc() 91 if (dmah->vaddr) { in __drm_legacy_pci_free() 94 for (addr = (unsigned long)dmah->vaddr, sz = dmah->size; in __drm_legacy_pci_free() 98 dma_free_coherent(&dev->pdev->dev, dmah->size, dmah->vaddr, in __drm_legacy_pci_free()
|
/linux-4.1.27/drivers/infiniband/hw/qib/ |
D | qib_keys.c | 180 isge->vaddr = (void *) sge->addr; in qib_lkey_ok() 227 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in qib_lkey_ok() 253 u32 len, u64 vaddr, u32 rkey, int acc) in qib_rkey_ok() argument 279 sge->vaddr = (void *) vaddr; in qib_rkey_ok() 292 off = vaddr - mr->iova; in qib_rkey_ok() 293 if (unlikely(vaddr < mr->iova || off + len > mr->length || in qib_rkey_ok() 326 sge->vaddr = mr->map[m]->segs[n].vaddr + off; in qib_rkey_ok() 379 mr->map[m]->segs[n].vaddr = (void *) page_list[i]; in qib_fast_reg_mr()
|
/linux-4.1.27/arch/avr32/mm/ |
D | dma-coherent.c | 16 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, int direction) in dma_cache_sync() argument 21 if (PXSEG(vaddr) == P2SEG) in dma_cache_sync() 26 invalidate_dcache_region(vaddr, size); in dma_cache_sync() 29 clean_dcache_region(vaddr, size); in dma_cache_sync() 32 flush_dcache_region(vaddr, size); in dma_cache_sync()
|
/linux-4.1.27/arch/m32r/include/asm/ |
D | page.h | 16 #define clear_user_page(page, vaddr, pg) clear_page(page) argument 17 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument 19 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument 20 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
|
/linux-4.1.27/arch/cris/arch-v32/drivers/pci/ |
D | dma.c | 43 void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument 47 if (!dma_release_from_coherent(dev, order, vaddr)) in dma_free_coherent() 48 free_pages((unsigned long)vaddr, order); in dma_free_coherent()
|
/linux-4.1.27/arch/alpha/boot/ |
D | bootpz.c | 79 unsigned long vaddr, kaddr; in check_range() local 86 for (vaddr = vstart; vaddr <= vend; vaddr += PAGE_SIZE) in check_range() 88 kaddr = (find_pa(vaddr) | PAGE_OFFSET); in check_range() 94 vaddr, kaddr, kstart, kend); in check_range()
|
/linux-4.1.27/fs/minix/ |
D | minix.h | 124 static inline int minix_find_first_zero_bit(const void *vaddr, unsigned size) in minix_find_first_zero_bit() argument 126 const unsigned short *p = vaddr, *addr = vaddr; in minix_find_first_zero_bit() 149 static inline int minix_test_bit(int nr, const void *vaddr) in minix_test_bit() argument 151 const unsigned short *p = vaddr; in minix_test_bit()
|
/linux-4.1.27/arch/alpha/kernel/ |
D | core_irongate.c | 319 unsigned long vaddr; in irongate_ioremap() local 385 for(baddr = addr, vaddr = (unsigned long)area->addr; in irongate_ioremap() 387 baddr += PAGE_SIZE, vaddr += PAGE_SIZE) in irongate_ioremap() 392 if (__alpha_remap_area_pages(vaddr, in irongate_ioremap() 402 vaddr = (unsigned long)area->addr + (addr & ~PAGE_MASK); in irongate_ioremap() 405 addr, size, vaddr); in irongate_ioremap() 407 return (void __iomem *)vaddr; in irongate_ioremap()
|
/linux-4.1.27/arch/ia64/include/asm/ |
D | cacheflush.h | 47 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 49 flush_icache_user_range(vma, page, vaddr, len); \ 51 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
/linux-4.1.27/arch/powerpc/include/asm/ |
D | cacheflush.h | 55 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 58 flush_icache_user_range(vma, page, vaddr, len); \ 60 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
D | highmem.h | 89 unsigned long idx, vaddr = (unsigned long) ptr; in kmap_atomic_to_page() local 92 if (vaddr < FIXADDR_START) in kmap_atomic_to_page() 95 idx = virt_to_fix(vaddr); in kmap_atomic_to_page()
|
D | dma-mapping.h | 28 void *vaddr, dma_addr_t dma_handle, 46 extern void __dma_free_coherent(size_t size, void *vaddr); 47 extern void __dma_sync(void *vaddr, size_t size, int direction); 218 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument 222 __dma_sync(vaddr, size, (int)direction); in dma_cache_sync()
|
/linux-4.1.27/drivers/tty/serial/8250/ |
D | 8250_acorn.c | 38 void __iomem *vaddr; member 57 info->vaddr = ecardm_iomap(ec, type->type, 0, 0); in serial_card_probe() 58 if (!info->vaddr) { in serial_card_probe() 74 uart.port.membase = info->vaddr + type->offset[i]; in serial_card_probe()
|
/linux-4.1.27/drivers/scsi/sym53c8xx_2/ |
D | sym_malloc.c | 226 void *vaddr; in ___get_dma_mem_cluster() local 232 vaddr = sym_m_get_dma_mem_cluster(mp, vbp); in ___get_dma_mem_cluster() 233 if (vaddr) { in ___get_dma_mem_cluster() 234 int hc = VTOB_HASH_CODE(vaddr); in ___get_dma_mem_cluster() 239 return vaddr; in ___get_dma_mem_cluster() 252 while (*vbpp && (*vbpp)->vaddr != m) in ___free_dma_mem_cluster() 370 while (vp && vp->vaddr != a) in __vtobus()
|
/linux-4.1.27/arch/alpha/include/asm/ |
D | cacheflush.h | 70 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 72 flush_icache_user_range(vma, page, vaddr, len); \ 74 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
/linux-4.1.27/drivers/media/pci/cx23885/ |
D | cx23885-alsa.c | 89 buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT); in cx23885_alsa_dma_init() 90 if (NULL == buf->vaddr) { in cx23885_alsa_dma_init() 96 (unsigned long)buf->vaddr, in cx23885_alsa_dma_init() 99 memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT); in cx23885_alsa_dma_init() 108 pg = vmalloc_to_page(buf->vaddr + i * PAGE_SIZE); in cx23885_alsa_dma_init() 119 vfree(buf->vaddr); in cx23885_alsa_dma_init() 120 buf->vaddr = NULL; in cx23885_alsa_dma_init() 154 vfree(buf->vaddr); in cx23885_alsa_dma_free() 155 buf->vaddr = NULL; in cx23885_alsa_dma_free() 416 substream->runtime->dma_area = chip->buf->vaddr; in snd_cx23885_hw_params()
|
/linux-4.1.27/drivers/parisc/ |
D | iommu-helpers.h | 30 unsigned long vaddr; in iommu_fill_pdir() local 69 vaddr = (unsigned long)sg_virt(startsg); in iommu_fill_pdir() 78 vaddr, hint); in iommu_fill_pdir() 79 vaddr += IOVP_SIZE; in iommu_fill_pdir()
|
/linux-4.1.27/crypto/ |
D | scatterwalk.c | 81 u8 *vaddr; in scatterwalk_copychunks() local 86 vaddr = scatterwalk_map(walk); in scatterwalk_copychunks() 87 memcpy_dir(buf, vaddr, len_this_page, out); in scatterwalk_copychunks() 88 scatterwalk_unmap(vaddr); in scatterwalk_copychunks()
|
/linux-4.1.27/arch/mips/netlogic/common/ |
D | nlm-dma.c | 70 void *vaddr, dma_addr_t dma_handle, struct dma_attrs *attrs) in nlm_dma_free_coherent() argument 74 if (dma_release_from_coherent(dev, order, vaddr)) in nlm_dma_free_coherent() 77 swiotlb_free_coherent(dev, size, vaddr, dma_handle); in nlm_dma_free_coherent()
|
/linux-4.1.27/arch/unicore32/include/asm/ |
D | cacheflush.h | 19 #define CACHE_COLOUR(vaddr) ((vaddr & (SHMLBA - 1)) >> PAGE_SHIFT) argument 120 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
/linux-4.1.27/fs/pstore/ |
D | ram_core.c | 391 void *vaddr; in persistent_ram_vmap() local 412 vaddr = vmap(pages, page_count, VM_MAP, prot); in persistent_ram_vmap() 415 return vaddr; in persistent_ram_vmap() 447 prz->vaddr = persistent_ram_vmap(start, size, memtype); in persistent_ram_buffer_map() 449 prz->vaddr = persistent_ram_iomap(start, size, memtype); in persistent_ram_buffer_map() 451 if (!prz->vaddr) { in persistent_ram_buffer_map() 457 prz->buffer = prz->vaddr + offset_in_page(start); in persistent_ram_buffer_map() 501 if (prz->vaddr) { in persistent_ram_free() 503 vunmap(prz->vaddr); in persistent_ram_free() 505 iounmap(prz->vaddr); in persistent_ram_free() [all …]
|
/linux-4.1.27/kernel/trace/ |
D | trace_uprobe.c | 33 unsigned long vaddr[]; member 127 void __user *vaddr = (void __force __user *) addr; \ 129 if (copy_from_user(&retval, vaddr, sizeof(type))) \ 165 void __user *vaddr = (void __force __user *) addr; in FETCH_FUNC_NAME() local 167 len = strnlen_user(vaddr, MAX_STRING_SIZE); in FETCH_FUNC_NAME() 180 udd = (void *) current->utask->vaddr; in translate_user_vaddr() 190 void *vaddr = (void *)translate_user_vaddr(offset); \ 192 FETCH_FUNC_NAME(memory, type)(regs, vaddr, dest); \ 799 entry->vaddr[0] = func; in __uprobe_trace_func() 800 entry->vaddr[1] = instruction_pointer(regs); in __uprobe_trace_func() [all …]
|
/linux-4.1.27/arch/m32r/mm/ |
D | fault.c | 348 void update_mmu_cache(struct vm_area_struct *vma, unsigned long vaddr, in update_mmu_cache() argument 363 vaddr = (vaddr & PAGE_MASK) | get_asid(); in update_mmu_cache() 370 if (*entry1++ == vaddr) { in update_mmu_cache() 378 if (*entry2++ == vaddr) { in update_mmu_cache() 405 : "r" (vaddr), "r" (pte_data), "i" (MMU_REG_BASE), in update_mmu_cache() 449 *entry1++ = vaddr; /* Set TLB tag */ in update_mmu_cache()
|
/linux-4.1.27/drivers/gpio/ |
D | gpio-ts5500.c | 64 #define TS5500_DIO_IN_OUT(vaddr, vbit, caddr, cbit) \ argument 66 .value_addr = vaddr, \ 99 #define TS5500_DIO_GROUP(vaddr, vbitfrom, caddr, cbit) \ argument 100 TS5500_DIO_IN_OUT(vaddr, vbitfrom + 0, caddr, cbit), \ 101 TS5500_DIO_IN_OUT(vaddr, vbitfrom + 1, caddr, cbit), \ 102 TS5500_DIO_IN_OUT(vaddr, vbitfrom + 2, caddr, cbit), \ 103 TS5500_DIO_IN_OUT(vaddr, vbitfrom + 3, caddr, cbit)
|
/linux-4.1.27/arch/arm/plat-samsung/ |
D | pm-debug.c | 59 unsigned long vaddr; in s3c_pm_uart_base() local 61 debug_ll_addr(&paddr, &vaddr); in s3c_pm_uart_base() 63 return (void __iomem *)vaddr; in s3c_pm_uart_base()
|
/linux-4.1.27/arch/arm64/mm/ |
D | copypage.c | 25 void __cpu_copy_user_page(void *kto, const void *kfrom, unsigned long vaddr) in __cpu_copy_user_page() argument 32 void __cpu_clear_user_page(void *kaddr, unsigned long vaddr) in __cpu_clear_user_page() argument
|
/linux-4.1.27/drivers/media/platform/exynos4-is/ |
D | fimc-is.c | 241 buf = is->memory.vaddr + is->setfile.base; in fimc_is_load_setfile() 246 pr_debug("mem vaddr: %p, setfile buf: %p\n", is->memory.vaddr, buf); in fimc_is_load_setfile() 317 memcpy(is->memory.vaddr, is->fw.f_w->data, is->fw.f_w->size); in fimc_is_start_firmware() 337 is->memory.vaddr = dma_alloc_coherent(dev, FIMC_IS_CPU_MEM_SIZE, in fimc_is_alloc_cpu_memory() 339 if (is->memory.vaddr == NULL) in fimc_is_alloc_cpu_memory() 343 memset(is->memory.vaddr, 0, is->memory.size); in fimc_is_alloc_cpu_memory() 350 dma_free_coherent(dev, is->memory.size, is->memory.vaddr, in fimc_is_alloc_cpu_memory() 355 is->is_p_region = (struct is_region *)(is->memory.vaddr + in fimc_is_alloc_cpu_memory() 361 is->is_shared_region = (struct is_share_region *)(is->memory.vaddr + in fimc_is_alloc_cpu_memory() 370 if (is->memory.vaddr == NULL) in fimc_is_free_cpu_memory() [all …]
|