| /linux-4.4.14/arch/m68k/include/asm/ |
| D | bitops.h | 31 static inline void bset_reg_set_bit(int nr, volatile unsigned long *vaddr) in bset_reg_set_bit() argument 33 char *p = (char *)vaddr + (nr ^ 31) / 8; in bset_reg_set_bit() 41 static inline void bset_mem_set_bit(int nr, volatile unsigned long *vaddr) in bset_mem_set_bit() argument 43 char *p = (char *)vaddr + (nr ^ 31) / 8; in bset_mem_set_bit() 50 static inline void bfset_mem_set_bit(int nr, volatile unsigned long *vaddr) in bfset_mem_set_bit() argument 54 : "d" (nr ^ 31), "o" (*vaddr) in bfset_mem_set_bit() 59 #define set_bit(nr, vaddr) bset_reg_set_bit(nr, vaddr) argument 61 #define set_bit(nr, vaddr) bset_mem_set_bit(nr, vaddr) argument 63 #define set_bit(nr, vaddr) (__builtin_constant_p(nr) ? \ argument 64 bset_mem_set_bit(nr, vaddr) : \ [all …]
|
| D | page_mm.h | 9 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument 58 #define clear_user_page(addr, vaddr, page) \ argument 62 #define copy_user_page(to, from, vaddr, page) \ argument 73 static inline unsigned long ___pa(void *vaddr) in ___pa() argument 80 : "0" (vaddr), "i" (m68k_fixup_memoffset)); in ___pa() 83 #define __pa(vaddr) ___pa((void *)(long)(vaddr)) argument 86 void *vaddr; in __va() local 90 : "=r" (vaddr) in __va() 92 return vaddr; in __va()
|
| D | page_no.h | 9 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument 15 #define clear_user_page(page, vaddr, pg) clear_page(page) argument 16 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument 18 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument 19 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr) 22 #define __pa(vaddr) ((unsigned long)(vaddr)) argument
|
| D | cacheflush_mm.h | 155 extern void cache_push_v(unsigned long vaddr, int len); 222 static inline void __flush_page_to_ram(void *vaddr) in __flush_page_to_ram() argument 226 addr = ((unsigned long) vaddr) & ~(PAGE_SIZE - 1); in __flush_page_to_ram() 239 : : "a" (__pa(vaddr))); in __flush_page_to_ram() 261 struct page *page, unsigned long vaddr, in copy_to_user_page() argument 264 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page() 266 flush_icache_user_range(vma, page, vaddr, len); in copy_to_user_page() 269 struct page *page, unsigned long vaddr, in copy_from_user_page() argument 272 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_from_user_page()
|
| /linux-4.4.14/arch/arm/mm/ |
| D | cache-xsc3l2.c | 100 unsigned long vaddr; in xsc3_l2_inv_range() local 107 vaddr = -1; /* to force the first mapping */ in xsc3_l2_inv_range() 113 vaddr = l2_map_va(start & ~(CACHE_LINE_SIZE - 1), vaddr); in xsc3_l2_inv_range() 114 xsc3_l2_clean_mva(vaddr); in xsc3_l2_inv_range() 115 xsc3_l2_inv_mva(vaddr); in xsc3_l2_inv_range() 123 vaddr = l2_map_va(start, vaddr); in xsc3_l2_inv_range() 124 xsc3_l2_inv_mva(vaddr); in xsc3_l2_inv_range() 132 vaddr = l2_map_va(start, vaddr); in xsc3_l2_inv_range() 133 xsc3_l2_clean_mva(vaddr); in xsc3_l2_inv_range() 134 xsc3_l2_inv_mva(vaddr); in xsc3_l2_inv_range() [all …]
|
| D | highmem.c | 23 unsigned long vaddr = __fix_to_virt(idx); in set_fixmap_pte() local 24 pte_t *ptep = pte_offset_kernel(pmd_off_k(vaddr), vaddr); in set_fixmap_pte() 27 local_flush_tlb_kernel_page(vaddr); in set_fixmap_pte() 30 static inline pte_t get_fixmap_pte(unsigned long vaddr) in get_fixmap_pte() argument 32 pte_t *ptep = pte_offset_kernel(pmd_off_k(vaddr), vaddr); in get_fixmap_pte() 58 unsigned long vaddr; in kmap_atomic() local 83 vaddr = __fix_to_virt(idx); in kmap_atomic() 89 BUG_ON(!pte_none(get_fixmap_pte(vaddr))); in kmap_atomic() 98 return (void *)vaddr; in kmap_atomic() 104 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local [all …]
|
| D | copypage-v6.c | 34 struct page *from, unsigned long vaddr, struct vm_area_struct *vma) in v6_copy_user_highpage_nonaliasing() argument 49 static void v6_clear_user_highpage_nonaliasing(struct page *page, unsigned long vaddr) in v6_clear_user_highpage_nonaliasing() argument 73 struct page *from, unsigned long vaddr, struct vm_area_struct *vma) in v6_copy_user_highpage_aliasing() argument 75 unsigned int offset = CACHE_COLOUR(vaddr); in v6_copy_user_highpage_aliasing() 106 static void v6_clear_user_highpage_aliasing(struct page *page, unsigned long vaddr) in v6_clear_user_highpage_aliasing() argument 108 unsigned long to = COPYPAGE_V6_TO + (CACHE_COLOUR(vaddr) << PAGE_SHIFT); in v6_clear_user_highpage_aliasing()
|
| D | copypage-xsc3.c | 74 unsigned long vaddr, struct vm_area_struct *vma) in xsc3_mc_copy_user_highpage() argument 80 flush_cache_page(vma, vaddr, page_to_pfn(from)); in xsc3_mc_copy_user_highpage() 91 void xsc3_mc_clear_user_highpage(struct page *page, unsigned long vaddr) in xsc3_mc_clear_user_highpage() argument
|
| D | copypage-feroceon.c | 71 unsigned long vaddr, struct vm_area_struct *vma) in feroceon_copy_user_highpage() argument 77 flush_cache_page(vma, vaddr, page_to_pfn(from)); in feroceon_copy_user_highpage() 83 void feroceon_clear_user_highpage(struct page *page, unsigned long vaddr) in feroceon_clear_user_highpage() argument
|
| D | copypage-v4wb.c | 51 unsigned long vaddr, struct vm_area_struct *vma) in v4wb_copy_user_highpage() argument 57 flush_cache_page(vma, vaddr, page_to_pfn(from)); in v4wb_copy_user_highpage() 68 void v4wb_clear_user_highpage(struct page *page, unsigned long vaddr) in v4wb_clear_user_highpage() argument
|
| D | flush.c | 41 static void flush_pfn_alias(unsigned long pfn, unsigned long vaddr) in flush_pfn_alias() argument 43 unsigned long to = FLUSH_ALIAS_START + (CACHE_COLOUR(vaddr) << PAGE_SHIFT); in flush_pfn_alias() 55 static void flush_icache_alias(unsigned long pfn, unsigned long vaddr, unsigned long len) in flush_icache_alias() argument 57 unsigned long va = FLUSH_ALIAS_START + (CACHE_COLOUR(vaddr) << PAGE_SHIFT); in flush_icache_alias() 58 unsigned long offset = vaddr & (PAGE_SIZE - 1); in flush_icache_alias() 118 #define flush_pfn_alias(pfn,vaddr) do { } while (0) argument 119 #define flush_icache_alias(pfn,vaddr,len) do { } while (0) argument
|
| /linux-4.4.14/arch/parisc/kernel/ |
| D | pci-dma.c | 83 unsigned long vaddr, in map_pte_uncached() argument 87 unsigned long orig_vaddr = vaddr; in map_pte_uncached() 89 vaddr &= ~PMD_MASK; in map_pte_uncached() 90 end = vaddr + size; in map_pte_uncached() 102 vaddr += PAGE_SIZE; in map_pte_uncached() 106 } while (vaddr < end); in map_pte_uncached() 110 static inline int map_pmd_uncached(pmd_t * pmd, unsigned long vaddr, in map_pmd_uncached() argument 114 unsigned long orig_vaddr = vaddr; in map_pmd_uncached() 116 vaddr &= ~PGDIR_MASK; in map_pmd_uncached() 117 end = vaddr + size; in map_pmd_uncached() [all …]
|
| /linux-4.4.14/arch/sh/mm/ |
| D | kmap.c | 18 #define kmap_get_fixmap_pte(vaddr) \ argument 19 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), (vaddr)) 25 unsigned long vaddr; in kmap_coherent_init() local 28 vaddr = __fix_to_virt(FIX_CMAP_BEGIN); in kmap_coherent_init() 29 kmap_coherent_pte = kmap_get_fixmap_pte(vaddr); in kmap_coherent_init() 35 unsigned long vaddr; in kmap_coherent() local 46 vaddr = __fix_to_virt(idx); in kmap_coherent() 51 return (void *)vaddr; in kmap_coherent() 57 unsigned long vaddr = (unsigned long)kvaddr & PAGE_MASK; in kunmap_coherent() local 58 enum fixed_addresses idx = __virt_to_fix(vaddr); in kunmap_coherent() [all …]
|
| D | init.c | 153 unsigned long vaddr, pte_t *lastpte) in page_table_kmap_check() argument 166 unsigned long vaddr; in page_table_range_init() local 168 vaddr = start; in page_table_range_init() 169 i = __pgd_offset(vaddr); in page_table_range_init() 170 j = __pud_offset(vaddr); in page_table_range_init() 171 k = __pmd_offset(vaddr); in page_table_range_init() 174 for ( ; (i < PTRS_PER_PGD) && (vaddr != end); pgd++, i++) { in page_table_range_init() 176 for ( ; (j < PTRS_PER_PUD) && (vaddr != end); pud++, j++) { in page_table_range_init() 181 for (; (k < PTRS_PER_PMD) && (vaddr != end); pmd++, k++) { in page_table_range_init() 183 pmd, vaddr, pte); in page_table_range_init() [all …]
|
| D | cache.c | 59 unsigned long vaddr, void *dst, const void *src, in copy_to_user_page() argument 64 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page() 74 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page() 78 unsigned long vaddr, void *dst, const void *src, in copy_from_user_page() argument 83 void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_from_user_page() 94 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument 102 vfrom = kmap_coherent(from, vaddr); in copy_user_highpage() 111 if (pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK) || in copy_user_highpage() 121 void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument 127 if (pages_do_alias((unsigned long)kaddr, vaddr & PAGE_MASK)) in clear_user_highpage()
|
| D | pmb.c | 137 static bool pmb_mapping_exists(unsigned long vaddr, phys_addr_t phys, in pmb_mapping_exists() argument 156 if ((vaddr < pmbe->vpn) || (vaddr >= (pmbe->vpn + pmbe->size))) in pmb_mapping_exists() 335 int pmb_bolt_mapping(unsigned long vaddr, phys_addr_t phys, in pmb_bolt_mapping() argument 345 if (!pmb_addr_valid(vaddr, size)) in pmb_bolt_mapping() 347 if (pmb_mapping_exists(vaddr, phys, size)) in pmb_bolt_mapping() 350 orig_addr = vaddr; in pmb_bolt_mapping() 353 flush_tlb_kernel_range(vaddr, vaddr + size); in pmb_bolt_mapping() 363 pmbe = pmb_alloc(vaddr, phys, pmb_flags | in pmb_bolt_mapping() 377 vaddr += pmbe->size; in pmb_bolt_mapping() 413 unsigned long vaddr; in pmb_remap_caller() local [all …]
|
| D | cache-sh4.c | 215 void *vaddr; in sh4_flush_cache_page() local 236 vaddr = NULL; in sh4_flush_cache_page() 246 vaddr = kmap_coherent(page, address); in sh4_flush_cache_page() 248 vaddr = kmap_atomic(page); in sh4_flush_cache_page() 250 address = (unsigned long)vaddr; in sh4_flush_cache_page() 259 if (vaddr) { in sh4_flush_cache_page() 261 kunmap_coherent(vaddr); in sh4_flush_cache_page() 263 kunmap_atomic(vaddr); in sh4_flush_cache_page()
|
| D | consistent.c | 68 void *vaddr, dma_addr_t dma_handle, in dma_generic_free_coherent() argument 78 iounmap(vaddr); in dma_generic_free_coherent() 81 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument 87 (void *)CAC_ADDR((unsigned long)vaddr) : vaddr; in dma_cache_sync()
|
| D | ioremap.c | 108 unsigned long vaddr = (unsigned long __force)addr; in __iounmap() local 114 if (iomapping_nontranslatable(vaddr)) in __iounmap() 129 p = remove_vm_area((void *)(vaddr & PAGE_MASK)); in __iounmap()
|
| /linux-4.4.14/arch/microblaze/mm/ |
| D | consistent.c | 64 unsigned long order, vaddr; in consistent_alloc() local 82 vaddr = __get_free_pages(gfp, order); in consistent_alloc() 83 if (!vaddr) in consistent_alloc() 90 flush_dcache_range(virt_to_phys((void *)vaddr), in consistent_alloc() 91 virt_to_phys((void *)vaddr) + size); in consistent_alloc() 94 ret = (void *)vaddr; in consistent_alloc() 113 free_pages(vaddr, order); in consistent_alloc() 120 *dma_handle = pa = __virt_to_phys(vaddr); in consistent_alloc() 129 page = virt_to_page(vaddr); in consistent_alloc() 151 free_pages(vaddr, order); in consistent_alloc() [all …]
|
| D | highmem.c | 37 unsigned long vaddr; in kmap_atomic_prot() local 48 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot() 52 set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot)); in kmap_atomic_prot() 53 local_flush_tlb_page(NULL, vaddr); in kmap_atomic_prot() 55 return (void *) vaddr; in kmap_atomic_prot() 61 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 64 if (vaddr < __fix_to_virt(FIX_KMAP_END)) { in __kunmap_atomic() 76 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in __kunmap_atomic() 82 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic() 83 local_flush_tlb_page(NULL, vaddr); in __kunmap_atomic()
|
| /linux-4.4.14/arch/sparc/mm/ |
| D | io-unit.c | 94 static unsigned long iounit_get_area(struct iounit_struct *iounit, unsigned long vaddr, int size) in iounit_get_area() argument 100 npages = ((vaddr & ~PAGE_MASK) + size + (PAGE_SIZE-1)) >> PAGE_SHIFT; in iounit_get_area() 109 IOD(("iounit_get_area(%08lx,%d[%d])=", vaddr, size, npages)); in iounit_get_area() 124 panic("iounit_get_area: Couldn't find free iopte slots for (%08lx,%d)\n", vaddr, size); in iounit_get_area() 132 iopte = MKIOPTE(__pa(vaddr & PAGE_MASK)); in iounit_get_area() 133 vaddr = IOUNIT_DMA_BASE + (scan << PAGE_SHIFT) + (vaddr & ~PAGE_MASK); in iounit_get_area() 138 IOD(("%08lx\n", vaddr)); in iounit_get_area() 139 return vaddr; in iounit_get_area() 142 static __u32 iounit_get_scsi_one(struct device *dev, char *vaddr, unsigned long len) in iounit_get_scsi_one() argument 148 ret = iounit_get_area(iounit, (unsigned long)vaddr, len); in iounit_get_scsi_one() [all …]
|
| D | highmem.c | 53 unsigned long vaddr; in kmap_atomic() local 63 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic() 67 __flush_cache_one(vaddr); in kmap_atomic() 78 __flush_tlb_one(vaddr); in kmap_atomic() 83 return (void*) vaddr; in kmap_atomic() 89 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 92 if (vaddr < FIXADDR_START) { // FIXME in __kunmap_atomic() 105 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN+idx)); in __kunmap_atomic() 109 __flush_cache_one(vaddr); in __kunmap_atomic() 118 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic() [all …]
|
| D | tlb.c | 69 static void tlb_batch_add_one(struct mm_struct *mm, unsigned long vaddr, in tlb_batch_add_one() argument 75 vaddr &= PAGE_MASK; in tlb_batch_add_one() 77 vaddr |= 0x1UL; in tlb_batch_add_one() 87 flush_tsb_user_page(mm, vaddr, huge); in tlb_batch_add_one() 88 global_flush_tlb_page(mm, vaddr); in tlb_batch_add_one() 103 tb->vaddrs[nr] = vaddr; in tlb_batch_add_one() 112 void tlb_batch_add(struct mm_struct *mm, unsigned long vaddr, in tlb_batch_add() argument 136 if ((paddr ^ vaddr) & (1 << 13)) in tlb_batch_add() 142 tlb_batch_add_one(mm, vaddr, pte_exec(orig), huge); in tlb_batch_add() 146 static void tlb_batch_pmd_scan(struct mm_struct *mm, unsigned long vaddr, in tlb_batch_pmd_scan() argument [all …]
|
| D | leon_mm.c | 35 unsigned long leon_swprobe(unsigned long vaddr, unsigned long *paddr) in leon_swprobe() argument 88 ptr += ((((vaddr) >> LEON_PGD_SH) & LEON_PGD_M) * 4); in leon_swprobe() 111 ptr += (((vaddr >> LEON_PMD_SH) & LEON_PMD_M) * 4); in leon_swprobe() 139 ptr += (((vaddr >> LEON_PTE_SH) & LEON_PTE_M) * 4); in leon_swprobe() 160 (vaddr & ~(-1 << LEON_PTE_SH)) | ((pte & ~0xff) << 4); in leon_swprobe() 164 (vaddr & ~(-1 << LEON_PMD_SH)) | ((pte & ~0xff) << 4); in leon_swprobe() 168 (vaddr & ~(-1 << LEON_PGD_SH)) | ((pte & ~0xff) << 4); in leon_swprobe() 172 paddr_calc = vaddr; in leon_swprobe()
|
| D | iommu.c | 207 static u32 iommu_get_scsi_one(struct device *dev, char *vaddr, unsigned int len) in iommu_get_scsi_one() argument 214 off = (unsigned long)vaddr & ~PAGE_MASK; in iommu_get_scsi_one() 216 page = virt_to_page((unsigned long)vaddr & PAGE_MASK); in iommu_get_scsi_one() 221 static __u32 iommu_get_scsi_one_gflush(struct device *dev, char *vaddr, unsigned long len) in iommu_get_scsi_one_gflush() argument 224 return iommu_get_scsi_one(dev, vaddr, len); in iommu_get_scsi_one_gflush() 227 static __u32 iommu_get_scsi_one_pflush(struct device *dev, char *vaddr, unsigned long len) in iommu_get_scsi_one_pflush() argument 229 unsigned long page = ((unsigned long) vaddr) & PAGE_MASK; in iommu_get_scsi_one_pflush() 231 while(page < ((unsigned long)(vaddr + len))) { in iommu_get_scsi_one_pflush() 235 return iommu_get_scsi_one(dev, vaddr, len); in iommu_get_scsi_one_pflush() 299 static void iommu_release_scsi_one(struct device *dev, __u32 vaddr, unsigned long len) in iommu_release_scsi_one() argument [all …]
|
| /linux-4.4.14/arch/arm/mach-ixp4xx/include/mach/ |
| D | io.h | 98 const u8 *vaddr, int count) in __indirect_writesb() argument 101 writeb(*vaddr++, bus_addr); in __indirect_writesb() 121 const u16 *vaddr, int count) in __indirect_writesw() argument 124 writew(*vaddr++, bus_addr); in __indirect_writesw() 140 const u32 *vaddr, int count) in __indirect_writesl() argument 143 writel(*vaddr++, bus_addr); in __indirect_writesl() 163 u8 *vaddr, u32 count) in __indirect_readsb() argument 166 *vaddr++ = readb(bus_addr); in __indirect_readsb() 186 u16 *vaddr, u32 count) in __indirect_readsw() argument 189 *vaddr++ = readw(bus_addr); in __indirect_readsw() [all …]
|
| /linux-4.4.14/arch/m68k/sun3/ |
| D | dvma.c | 23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) in dvma_page() argument 36 if(ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] != pte) { in dvma_page() 37 sun3_put_pte(vaddr, pte); in dvma_page() 38 ptelist[(vaddr & 0xff000) >> PAGE_SHIFT] = pte; in dvma_page() 41 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page() 50 unsigned long vaddr; in dvma_map_iommu() local 52 vaddr = dvma_btov(baddr); in dvma_map_iommu() 54 end = vaddr + len; in dvma_map_iommu() 56 while(vaddr < end) { in dvma_map_iommu() 57 dvma_page(kaddr, vaddr); in dvma_map_iommu() [all …]
|
| D | mmu_emu.c | 117 void print_pte_vaddr (unsigned long vaddr) in print_pte_vaddr() argument 119 printk (" vaddr=%lx [%02lx]", vaddr, sun3_get_segmap (vaddr)); in print_pte_vaddr() 120 print_pte (__pte (sun3_get_pte (vaddr))); in print_pte_vaddr() 282 inline void mmu_emu_map_pmeg (int context, int vaddr) in mmu_emu_map_pmeg() argument 288 vaddr &= ~SUN3_PMEG_MASK; in mmu_emu_map_pmeg() 297 curr_pmeg, context, vaddr); in mmu_emu_map_pmeg() 309 if(vaddr >= PAGE_OFFSET) { in mmu_emu_map_pmeg() 315 sun3_put_segmap (vaddr, curr_pmeg); in mmu_emu_map_pmeg() 325 sun3_put_segmap (vaddr, curr_pmeg); in mmu_emu_map_pmeg() 328 pmeg_vaddr[curr_pmeg] = vaddr; in mmu_emu_map_pmeg() [all …]
|
| /linux-4.4.14/drivers/media/v4l2-core/ |
| D | videobuf2-vmalloc.c | 25 void *vaddr; member 46 buf->vaddr = vmalloc_user(buf->size); in vb2_vmalloc_alloc() 52 if (!buf->vaddr) { in vb2_vmalloc_alloc() 67 vfree(buf->vaddr); in vb2_vmalloc_put() 72 static void *vb2_vmalloc_get_userptr(void *alloc_ctx, unsigned long vaddr, in vb2_vmalloc_get_userptr() argument 85 offset = vaddr & ~PAGE_MASK; in vb2_vmalloc_get_userptr() 87 vec = vb2_create_framevec(vaddr, size, dma_dir == DMA_FROM_DEVICE); in vb2_vmalloc_get_userptr() 102 buf->vaddr = (__force void *) in vb2_vmalloc_get_userptr() 105 buf->vaddr = vm_map_ram(frame_vector_pages(vec), n_pages, -1, in vb2_vmalloc_get_userptr() 109 if (!buf->vaddr) in vb2_vmalloc_get_userptr() [all …]
|
| D | videobuf-vmalloc.c | 105 __func__, i, mem->vaddr); in videobuf_vm_close() 107 vfree(mem->vaddr); in videobuf_vm_close() 108 mem->vaddr = NULL; in videobuf_vm_close() 173 if (!mem->vaddr) { in __videobuf_iolock() 192 mem->vaddr = vmalloc_user(pages); in __videobuf_iolock() 193 if (!mem->vaddr) { in __videobuf_iolock() 198 mem->vaddr, pages); in __videobuf_iolock() 257 mem->vaddr = vmalloc_user(pages); in __videobuf_mmap_mapper() 258 if (!mem->vaddr) { in __videobuf_mmap_mapper() 262 dprintk(1, "vmalloc is at addr %p (%d pages)\n", mem->vaddr, pages); in __videobuf_mmap_mapper() [all …]
|
| D | videobuf-dma-contig.c | 28 void *vaddr; member 45 mem->vaddr = dma_alloc_coherent(dev, mem->size, in __videobuf_dc_alloc() 48 if (!mem->vaddr) { in __videobuf_dc_alloc() 53 dev_dbg(dev, "dma mapped data is at %p (%ld)\n", mem->vaddr, mem->size); in __videobuf_dc_alloc() 61 dma_free_coherent(dev, mem->size, mem->vaddr, mem->dma_handle); in __videobuf_dc_free() 63 mem->vaddr = NULL; in __videobuf_dc_free() 117 i, mem->vaddr); in videobuf_vm_close() 120 mem->vaddr = NULL; in videobuf_vm_close() 233 return mem->vaddr; in __videobuf_to_vaddr() 250 if (!mem->vaddr) { in __videobuf_iolock() [all …]
|
| D | videobuf2-dma-sg.c | 39 void *vaddr; member 120 buf->vaddr = NULL; in vb2_dma_sg_alloc() 193 if (buf->vaddr) in vb2_dma_sg_put() 194 vm_unmap_ram(buf->vaddr, buf->num_pages); in vb2_dma_sg_put() 229 static void *vb2_dma_sg_get_userptr(void *alloc_ctx, unsigned long vaddr, in vb2_dma_sg_get_userptr() argument 244 buf->vaddr = NULL; in vb2_dma_sg_get_userptr() 247 buf->offset = vaddr & ~PAGE_MASK; in vb2_dma_sg_get_userptr() 250 vec = vb2_create_framevec(vaddr, size, buf->dma_dir == DMA_FROM_DEVICE); in vb2_dma_sg_get_userptr() 302 if (buf->vaddr) in vb2_dma_sg_put_userptr() 303 vm_unmap_ram(buf->vaddr, buf->num_pages); in vb2_dma_sg_put_userptr() [all …]
|
| D | videobuf2-dma-contig.c | 30 void *vaddr; member 81 if (!buf->vaddr && buf->db_attach) in vb2_dc_vaddr() 82 buf->vaddr = dma_buf_vmap(buf->db_attach->dmabuf); in vb2_dc_vaddr() 84 return buf->vaddr; in vb2_dc_vaddr() 134 dma_free_coherent(buf->dev, buf->size, buf->vaddr, buf->dma_addr); in vb2_dc_put() 150 buf->vaddr = dma_alloc_coherent(dev, size, &buf->dma_addr, in vb2_dc_alloc() 152 if (!buf->vaddr) { in vb2_dc_alloc() 188 ret = dma_mmap_coherent(buf->dev, vma, buf->vaddr, in vb2_dc_mmap() 332 return buf->vaddr + pgnum * PAGE_SIZE; in vb2_dc_dmabuf_ops_kmap() 339 return buf->vaddr; in vb2_dc_dmabuf_ops_vmap() [all …]
|
| /linux-4.4.14/drivers/lguest/ |
| D | page_tables.c | 83 static pgd_t *spgd_addr(struct lg_cpu *cpu, u32 i, unsigned long vaddr) in spgd_addr() argument 85 unsigned int index = pgd_index(vaddr); in spgd_addr() 97 static pmd_t *spmd_addr(struct lg_cpu *cpu, pgd_t spgd, unsigned long vaddr) in spmd_addr() argument 99 unsigned int index = pmd_index(vaddr); in spmd_addr() 115 static pte_t *spte_addr(struct lg_cpu *cpu, pgd_t spgd, unsigned long vaddr) in spte_addr() argument 118 pmd_t *pmd = spmd_addr(cpu, spgd, vaddr); in spte_addr() 129 return &page[pte_index(vaddr)]; in spte_addr() 136 static unsigned long gpgd_addr(struct lg_cpu *cpu, unsigned long vaddr) in gpgd_addr() argument 138 unsigned int index = vaddr >> (PGDIR_SHIFT); in gpgd_addr() 144 static unsigned long gpmd_addr(pgd_t gpgd, unsigned long vaddr) in gpmd_addr() argument [all …]
|
| /linux-4.4.14/arch/mips/mm/ |
| D | highmem.c | 47 unsigned long vaddr; in kmap_atomic() local 57 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic() 62 local_flush_tlb_one((unsigned long)vaddr); in kmap_atomic() 64 return (void*) vaddr; in kmap_atomic() 70 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 73 if (vaddr < FIXADDR_START) { // FIXME in __kunmap_atomic() 84 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in __kunmap_atomic() 90 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic() 91 local_flush_tlb_one(vaddr); in __kunmap_atomic() 106 unsigned long vaddr; in kmap_atomic_pfn() local [all …]
|
| D | pgtable-32.c | 35 unsigned long vaddr; in pagetable_init() local 54 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; in pagetable_init() 55 fixrange_init(vaddr, vaddr + FIXADDR_SIZE, pgd_base); in pagetable_init() 61 vaddr = PKMAP_BASE; in pagetable_init() 62 fixrange_init(vaddr, vaddr + PAGE_SIZE*LAST_PKMAP, pgd_base); in pagetable_init() 64 pgd = swapper_pg_dir + __pgd_offset(vaddr); in pagetable_init() 65 pud = pud_offset(pgd, vaddr); in pagetable_init() 66 pmd = pmd_offset(pud, vaddr); in pagetable_init() 67 pte = pte_offset_kernel(pmd, vaddr); in pagetable_init()
|
| D | init.c | 88 unsigned long vaddr, flags, entrylo; in __kmap_pgprot() local 99 vaddr = __fix_to_virt(FIX_CMAP_END - idx); in __kmap_pgprot() 109 write_c0_entryhi(vaddr & (PAGE_MASK << 1)); in __kmap_pgprot() 126 return (void*) vaddr; in __kmap_pgprot() 162 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument 169 vfrom = kmap_coherent(from, vaddr); in copy_user_highpage() 178 pages_do_alias((unsigned long)vto, vaddr & PAGE_MASK)) in copy_user_highpage() 186 struct page *page, unsigned long vaddr, void *dst, const void *src, in copy_to_user_page() argument 191 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page() 200 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page() [all …]
|
| D | pgtable-64.c | 97 unsigned long vaddr; in pagetable_init() local 109 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; in pagetable_init() 110 fixrange_init(vaddr, vaddr + FIXADDR_SIZE, pgd_base); in pagetable_init()
|
| /linux-4.4.14/arch/m68k/sun3x/ |
| D | dvma.c | 83 unsigned long vaddr, int len) in dvma_map_cpu() argument 90 vaddr &= PAGE_MASK; in dvma_map_cpu() 92 end = PAGE_ALIGN(vaddr + len); in dvma_map_cpu() 96 kaddr, vaddr); in dvma_map_cpu() 98 pgd = pgd_offset_k(vaddr); in dvma_map_cpu() 104 if((pmd = pmd_alloc(&init_mm, pgd, vaddr)) == NULL) { in dvma_map_cpu() 109 if((end & PGDIR_MASK) > (vaddr & PGDIR_MASK)) in dvma_map_cpu() 110 end2 = (vaddr + (PGDIR_SIZE-1)) & PGDIR_MASK; in dvma_map_cpu() 118 if((pte = pte_alloc_kernel(pmd, vaddr)) == NULL) { in dvma_map_cpu() 123 if((end2 & PMD_MASK) > (vaddr & PMD_MASK)) in dvma_map_cpu() [all …]
|
| /linux-4.4.14/mm/ |
| D | highmem.c | 154 struct page *kmap_to_page(void *vaddr) in kmap_to_page() argument 156 unsigned long addr = (unsigned long)vaddr; in kmap_to_page() 219 unsigned long vaddr; in map_new_virtual() local 261 vaddr = PKMAP_ADDR(last_pkmap_nr); in map_new_virtual() 262 set_pte_at(&init_mm, vaddr, in map_new_virtual() 266 set_page_address(page, (void *)vaddr); in map_new_virtual() 268 return vaddr; in map_new_virtual() 281 unsigned long vaddr; in kmap_high() local 288 vaddr = (unsigned long)page_address(page); in kmap_high() 289 if (!vaddr) in kmap_high() [all …]
|
| D | dmapool.c | 58 void *vaddr; member 217 *(int *)(page->vaddr + offset) = next; in pool_initialise_page() 229 page->vaddr = dma_alloc_coherent(pool->dev, pool->allocation, in pool_alloc_page() 231 if (page->vaddr) { in pool_alloc_page() 233 memset(page->vaddr, POOL_POISON_FREED, pool->allocation); in pool_alloc_page() 255 memset(page->vaddr, POOL_POISON_FREED, pool->allocation); in pool_free_page() 257 dma_free_coherent(pool->dev, pool->allocation, page->vaddr, dma); in pool_free_page() 295 pool->name, page->vaddr); in dma_pool_destroy() 299 pool->name, page->vaddr); in dma_pool_destroy() 350 page->offset = *(int *)(page->vaddr + offset); in dma_pool_alloc() [all …]
|
| /linux-4.4.14/arch/x86/include/asm/ |
| D | pmem.h | 78 static inline void __arch_wb_cache_pmem(void *vaddr, size_t size) in __arch_wb_cache_pmem() argument 82 void *vend = vaddr + size; in __arch_wb_cache_pmem() 85 for (p = (void *)((unsigned long)vaddr & ~clflush_mask); in __arch_wb_cache_pmem() 111 void *vaddr = (void __force *)addr; in arch_copy_from_iter_pmem() local 115 len = copy_from_iter_nocache(vaddr, bytes, i); in arch_copy_from_iter_pmem() 118 __arch_wb_cache_pmem(vaddr, bytes); in arch_copy_from_iter_pmem() 133 void *vaddr = (void __force *)addr; in arch_clear_pmem() local 136 if (size == PAGE_SIZE && ((unsigned long)vaddr & ~PAGE_MASK) == 0) in arch_clear_pmem() 137 clear_page(vaddr); in arch_clear_pmem() 139 memset(vaddr, 0, size); in arch_clear_pmem() [all …]
|
| D | page.h | 24 static inline void clear_user_page(void *page, unsigned long vaddr, in clear_user_page() argument 30 static inline void copy_user_page(void *to, void *from, unsigned long vaddr, in copy_user_page() argument 36 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument 37 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
|
| D | pgtable_32.h | 60 #define kpte_clear_flush(ptep, vaddr) \ argument 62 pte_clear(&init_mm, (vaddr), (ptep)); \ 63 __flush_tlb_one((vaddr)); \
|
| /linux-4.4.14/arch/sparc/include/asm/ |
| D | viking.h | 212 static inline unsigned long viking_hwprobe(unsigned long vaddr) in viking_hwprobe() argument 216 vaddr &= PAGE_MASK; in viking_hwprobe() 220 : "r" (vaddr | 0x400), "i" (ASI_M_FLUSH_PROBE)); in viking_hwprobe() 227 : "r" (vaddr | 0x200), "i" (ASI_M_FLUSH_PROBE)); in viking_hwprobe() 229 vaddr &= ~SRMMU_PGDIR_MASK; in viking_hwprobe() 230 vaddr >>= PAGE_SHIFT; in viking_hwprobe() 231 return val | (vaddr << 8); in viking_hwprobe() 237 : "r" (vaddr | 0x100), "i" (ASI_M_FLUSH_PROBE)); in viking_hwprobe() 239 vaddr &= ~SRMMU_REAL_PMD_MASK; in viking_hwprobe() 240 vaddr >>= PAGE_SHIFT; in viking_hwprobe() [all …]
|
| D | tlbflush_64.h | 20 void flush_tsb_user_page(struct mm_struct *mm, unsigned long vaddr, bool huge); 49 void __flush_tlb_page(unsigned long context, unsigned long vaddr); 54 static inline void global_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr) in global_flush_tlb_page() argument 56 __flush_tlb_page(CTX_HWBITS(mm->context), vaddr); in global_flush_tlb_page() 62 void smp_flush_tlb_page(struct mm_struct *mm, unsigned long vaddr); 64 #define global_flush_tlb_page(mm, vaddr) \ argument 65 smp_flush_tlb_page(mm, vaddr)
|
| D | cacheflush_64.h | 57 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 59 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 61 flush_ptrace_access(vma, page, vaddr, src, len, 0); \ 64 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument 66 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 68 flush_ptrace_access(vma, page, vaddr, dst, len, 1); \
|
| D | cacheflush_32.h | 21 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 23 flush_cache_page(vma, vaddr, page_to_pfn(page));\ 26 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument 28 flush_cache_page(vma, vaddr, page_to_pfn(page));\
|
| D | dma.h | 110 #define mmu_get_scsi_one(dev,vaddr,len) \ argument 111 sparc32_dma_ops->get_scsi_one(dev, vaddr, len) 114 #define mmu_release_scsi_one(dev,vaddr,len) \ argument 115 sparc32_dma_ops->release_scsi_one(dev, vaddr,len)
|
| D | oplib_64.h | 155 unsigned long vaddr); 159 unsigned long vaddr); 175 unsigned long vaddr, unsigned long paddr); 176 void prom_unmap(unsigned long size, unsigned long vaddr);
|
| /linux-4.4.14/arch/x86/mm/ |
| D | init_32.c | 108 pmd_t * __init populate_extra_pmd(unsigned long vaddr) in populate_extra_pmd() argument 110 int pgd_idx = pgd_index(vaddr); in populate_extra_pmd() 111 int pmd_idx = pmd_index(vaddr); in populate_extra_pmd() 116 pte_t * __init populate_extra_pte(unsigned long vaddr) in populate_extra_pte() argument 118 int pte_idx = pte_index(vaddr); in populate_extra_pte() 121 pmd = populate_extra_pmd(vaddr); in populate_extra_pte() 133 unsigned long vaddr; in page_table_range_init_count() local 138 vaddr = start; in page_table_range_init_count() 139 pgd_idx = pgd_index(vaddr); in page_table_range_init_count() 140 pmd_idx = pmd_index(vaddr); in page_table_range_init_count() [all …]
|
| D | highmem_32.c | 35 unsigned long vaddr; in kmap_atomic_prot() local 46 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot() 51 return (void *)vaddr; in kmap_atomic_prot() 73 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 75 if (vaddr >= __fix_to_virt(FIX_KMAP_END) && in __kunmap_atomic() 76 vaddr <= __fix_to_virt(FIX_KMAP_BEGIN)) { in __kunmap_atomic() 83 WARN_ON_ONCE(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in __kunmap_atomic() 91 kpte_clear_flush(kmap_pte-idx, vaddr); in __kunmap_atomic() 97 BUG_ON(vaddr < PAGE_OFFSET); in __kunmap_atomic() 98 BUG_ON(vaddr >= (unsigned long)high_memory); in __kunmap_atomic()
|
| D | pgtable_32.c | 27 void set_pte_vaddr(unsigned long vaddr, pte_t pteval) in set_pte_vaddr() argument 34 pgd = swapper_pg_dir + pgd_index(vaddr); in set_pte_vaddr() 39 pud = pud_offset(pgd, vaddr); in set_pte_vaddr() 44 pmd = pmd_offset(pud, vaddr); in set_pte_vaddr() 49 pte = pte_offset_kernel(pmd, vaddr); in set_pte_vaddr() 51 set_pte_at(&init_mm, vaddr, pte, pteval); in set_pte_vaddr() 53 pte_clear(&init_mm, vaddr, pte); in set_pte_vaddr() 59 __flush_tlb_one(vaddr); in set_pte_vaddr()
|
| D | iomap_32.c | 59 unsigned long vaddr; in kmap_atomic_prot_pfn() local 67 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot_pfn() 71 return (void *)vaddr; in kmap_atomic_prot_pfn() 98 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in iounmap_atomic() local 100 if (vaddr >= __fix_to_virt(FIX_KMAP_END) && in iounmap_atomic() 101 vaddr <= __fix_to_virt(FIX_KMAP_BEGIN)) { in iounmap_atomic() 108 WARN_ON_ONCE(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in iounmap_atomic() 116 kpte_clear_flush(kmap_pte-idx, vaddr); in iounmap_atomic()
|
| D | ioremap.c | 31 int ioremap_change_attr(unsigned long vaddr, unsigned long size, in ioremap_change_attr() argument 40 err = _set_memory_uc(vaddr, nrpages); in ioremap_change_attr() 43 err = _set_memory_wc(vaddr, nrpages); in ioremap_change_attr() 46 err = _set_memory_wt(vaddr, nrpages); in ioremap_change_attr() 49 err = _set_memory_wb(vaddr, nrpages); in ioremap_change_attr() 86 unsigned long offset, vaddr; in __ioremap_caller() local 182 vaddr = (unsigned long) area->addr; in __ioremap_caller() 187 if (ioremap_page_range(vaddr, vaddr + size, phys_addr, prot)) in __ioremap_caller() 190 ret_addr = (void __iomem *) (vaddr + offset); in __ioremap_caller() 400 void *vaddr; in xlate_dev_mem_ptr() local [all …]
|
| D | init_64.c | 233 static pud_t *fill_pud(pgd_t *pgd, unsigned long vaddr) in fill_pud() argument 242 return pud_offset(pgd, vaddr); in fill_pud() 245 static pmd_t *fill_pmd(pud_t *pud, unsigned long vaddr) in fill_pmd() argument 254 return pmd_offset(pud, vaddr); in fill_pmd() 257 static pte_t *fill_pte(pmd_t *pmd, unsigned long vaddr) in fill_pte() argument 265 return pte_offset_kernel(pmd, vaddr); in fill_pte() 268 void set_pte_vaddr_pud(pud_t *pud_page, unsigned long vaddr, pte_t new_pte) in set_pte_vaddr_pud() argument 274 pud = pud_page + pud_index(vaddr); in set_pte_vaddr_pud() 275 pmd = fill_pmd(pud, vaddr); in set_pte_vaddr_pud() 276 pte = fill_pte(pmd, vaddr); in set_pte_vaddr_pud() [all …]
|
| D | pageattr.c | 32 unsigned long *vaddr; member 130 void clflush_cache_range(void *vaddr, unsigned int size) in clflush_cache_range() argument 133 void *vend = vaddr + size; in clflush_cache_range() 138 for (p = (void *)((unsigned long)vaddr & ~clflush_mask); in clflush_cache_range() 1117 static int __cpa_process_fault(struct cpa_data *cpa, unsigned long vaddr, in __cpa_process_fault() argument 1121 return populate_pgd(cpa, vaddr); in __cpa_process_fault() 1136 if (within(vaddr, PAGE_OFFSET, in __cpa_process_fault() 1139 cpa->pfn = __pa(vaddr) >> PAGE_SHIFT; in __cpa_process_fault() 1143 "vaddr = %lx cpa->vaddr = %lx\n", vaddr, in __cpa_process_fault() 1144 *cpa->vaddr); in __cpa_process_fault() [all …]
|
| /linux-4.4.14/drivers/block/ |
| D | cciss.h | 66 void __iomem *vaddr; member 226 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command() 227 readl(h->vaddr + SA5_SCRATCHPAD_OFFSET); in SA5_submit_command() 243 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 244 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 249 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 250 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 263 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask() 264 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask() 269 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5B_intr_mask() [all …]
|
| D | smart1,2.h | 44 writel(c->busaddr, h->vaddr + S42XX_REQUEST_PORT_OFFSET); in smart4_submit_command() 56 writel(0, h->vaddr + S42XX_REPLY_INTR_MASK_OFFSET); in smart4_intr_mask() 60 h->vaddr + S42XX_REPLY_INTR_MASK_OFFSET); in smart4_intr_mask() 72 return (!readl(h->vaddr + S42XX_REQUEST_PORT_OFFSET)); in smart4_fifo_full() 82 = readl(h->vaddr + S42XX_REPLY_PORT_OFFSET); in smart4_completed() 90 writel(0, h->vaddr + S42XX_REPLY_PORT_OFFSET); in smart4_completed() 103 readl(h->vaddr + S42XX_INTR_STATUS); in smart4_intr_pending() 123 writel(c->busaddr, h->vaddr + COMMAND_FIFO); in smart2_submit_command() 128 writel(val, h->vaddr + INTR_MASK); in smart2_intr_mask() 133 return readl(h->vaddr + COMMAND_FIFO); in smart2_fifo_full() [all …]
|
| /linux-4.4.14/drivers/base/ |
| D | dma-mapping.c | 22 void *vaddr; member 30 dma_free_coherent(dev, this->size, this->vaddr, this->dma_handle); in dmam_coherent_release() 37 dma_free_noncoherent(dev, this->size, this->vaddr, this->dma_handle); in dmam_noncoherent_release() 44 if (this->vaddr == match->vaddr) { in dmam_match() 69 void *vaddr; in dmam_alloc_coherent() local 75 vaddr = dma_alloc_coherent(dev, size, dma_handle, gfp); in dmam_alloc_coherent() 76 if (!vaddr) { in dmam_alloc_coherent() 81 dr->vaddr = vaddr; in dmam_alloc_coherent() 87 return vaddr; in dmam_alloc_coherent() 100 void dmam_free_coherent(struct device *dev, size_t size, void *vaddr, in dmam_free_coherent() argument [all …]
|
| D | dma-coherent.c | 213 int dma_release_from_coherent(struct device *dev, int order, void *vaddr) in dma_release_from_coherent() argument 217 if (mem && vaddr >= mem->virt_base && vaddr < in dma_release_from_coherent() 219 int page = (vaddr - mem->virt_base) >> PAGE_SHIFT; in dma_release_from_coherent() 247 void *vaddr, size_t size, int *ret) in dma_mmap_from_coherent() argument 251 if (mem && vaddr >= mem->virt_base && vaddr + size <= in dma_mmap_from_coherent() 254 int start = (vaddr - mem->virt_base) >> PAGE_SHIFT; in dma_mmap_from_coherent()
|
| /linux-4.4.14/arch/powerpc/mm/ |
| D | highmem.c | 34 unsigned long vaddr; in kmap_atomic_prot() local 44 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot() 48 __set_pte_at(&init_mm, vaddr, kmap_pte-idx, mk_pte(page, prot), 1); in kmap_atomic_prot() 49 local_flush_tlb_page(NULL, vaddr); in kmap_atomic_prot() 51 return (void*) vaddr; in kmap_atomic_prot() 57 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 60 if (vaddr < __fix_to_virt(FIX_KMAP_END)) { in __kunmap_atomic() 73 BUG_ON(vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)); in __kunmap_atomic() 79 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic() 80 local_flush_tlb_page(NULL, vaddr); in __kunmap_atomic()
|
| D | hash_utils_64.c | 186 unsigned long vaddr, paddr; in htab_bolt_mapping() local 198 for (vaddr = vstart, paddr = pstart; vaddr < vend; in htab_bolt_mapping() 199 vaddr += step, paddr += step) { in htab_bolt_mapping() 201 unsigned long vsid = get_kernel_vsid(vaddr, ssize); in htab_bolt_mapping() 202 unsigned long vpn = hpt_vpn(vaddr, vsid, ssize); in htab_bolt_mapping() 211 if (overlaps_kernel_text(vaddr, vaddr + step)) in htab_bolt_mapping() 215 if (overlaps_kvm_tmp(vaddr, vaddr + step)) in htab_bolt_mapping() 229 overlaps_interrupt_vector_text(vaddr, vaddr + step)) in htab_bolt_mapping() 253 unsigned long vaddr; in htab_remove_mapping() local 265 for (vaddr = vstart; vaddr < vend; vaddr += step) in htab_remove_mapping() [all …]
|
| D | dma-noncoherent.c | 218 unsigned long vaddr = c->vm_start; in __dma_alloc_coherent() local 230 map_page(vaddr, page_to_phys(page), in __dma_alloc_coherent() 233 vaddr += PAGE_SIZE; in __dma_alloc_coherent() 257 void __dma_free_coherent(size_t size, void *vaddr) in __dma_free_coherent() argument 266 c = ppc_vm_region_find(&consistent_head, (unsigned long)vaddr); in __dma_free_coherent() 309 __func__, vaddr); in __dma_free_coherent() 317 void __dma_sync(void *vaddr, size_t size, int direction) in __dma_sync() argument 319 unsigned long start = (unsigned long)vaddr; in __dma_sync()
|
| /linux-4.4.14/arch/metag/mm/ |
| D | highmem.c | 43 unsigned long vaddr; in kmap_atomic() local 53 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic() 59 return (void *)vaddr; in kmap_atomic() 65 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 78 pte_clear(&init_mm, vaddr, kmap_pte-idx); in __kunmap_atomic() 79 flush_tlb_kernel_range(vaddr, vaddr + PAGE_SIZE); in __kunmap_atomic() 96 unsigned long vaddr; in kmap_atomic_pfn() local 104 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_pfn() 109 flush_tlb_kernel_range(vaddr, vaddr + PAGE_SIZE); in kmap_atomic_pfn() 111 return (void *)vaddr; in kmap_atomic_pfn()
|
| D | init.c | 270 unsigned long vaddr; in allocate_pgtables() local 272 vaddr = start; in allocate_pgtables() 273 i = pgd_index(vaddr); in allocate_pgtables() 274 j = pmd_index(vaddr); in allocate_pgtables() 277 for ( ; (i < PTRS_PER_PGD) && (vaddr != end); pgd++, i++) { in allocate_pgtables() 279 for (; (j < PTRS_PER_PMD) && (vaddr != end); pmd++, j++) { in allocate_pgtables() 280 vaddr += PMD_SIZE; in allocate_pgtables() 294 unsigned long vaddr, end; in fixedrange_init() local 303 vaddr = __fix_to_virt(__end_of_fixed_addresses - 1) & PMD_MASK; in fixedrange_init() 305 allocate_pgtables(vaddr, end); in fixedrange_init() [all …]
|
| D | mmu-meta2.c | 18 unsigned long mmu_read_first_level_page(unsigned long vaddr) in mmu_read_first_level_page() argument 25 if (is_global_space(vaddr)) in mmu_read_first_level_page() 26 vaddr &= ~0x80000000; in mmu_read_first_level_page() 28 offset = vaddr >> PGDIR_SHIFT; in mmu_read_first_level_page() 53 unsigned long mmu_read_second_level_page(unsigned long vaddr) in mmu_read_second_level_page() argument 55 return __builtin_meta2_cacherd((void *)(vaddr & PAGE_MASK)); in mmu_read_second_level_page()
|
| D | mmu-meta1.c | 111 unsigned long mmu_read_first_level_page(unsigned long vaddr) in mmu_read_first_level_page() argument 113 return metag_in32(pgd_entry_addr(vaddr)); in mmu_read_first_level_page() 116 unsigned long mmu_read_second_level_page(unsigned long vaddr) in mmu_read_second_level_page() argument 118 return metag_in32(pgtable_entry_addr(vaddr)); in mmu_read_second_level_page()
|
| /linux-4.4.14/arch/frv/mm/ |
| D | cache-page.c | 25 void *vaddr; in flush_dcache_page() local 29 vaddr = kmap_atomic_primary(page); in flush_dcache_page() 31 frv_dcache_writeback((unsigned long) vaddr, (unsigned long) vaddr + PAGE_SIZE); in flush_dcache_page() 33 kunmap_atomic_primary(vaddr); in flush_dcache_page() 53 void *vaddr; in flush_icache_user_range() local 57 vaddr = kmap_atomic_primary(page); in flush_icache_user_range() 59 start = (start & ~PAGE_MASK) | (unsigned long) vaddr; in flush_icache_user_range() 62 kunmap_atomic_primary(vaddr); in flush_icache_user_range()
|
| D | dma-alloc.c | 141 void consistent_free(void *vaddr) in consistent_free() argument 145 vfree(vaddr); in consistent_free() 151 void consistent_sync(void *vaddr, size_t size, int direction) in consistent_sync() argument 153 unsigned long start = (unsigned long) vaddr; in consistent_sync()
|
| /linux-4.4.14/arch/um/kernel/ |
| D | mem.c | 94 unsigned long vaddr; in fixrange_init() local 96 vaddr = start; in fixrange_init() 97 i = pgd_index(vaddr); in fixrange_init() 98 j = pmd_index(vaddr); in fixrange_init() 101 for ( ; (i < PTRS_PER_PGD) && (vaddr < end); pgd++, i++) { in fixrange_init() 102 pud = pud_offset(pgd, vaddr); in fixrange_init() 105 pmd = pmd_offset(pud, vaddr); in fixrange_init() 106 for (; (j < PTRS_PER_PMD) && (vaddr < end); pmd++, j++) { in fixrange_init() 108 vaddr += PMD_SIZE; in fixrange_init() 123 unsigned long v, vaddr = FIXADDR_USER_START; in fixaddr_user_init() local [all …]
|
| /linux-4.4.14/arch/tile/mm/ |
| D | highmem.c | 20 #define kmap_get_pte(vaddr) \ argument 21 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)),\ 22 (vaddr)), (vaddr)) 200 unsigned long vaddr; in kmap_atomic_prot() local 215 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic_prot() 216 pte = kmap_get_pte(vaddr); in kmap_atomic_prot() 220 kmap_atomic_register(page, type, vaddr, pte, mk_pte(page, prot)); in kmap_atomic_prot() 222 return (void *)vaddr; in kmap_atomic_prot() 235 unsigned long vaddr = (unsigned long) kvaddr & PAGE_MASK; in __kunmap_atomic() local 237 if (vaddr >= __fix_to_virt(FIX_KMAP_END) && in __kunmap_atomic() [all …]
|
| D | init.c | 180 unsigned long vaddr; in page_table_range_init() local 183 for (vaddr = start; vaddr < end; vaddr += PMD_SIZE) { in page_table_range_init() 184 pmd_t *pmd = get_pmd(pgd, vaddr); in page_table_range_init() 643 unsigned long vaddr; in permanent_kmaps_init() local 645 vaddr = PKMAP_BASE; in permanent_kmaps_init() 646 page_table_range_init(vaddr, vaddr + PAGE_SIZE*LAST_PKMAP, pgd_base); in permanent_kmaps_init() 648 pgd = swapper_pg_dir + pgd_index(vaddr); in permanent_kmaps_init() 649 pud = pud_offset(pgd, vaddr); in permanent_kmaps_init() 650 pmd = pmd_offset(pud, vaddr); in permanent_kmaps_init() 651 pte = pte_offset_kernel(pmd, vaddr); in permanent_kmaps_init()
|
| /linux-4.4.14/arch/xtensa/mm/ |
| D | cache.c | 64 unsigned long vaddr) in kmap_invalidate_coherent() argument 66 if (!DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) { in kmap_invalidate_coherent() 84 unsigned long vaddr, unsigned long *paddr) in coherent_kvaddr() argument 86 if (PageHighMem(page) || !DCACHE_ALIAS_EQ(page_to_phys(page), vaddr)) { in coherent_kvaddr() 88 return (void *)(base + (vaddr & DCACHE_ALIAS_MASK)); in coherent_kvaddr() 95 void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument 98 void *kvaddr = coherent_kvaddr(page, TLBTEMP_BASE_1, vaddr, &paddr); in clear_user_highpage() 101 kmap_invalidate_coherent(page, vaddr); in clear_user_highpage() 108 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument 111 void *dst_vaddr = coherent_kvaddr(dst, TLBTEMP_BASE_1, vaddr, in copy_user_highpage() [all …]
|
| D | mmu.c | 21 static void * __init init_pmd(unsigned long vaddr, unsigned long n_pages) in init_pmd() argument 23 pgd_t *pgd = pgd_offset_k(vaddr); in init_pmd() 24 pmd_t *pmd = pmd_offset(pgd, vaddr); in init_pmd() 31 __func__, vaddr, n_pages); in init_pmd()
|
| D | highmem.c | 43 unsigned long vaddr; in kmap_atomic() local 52 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic() 58 return (void *)vaddr; in kmap_atomic()
|
| /linux-4.4.14/arch/avr32/include/asm/ |
| D | cacheflush.h | 26 static inline void invalidate_dcache_line(void *vaddr) in invalidate_dcache_line() argument 30 : "r"(vaddr), "n"(CACHE_OP_DCACHE_INVALIDATE) in invalidate_dcache_line() 38 static inline void clean_dcache_line(void *vaddr) in clean_dcache_line() argument 42 : "r"(vaddr), "n"(CACHE_OP_DCACHE_CLEAN) in clean_dcache_line() 50 static inline void flush_dcache_line(void *vaddr) in flush_dcache_line() argument 54 : "r"(vaddr), "n"(CACHE_OP_DCACHE_CLEAN_INVAL) in flush_dcache_line() 62 static inline void invalidate_icache_line(void *vaddr) in invalidate_icache_line() argument 66 : "r"(vaddr), "n"(CACHE_OP_ICACHE_INVALIDATE) in invalidate_icache_line() 122 unsigned long vaddr, void *dst, const void *src, 126 struct page *page, unsigned long vaddr, void *dst, in copy_from_user_page() argument
|
| /linux-4.4.14/arch/xtensa/include/asm/ |
| D | fixmap.h | 66 static inline unsigned long virt_to_fix(const unsigned long vaddr) in virt_to_fix() argument 68 BUG_ON(vaddr >= FIXADDR_TOP || vaddr < FIXADDR_START); in virt_to_fix() 69 return __virt_to_fix(vaddr); in virt_to_fix() 74 #define kmap_get_fixmap_pte(vaddr) \ argument 76 pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), \ 77 (vaddr) \
|
| D | page.h | 149 extern void clear_page_alias(void *vaddr, unsigned long paddr); 154 void clear_user_highpage(struct page *page, unsigned long vaddr); 157 unsigned long vaddr, struct vm_area_struct *vma); 159 # define clear_user_page(page, vaddr, pg) clear_page(page) argument 160 # define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
|
| /linux-4.4.14/arch/frv/include/asm/ |
| D | virtconvert.h | 24 #define phys_to_virt(vaddr) ((void *) ((unsigned long)(vaddr) + PAGE_OFFSET)) argument 25 #define virt_to_phys(vaddr) ((unsigned long) (vaddr) - PAGE_OFFSET) argument 29 #define phys_to_virt(vaddr) ((void *) (vaddr)) argument 30 #define virt_to_phys(vaddr) ((unsigned long) (vaddr)) argument
|
| D | page.h | 11 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument 17 #define clear_user_page(pgaddr, vaddr, page) memset((pgaddr), 0, PAGE_SIZE) argument 18 #define copy_user_page(vto, vfrom, vaddr, topg) memcpy((vto), (vfrom), PAGE_SIZE) argument 45 #define __pa(vaddr) virt_to_phys((void *) (unsigned long) (vaddr)) argument
|
| D | pci.h | 29 extern void consistent_free(void *vaddr); 30 extern void consistent_sync(void *vaddr, size_t size, int direction); 39 void *vaddr, dma_addr_t dma_handle);
|
| D | cacheflush.h | 96 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 99 flush_icache_user_range((vma), (page), (vaddr), (len)); \ 102 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
| /linux-4.4.14/drivers/misc/sgi-gru/ |
| D | grufault.c | 61 struct vm_area_struct *gru_find_vma(unsigned long vaddr) in gru_find_vma() argument 65 vma = find_vma(current->mm, vaddr); in gru_find_vma() 66 if (vma && vma->vm_start <= vaddr && vma->vm_ops == &gru_vm_ops) in gru_find_vma() 79 static struct gru_thread_state *gru_find_lock_gts(unsigned long vaddr) in gru_find_lock_gts() argument 86 vma = gru_find_vma(vaddr); in gru_find_lock_gts() 88 gts = gru_find_thread_state(vma, TSID(vaddr, vma)); in gru_find_lock_gts() 96 static struct gru_thread_state *gru_alloc_locked_gts(unsigned long vaddr) in gru_alloc_locked_gts() argument 103 vma = gru_find_vma(vaddr); in gru_alloc_locked_gts() 107 gts = gru_alloc_thread_state(vma, TSID(vaddr, vma)); in gru_alloc_locked_gts() 191 unsigned long vaddr, int write, in non_atomic_pte_lookup() argument [all …]
|
| D | gruhandles.c | 152 unsigned long vaddr, unsigned long vaddrmask, in tgh_invalidate() argument 156 tgh->vaddr = vaddr; in tgh_invalidate() 170 unsigned long vaddr, int asid, int dirty, in tfh_write_only() argument 174 tfh->fillvaddr = vaddr; in tfh_write_only() 186 unsigned long vaddr, int asid, int dirty, in tfh_write_restart() argument 190 tfh->fillvaddr = vaddr; in tfh_write_restart()
|
| /linux-4.4.14/arch/arm/kernel/ |
| D | crash_dump.c | 37 void *vaddr; in copy_oldmem_page() local 42 vaddr = ioremap(__pfn_to_phys(pfn), PAGE_SIZE); in copy_oldmem_page() 43 if (!vaddr) in copy_oldmem_page() 47 if (copy_to_user(buf, vaddr + offset, csize)) { in copy_oldmem_page() 48 iounmap(vaddr); in copy_oldmem_page() 52 memcpy(buf, vaddr + offset, csize); in copy_oldmem_page() 55 iounmap(vaddr); in copy_oldmem_page()
|
| /linux-4.4.14/arch/x86/kernel/ |
| D | crash_dump_64.c | 29 void *vaddr; in copy_oldmem_page() local 34 vaddr = ioremap_cache(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page() 35 if (!vaddr) in copy_oldmem_page() 39 if (copy_to_user(buf, vaddr + offset, csize)) { in copy_oldmem_page() 40 iounmap(vaddr); in copy_oldmem_page() 44 memcpy(buf, vaddr + offset, csize); in copy_oldmem_page() 47 iounmap(vaddr); in copy_oldmem_page()
|
| D | crash_dump_32.c | 53 void *vaddr; in copy_oldmem_page() local 61 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page() 64 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_page() 65 kunmap_atomic(vaddr); in copy_oldmem_page() 70 kunmap_atomic(vaddr); in copy_oldmem_page() 73 copy_page(kdump_buf_page, vaddr); in copy_oldmem_page() 74 kunmap_atomic(vaddr); in copy_oldmem_page()
|
| D | pci-swiotlb.c | 21 void *vaddr; in x86_swiotlb_alloc_coherent() local 30 vaddr = dma_generic_alloc_coherent(hwdev, size, dma_handle, flags, in x86_swiotlb_alloc_coherent() 32 if (vaddr) in x86_swiotlb_alloc_coherent() 33 return vaddr; in x86_swiotlb_alloc_coherent() 39 void *vaddr, dma_addr_t dma_addr, in x86_swiotlb_free_coherent() argument 43 swiotlb_free_coherent(dev, size, vaddr, dma_addr); in x86_swiotlb_free_coherent() 45 dma_generic_free_coherent(dev, size, vaddr, dma_addr, attrs); in x86_swiotlb_free_coherent()
|
| D | tboot.c | 123 static int map_tboot_page(unsigned long vaddr, unsigned long pfn, in map_tboot_page() argument 131 pgd = pgd_offset(&tboot_mm, vaddr); in map_tboot_page() 132 pud = pud_alloc(&tboot_mm, pgd, vaddr); in map_tboot_page() 135 pmd = pmd_alloc(&tboot_mm, pud, vaddr); in map_tboot_page() 138 pte = pte_alloc_map(&tboot_mm, NULL, pmd, vaddr); in map_tboot_page() 141 set_pte_at(&tboot_mm, vaddr, pte, pfn_pte(pfn, prot)); in map_tboot_page() 146 static int map_tboot_pages(unsigned long vaddr, unsigned long start_pfn, in map_tboot_pages() argument 154 for (; nr > 0; nr--, vaddr += PAGE_SIZE, start_pfn++) { in map_tboot_pages() 155 if (map_tboot_page(vaddr, start_pfn, PAGE_KERNEL_EXEC)) in map_tboot_pages()
|
| D | machine_kexec_32.c | 104 unsigned long vaddr, unsigned long paddr) in machine_kexec_page_table_set_one() argument 108 pgd += pgd_index(vaddr); in machine_kexec_page_table_set_one() 113 pud = pud_offset(pgd, vaddr); in machine_kexec_page_table_set_one() 114 pmd = pmd_offset(pud, vaddr); in machine_kexec_page_table_set_one() 117 pte = pte_offset_kernel(pmd, vaddr); in machine_kexec_page_table_set_one()
|
| /linux-4.4.14/arch/parisc/include/asm/ |
| D | cacheflush.h | 51 static inline void flush_kernel_vmap_range(void *vaddr, int size) in flush_kernel_vmap_range() argument 53 unsigned long start = (unsigned long)vaddr; in flush_kernel_vmap_range() 57 static inline void invalidate_kernel_vmap_range(void *vaddr, int size) in invalidate_kernel_vmap_range() argument 59 unsigned long start = (unsigned long)vaddr; in invalidate_kernel_vmap_range() 60 void *cursor = vaddr; in invalidate_kernel_vmap_range() 62 for ( ; cursor < vaddr + size; cursor += PAGE_SIZE) { in invalidate_kernel_vmap_range() 92 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 94 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ 99 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument 101 flush_cache_page(vma, vaddr, page_to_pfn(page)); \ [all …]
|
| D | dma-mapping.h | 13 void (*free_consistent)(struct device *dev, size_t size, void *vaddr, dma_addr_t iova); 68 void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument 70 hppa_dma_ops->free_consistent(dev, size, vaddr, dma_handle); in dma_free_coherent() 75 void *vaddr, dma_addr_t dma_handle) in dma_free_noncoherent() argument 77 hppa_dma_ops->free_consistent(dev, size, vaddr, dma_handle); in dma_free_noncoherent() 191 dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument 195 flush_kernel_dcache_range((unsigned long)vaddr, size); in dma_cache_sync()
|
| /linux-4.4.14/arch/m68k/mm/ |
| D | cache.c | 14 static unsigned long virt_to_phys_slow(unsigned long vaddr) in virt_to_phys_slow() argument 36 : "0" (vaddr)); in virt_to_phys_slow() 46 : "a" (vaddr)); in virt_to_phys_slow() 49 return (mmusr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow() 57 : "a" (vaddr), "d" (get_fs().seg)); in virt_to_phys_slow() 63 return (*descaddr & 0xfe000000) | (vaddr & 0x01ffffff); in virt_to_phys_slow() 65 return (*descaddr & 0xfffc0000) | (vaddr & 0x0003ffff); in virt_to_phys_slow() 67 return (*descaddr & PAGE_MASK) | (vaddr & ~PAGE_MASK); in virt_to_phys_slow()
|
| /linux-4.4.14/arch/sh/kernel/ |
| D | crash_dump.c | 28 void *vaddr; in copy_oldmem_page() local 33 vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); in copy_oldmem_page() 36 if (copy_to_user(buf, (vaddr + offset), csize)) { in copy_oldmem_page() 37 iounmap(vaddr); in copy_oldmem_page() 41 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_page() 43 iounmap(vaddr); in copy_oldmem_page()
|
| /linux-4.4.14/arch/mn10300/include/asm/ |
| D | highmem.h | 75 unsigned long vaddr; in kmap_atomic() local 85 vaddr = __fix_to_virt(FIX_KMAP_BEGIN + idx); in kmap_atomic() 91 local_flush_tlb_one(vaddr); in kmap_atomic() 93 return (void *)vaddr; in kmap_atomic() 96 static inline void __kunmap_atomic(unsigned long vaddr) in __kunmap_atomic() argument 100 if (vaddr < FIXADDR_START) { /* FIXME */ in __kunmap_atomic() 113 if (vaddr != __fix_to_virt(FIX_KMAP_BEGIN + idx)) in __kunmap_atomic() 121 local_flush_tlb_one(vaddr); in __kunmap_atomic()
|
| /linux-4.4.14/arch/powerpc/kernel/ |
| D | crash_dump.c | 72 static size_t copy_oldmem_vaddr(void *vaddr, char *buf, size_t csize, in copy_oldmem_vaddr() argument 76 if (copy_to_user((char __user *)buf, (vaddr + offset), csize)) in copy_oldmem_vaddr() 79 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_vaddr() 100 void *vaddr; in copy_oldmem_page() local 110 vaddr = __va(paddr); in copy_oldmem_page() 111 csize = copy_oldmem_vaddr(vaddr, buf, csize, offset, userbuf); in copy_oldmem_page() 113 vaddr = __ioremap(paddr, PAGE_SIZE, 0); in copy_oldmem_page() 114 csize = copy_oldmem_vaddr(vaddr, buf, csize, offset, userbuf); in copy_oldmem_page() 115 iounmap(vaddr); in copy_oldmem_page()
|
| D | io-workarounds.c | 28 static struct iowa_bus *iowa_pci_find(unsigned long vaddr, unsigned long paddr) in iowa_pci_find() argument 38 if (vaddr) { in iowa_pci_find() 41 if ((vaddr >= vstart) && (vaddr <= vend)) in iowa_pci_find() 68 unsigned long vaddr, paddr; in iowa_mem_find_bus() local 71 vaddr = (unsigned long)PCI_FIX_ADDR(addr); in iowa_mem_find_bus() 72 if (vaddr < PHB_IO_BASE || vaddr >= PHB_IO_END) in iowa_mem_find_bus() 78 ptep = __find_linux_pte_or_hugepte(init_mm.pgd, vaddr, in iowa_mem_find_bus() 86 bus = iowa_pci_find(vaddr, paddr); in iowa_mem_find_bus() 104 unsigned long vaddr = (unsigned long)pci_io_base + port; in iowa_pio_find_bus() local 105 return iowa_pci_find(vaddr, 0); in iowa_pio_find_bus()
|
| D | iommu.c | 460 unsigned long vaddr, npages, entry, slen; in ppc_iommu_map_sg() local 469 vaddr = (unsigned long) sg_virt(s); in ppc_iommu_map_sg() 470 npages = iommu_num_pages(vaddr, slen, IOMMU_PAGE_SIZE(tbl)); in ppc_iommu_map_sg() 473 (vaddr & ~PAGE_MASK) == 0) in ppc_iommu_map_sg() 478 DBG(" - vaddr: %lx, size: %lx\n", vaddr, slen); in ppc_iommu_map_sg() 484 "vaddr %lx npages %lu\n", tbl, vaddr, in ppc_iommu_map_sg() 499 vaddr & IOMMU_PAGE_MASK(tbl), in ppc_iommu_map_sg() 559 unsigned long vaddr, npages; in ppc_iommu_map_sg() local 561 vaddr = s->dma_address & IOMMU_PAGE_MASK(tbl); in ppc_iommu_map_sg() 564 __iommu_free(tbl, vaddr, npages); in ppc_iommu_map_sg() [all …]
|
| D | fadump.c | 549 void *vaddr; in fadump_cpu_notes_buf_alloc() local 554 vaddr = (void *)__get_free_pages(GFP_KERNEL|__GFP_ZERO, order); in fadump_cpu_notes_buf_alloc() 555 if (!vaddr) in fadump_cpu_notes_buf_alloc() 559 page = virt_to_page(vaddr); in fadump_cpu_notes_buf_alloc() 562 return vaddr; in fadump_cpu_notes_buf_alloc() 565 static void fadump_cpu_notes_buf_free(unsigned long vaddr, unsigned long size) in fadump_cpu_notes_buf_free() argument 572 page = virt_to_page(vaddr); in fadump_cpu_notes_buf_free() 597 void *vaddr; in fadump_build_cpu_notes() local 607 vaddr = __va(addr); in fadump_build_cpu_notes() 609 reg_header = vaddr; in fadump_build_cpu_notes() [all …]
|
| /linux-4.4.14/include/linux/ |
| D | highmem.h | 23 static inline void flush_kernel_vmap_range(void *vaddr, int size) in flush_kernel_vmap_range() argument 26 static inline void invalidate_kernel_vmap_range(void *vaddr, int size) in invalidate_kernel_vmap_range() argument 133 static inline void clear_user_highpage(struct page *page, unsigned long vaddr) in clear_user_highpage() argument 136 clear_user_page(addr, vaddr, page); in clear_user_highpage() 159 unsigned long vaddr) in __alloc_zeroed_user_highpage() argument 162 vma, vaddr); in __alloc_zeroed_user_highpage() 165 clear_user_highpage(page, vaddr); in __alloc_zeroed_user_highpage() 181 unsigned long vaddr) in alloc_zeroed_user_highpage_movable() argument 183 return __alloc_zeroed_user_highpage(__GFP_MOVABLE, vma, vaddr); in alloc_zeroed_user_highpage_movable() 226 unsigned long vaddr, struct vm_area_struct *vma) in copy_user_highpage() argument [all …]
|
| D | io-mapping.h | 97 io_mapping_unmap_atomic(void __iomem *vaddr) in io_mapping_unmap_atomic() argument 99 iounmap_atomic(vaddr); in io_mapping_unmap_atomic() 114 io_mapping_unmap(void __iomem *vaddr) in io_mapping_unmap() argument 116 iounmap(vaddr); in io_mapping_unmap() 150 io_mapping_unmap_atomic(void __iomem *vaddr) in io_mapping_unmap_atomic() argument 164 io_mapping_unmap(void __iomem *vaddr) in io_mapping_unmap() argument
|
| D | uprobes.h | 79 unsigned long vaddr; member 117 extern int set_swbp(struct arch_uprobe *aup, struct mm_struct *mm, unsigned long vaddr); 118 extern int set_orig_insn(struct arch_uprobe *aup, struct mm_struct *mm, unsigned long vaddr); 123 extern int uprobe_write_opcode(struct mm_struct *mm, unsigned long vaddr, uprobe_opcode_t); 149 extern void arch_uprobe_copy_ixol(struct page *page, unsigned long vaddr,
|
| /linux-4.4.14/arch/mips/kernel/ |
| D | crash_dump.c | 29 void *vaddr; in copy_oldmem_page() local 34 vaddr = kmap_atomic_pfn(pfn); in copy_oldmem_page() 37 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_page() 38 kunmap_atomic(vaddr); in copy_oldmem_page() 45 copy_page(kdump_buf_page, vaddr); in copy_oldmem_page() 46 kunmap_atomic(vaddr); in copy_oldmem_page()
|
| D | mips-r2-to-r6-emul.c | 908 unsigned long vaddr; in mipsr2_decoder() local 1201 vaddr = regs->regs[MIPSInst_RS(inst)] + MIPSInst_SIMM(inst); in mipsr2_decoder() 1202 if (!access_ok(VERIFY_READ, vaddr, 4)) { in mipsr2_decoder() 1203 current->thread.cp0_baduaddr = vaddr; in mipsr2_decoder() 1262 "+&r"(vaddr), "+&r"(err) in mipsr2_decoder() 1274 vaddr = regs->regs[MIPSInst_RS(inst)] + MIPSInst_SIMM(inst); in mipsr2_decoder() 1275 if (!access_ok(VERIFY_READ, vaddr, 4)) { in mipsr2_decoder() 1276 current->thread.cp0_baduaddr = vaddr; in mipsr2_decoder() 1337 "+&r"(vaddr), "+&r"(err) in mipsr2_decoder() 1348 vaddr = regs->regs[MIPSInst_RS(inst)] + MIPSInst_SIMM(inst); in mipsr2_decoder() [all …]
|
| D | uprobes.c | 249 instruction_pointer_set(regs, utask->vaddr); in arch_uprobe_abort_xol() 278 unsigned long vaddr) in set_swbp() argument 280 return uprobe_write_opcode(mm, vaddr, UPROBE_SWBP_INSN); in set_swbp() 295 unsigned long vaddr) in set_orig_insn() argument 297 return uprobe_write_opcode(mm, vaddr, in set_orig_insn() 301 void __weak arch_uprobe_copy_ixol(struct page *page, unsigned long vaddr, in arch_uprobe_copy_ixol() argument 308 memcpy(kaddr + (vaddr & ~PAGE_MASK), src, len); in arch_uprobe_copy_ixol() 316 flush_icache_range(vaddr, vaddr + len); in arch_uprobe_copy_ixol()
|
| /linux-4.4.14/arch/hexagon/include/asm/ |
| D | fixmap.h | 31 #define kmap_get_fixmap_pte(vaddr) \ argument 32 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), \ 33 (vaddr)), (vaddr)), (vaddr))
|
| /linux-4.4.14/drivers/scsi/ |
| D | hpsa.h | 163 void __iomem *vaddr; member 411 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command() 412 (void) readl(h->vaddr + SA5_SCRATCHPAD_OFFSET); in SA5_submit_command() 418 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command_no_read() 424 writel(c->busaddr, h->vaddr + SA5_REQUEST_PORT_OFFSET); in SA5_submit_command_ioaccel2() 436 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 437 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 441 h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 442 (void) readl(h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_intr_mask() 450 writel(0, h->vaddr + SA5_REPLY_INTR_MASK_OFFSET); in SA5_performant_intr_mask() [all …]
|
| D | sun3_scsi.c | 402 unsigned char *vaddr; in sun3scsi_dma_finish() local 404 vaddr = (unsigned char *)dvma_vmetov(sun3_dma_orig_addr); in sun3scsi_dma_finish() 406 vaddr += (sun3_dma_orig_count - fifo); in sun3scsi_dma_finish() 407 vaddr--; in sun3scsi_dma_finish() 411 *vaddr = (dregs->bpack_lo & 0xff00) >> 8; in sun3scsi_dma_finish() 412 vaddr--; in sun3scsi_dma_finish() 415 *vaddr = (dregs->bpack_hi & 0x00ff); in sun3scsi_dma_finish() 416 vaddr--; in sun3scsi_dma_finish() 419 *vaddr = (dregs->bpack_hi & 0xff00) >> 8; in sun3scsi_dma_finish() 448 unsigned char *vaddr; in sun3scsi_dma_finish() local [all …]
|
| D | NCR_Q720.c | 79 int irq, int slot, __u32 paddr, void __iomem *vaddr) in NCR_Q720_probe_one() argument 84 __u8 scsr1 = readb(vaddr + NCR_Q720_SCSR_OFFSET + 1); in NCR_Q720_probe_one() 85 __u8 differential = readb(vaddr + NCR_Q720_SCSR_OFFSET) & 0x20; in NCR_Q720_probe_one() 94 writeb(scsr1, vaddr + NCR_Q720_SCSR_OFFSET + 1); in NCR_Q720_probe_one() 96 version = readb(vaddr + 0x18) >> 4; in NCR_Q720_probe_one() 106 device.slot.base_v = vaddr; in NCR_Q720_probe_one() 118 scsr1 = readb(vaddr + NCR_Q720_SCSR_OFFSET + 1); in NCR_Q720_probe_one() 121 writeb(scsr1, vaddr + NCR_Q720_SCSR_OFFSET + 1); in NCR_Q720_probe_one()
|
| /linux-4.4.14/arch/arm/include/asm/ |
| D | page.h | 114 void (*cpu_clear_user_highpage)(struct page *page, unsigned long vaddr); 116 unsigned long vaddr, struct vm_area_struct *vma); 130 extern void __cpu_clear_user_highpage(struct page *page, unsigned long vaddr); 132 unsigned long vaddr, struct vm_area_struct *vma); 135 #define clear_user_highpage(page,vaddr) \ argument 136 __cpu_clear_user_highpage(page, vaddr) 139 #define copy_user_highpage(to,from,vaddr,vma) \ argument 140 __cpu_copy_user_highpage(to, from, vaddr, vma)
|
| D | page-nommu.h | 20 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument 26 #define clear_user_page(page, vaddr, pg) clear_page(page) argument 27 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
|
| /linux-4.4.14/arch/arc/mm/ |
| D | cache.c | 28 void (*_cache_line_loop_ic_fn)(phys_addr_t paddr, unsigned long vaddr, 219 void __cache_line_loop_v2(phys_addr_t paddr, unsigned long vaddr, in __cache_line_loop_v2() argument 242 vaddr &= CACHE_LINE_MASK; in __cache_line_loop_v2() 248 paddr |= (vaddr >> PAGE_SHIFT) & 0x1F; in __cache_line_loop_v2() 261 void __cache_line_loop_v3(phys_addr_t paddr, unsigned long vaddr, in __cache_line_loop_v3() argument 285 vaddr &= CACHE_LINE_MASK; in __cache_line_loop_v3() 312 write_aux_reg(aux_cmd, vaddr); in __cache_line_loop_v3() 313 vaddr += L1_CACHE_BYTES; in __cache_line_loop_v3() 331 void __cache_line_loop_v4(phys_addr_t paddr, unsigned long vaddr, in __cache_line_loop_v4() argument 451 static inline void __dc_line_op(phys_addr_t paddr, unsigned long vaddr, in __dc_line_op() argument [all …]
|
| D | ioremap.c | 44 void __iomem *vaddr; in ioremap_prot() local 73 vaddr = (void __iomem *)area->addr; in ioremap_prot() 74 if (ioremap_page_range((unsigned long)vaddr, in ioremap_prot() 75 (unsigned long)vaddr + size, paddr, prot)) { in ioremap_prot() 76 vunmap((void __force *)vaddr); in ioremap_prot() 79 return (void __iomem *)(off + (char __iomem *)vaddr); in ioremap_prot()
|
| D | highmem.c | 68 unsigned long vaddr; in kmap_atomic() local 77 vaddr = FIXMAP_ADDR(idx); in kmap_atomic() 79 set_pte_at(&init_mm, vaddr, fixmap_page_table + idx, in kmap_atomic() 82 return (void *)vaddr; in kmap_atomic()
|
| /linux-4.4.14/arch/frv/mb93090-mb00/ |
| D | pci-dma.c | 34 void dma_free_coherent(struct device *hwdev, size_t size, void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument 36 consistent_free(vaddr); in dma_free_coherent() 57 void *vaddr; in dma_map_sg() local 66 vaddr = kmap_atomic_primary(sg_page(sg)); in dma_map_sg() 68 frv_dcache_writeback((unsigned long) vaddr, in dma_map_sg() 69 (unsigned long) vaddr + PAGE_SIZE); in dma_map_sg() 73 kunmap_atomic_primary(vaddr); in dma_map_sg()
|
| /linux-4.4.14/drivers/xen/xenbus/ |
| D | xenbus_client.c | 78 void **vaddr); 79 int (*unmap)(struct xenbus_device *dev, void *vaddr); 380 int xenbus_grant_ring(struct xenbus_device *dev, void *vaddr, in xenbus_grant_ring() argument 388 virt_to_gfn(vaddr), 0); in xenbus_grant_ring() 396 vaddr = vaddr + XEN_PAGE_SIZE; in xenbus_grant_ring() 470 unsigned int nr_grefs, void **vaddr) in xenbus_map_ring_valloc() argument 472 return ring_ops->map(dev, gnt_refs, nr_grefs, vaddr); in xenbus_map_ring_valloc() 544 void **vaddr) in xenbus_map_ring_valloc_pv() argument 554 *vaddr = NULL; in xenbus_map_ring_valloc_pv() 586 *vaddr = area->addr; in xenbus_map_ring_valloc_pv() [all …]
|
| /linux-4.4.14/arch/score/include/asm/ |
| D | page.h | 36 #define clear_user_page(pgaddr, vaddr, page) memset((pgaddr), 0, PAGE_SIZE) argument 37 #define copy_user_page(vto, vfrom, vaddr, topg) \ argument 67 #define virt_to_pfn(vaddr) (phys_to_pfn((__pa(vaddr)))) argument 70 #define virt_to_page(vaddr) (pfn_to_page(virt_to_pfn(vaddr))) argument 83 #define virt_addr_valid(vaddr) (pfn_valid(virt_to_pfn(vaddr))) argument
|
| D | cacheflush.h | 38 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument 41 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 45 flush_cache_page(vma, vaddr, page_to_pfn(page));\
|
| D | tlbflush.h | 23 extern void local_flush_tlb_one(unsigned long vaddr); 32 #define flush_tlb_one(vaddr) local_flush_tlb_one(vaddr) argument
|
| /linux-4.4.14/arch/nios2/include/asm/ |
| D | dma-mapping.h | 17 static inline void __dma_sync_for_device(void *vaddr, size_t size, in __dma_sync_for_device() argument 22 invalidate_dcache_range((unsigned long)vaddr, in __dma_sync_for_device() 23 (unsigned long)(vaddr + size)); in __dma_sync_for_device() 31 flush_dcache_range((unsigned long)vaddr, in __dma_sync_for_device() 32 (unsigned long)(vaddr + size)); in __dma_sync_for_device() 39 static inline void __dma_sync_for_cpu(void *vaddr, size_t size, in __dma_sync_for_cpu() argument 45 invalidate_dcache_range((unsigned long)vaddr, in __dma_sync_for_cpu() 46 (unsigned long)(vaddr + size)); in __dma_sync_for_cpu() 62 void *vaddr, dma_addr_t dma_handle); 125 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument
|
| D | page.h | 53 extern void clear_user_page(void *addr, unsigned long vaddr, struct page *page); 54 extern void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, 93 # define virt_to_page(vaddr) pfn_to_page(PFN_DOWN(virt_to_phys(vaddr))) argument 94 # define virt_addr_valid(vaddr) pfn_valid(PFN_DOWN(virt_to_phys(vaddr))) argument
|
| D | io.h | 56 #define phys_to_virt(vaddr) \ argument 57 ((void *)((unsigned long)(vaddr) | CONFIG_NIOS2_KERNEL_REGION_BASE)) 59 #define virt_to_phys(vaddr) \ argument 60 ((unsigned long)((unsigned long)(vaddr) & ~0xE0000000))
|
| /linux-4.4.14/kernel/events/ |
| D | uprobes.c | 110 unsigned long vaddr; /* Page(s) of instruction slots */ member 136 static loff_t vaddr_to_offset(struct vm_area_struct *vma, unsigned long vaddr) in vaddr_to_offset() argument 138 return ((loff_t)vma->vm_pgoff << PAGE_SHIFT) + (vaddr - vma->vm_start); in vaddr_to_offset() 233 static void copy_from_page(struct page *page, unsigned long vaddr, void *dst, int len) in copy_from_page() argument 236 memcpy(dst, kaddr + (vaddr & ~PAGE_MASK), len); in copy_from_page() 240 static void copy_to_page(struct page *page, unsigned long vaddr, const void *src, int len) in copy_to_page() argument 243 memcpy(kaddr + (vaddr & ~PAGE_MASK), src, len); in copy_to_page() 247 static int verify_opcode(struct page *page, unsigned long vaddr, uprobe_opcode_t *new_opcode) in verify_opcode() argument 261 copy_from_page(page, vaddr, &old_opcode, UPROBE_SWBP_INSN_SIZE); in verify_opcode() 292 int uprobe_write_opcode(struct mm_struct *mm, unsigned long vaddr, in uprobe_write_opcode() argument [all …]
|
| /linux-4.4.14/arch/s390/include/asm/ |
| D | idals.h | 29 idal_is_needed(void *vaddr, unsigned int length) in idal_is_needed() argument 31 return ((__pa(vaddr) + length - 1) >> 31) != 0; in idal_is_needed() 38 static inline unsigned int idal_nr_words(void *vaddr, unsigned int length) in idal_nr_words() argument 40 return ((__pa(vaddr) & (IDA_BLOCK_SIZE-1)) + length + in idal_nr_words() 48 void *vaddr, unsigned int length) in idal_create_words() argument 53 paddr = __pa(vaddr); in idal_create_words() 70 set_normalized_cda(struct ccw1 * ccw, void *vaddr) in set_normalized_cda() argument 77 nridaws = idal_nr_words(vaddr, ccw->count); in set_normalized_cda() 83 idal_create_words(idal, vaddr, ccw->count); in set_normalized_cda() 85 vaddr = idal; in set_normalized_cda() [all …]
|
| D | page.h | 59 #define clear_user_page(page, vaddr, pg) clear_page(page) argument 60 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument 62 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument 63 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
|
| /linux-4.4.14/arch/metag/include/asm/ |
| D | fixmap.h | 56 #define kmap_get_fixmap_pte(vaddr) \ argument 58 pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), \ 59 (vaddr) \
|
| /linux-4.4.14/arch/mips/include/asm/ |
| D | fixmap.h | 72 #define kmap_get_fixmap_pte(vaddr) \ argument 73 pte_offset_kernel(pmd_offset(pud_offset(pgd_offset_k(vaddr), (vaddr)), (vaddr)), (vaddr))
|
| D | cacheflush.h | 101 struct page *page, unsigned long vaddr, void *dst, const void *src, 105 struct page *page, unsigned long vaddr, void *dst, const void *src, 135 extern void (*__flush_kernel_vmap_range)(unsigned long vaddr, int size); 137 static inline void flush_kernel_vmap_range(void *vaddr, int size) in flush_kernel_vmap_range() argument 140 __flush_kernel_vmap_range((unsigned long) vaddr, size); in flush_kernel_vmap_range() 143 static inline void invalidate_kernel_vmap_range(void *vaddr, int size) in invalidate_kernel_vmap_range() argument 146 __flush_kernel_vmap_range((unsigned long) vaddr, size); in invalidate_kernel_vmap_range()
|
| D | tlbflush.h | 23 extern void local_flush_tlb_one(unsigned long vaddr); 33 extern void flush_tlb_one(unsigned long vaddr); 43 #define flush_tlb_one(vaddr) local_flush_tlb_one(vaddr) argument
|
| /linux-4.4.14/arch/mips/boot/ |
| D | elf2ecoff.c | 58 unsigned long vaddr; member 105 if (base->vaddr + base->len != new->vaddr) { in combine() 107 base->len = new->vaddr - base->vaddr; in combine() 282 text.vaddr = data.vaddr = bss.vaddr = 0; in main() 362 ndata.vaddr = ph[i].p_vaddr; in main() 364 nbss.vaddr = ph[i].p_vaddr + ph[i].p_filesz; in main() 372 ntxt.vaddr = ph[i].p_vaddr; in main() 392 if (text.vaddr > data.vaddr || data.vaddr > bss.vaddr || in main() 393 text.vaddr + text.len > data.vaddr in main() 394 || data.vaddr + data.len > bss.vaddr) { in main() [all …]
|
| /linux-4.4.14/arch/powerpc/platforms/44x/ |
| D | canyonlands.c | 69 void __iomem *vaddr; in ppc460ex_canyonlands_fixup() local 94 vaddr = of_iomap(np, 0); in ppc460ex_canyonlands_fixup() 97 if (!vaddr) { in ppc460ex_canyonlands_fixup() 117 setbits32((vaddr + GPIO0_OSRH), 0x42000000); in ppc460ex_canyonlands_fixup() 118 setbits32((vaddr + GPIO0_TSRH), 0x42000000); in ppc460ex_canyonlands_fixup() 120 iounmap(vaddr); in ppc460ex_canyonlands_fixup()
|
| /linux-4.4.14/arch/sh/boards/mach-sdk7786/ |
| D | sram.c | 25 void __iomem *vaddr; in fpga_sram_init() local 53 vaddr = ioremap(phys, SZ_2K); in fpga_sram_init() 54 if (unlikely(!vaddr)) { in fpga_sram_init() 63 ret = gen_pool_add(sram_pool, (unsigned long)vaddr, SZ_2K, -1); in fpga_sram_init() 66 iounmap(vaddr); in fpga_sram_init()
|
| /linux-4.4.14/arch/mips/dec/ |
| D | kn01-berr.c | 83 long asid, entryhi, vaddr; in dec_kn01_be_backend() local 105 vaddr = regs->regs[insn.i_format.rs] + in dec_kn01_be_backend() 108 vaddr = (long)pc; in dec_kn01_be_backend() 109 if (KSEGX(vaddr) == CKSEG0 || KSEGX(vaddr) == CKSEG1) in dec_kn01_be_backend() 110 address = CPHYSADDR(vaddr); in dec_kn01_be_backend() 115 entryhi |= vaddr & ~(PAGE_SIZE - 1); in dec_kn01_be_backend() 123 offset = vaddr & (PAGE_SIZE - 1); in dec_kn01_be_backend()
|
| /linux-4.4.14/arch/arc/include/asm/ |
| D | cacheflush.h | 35 void __sync_icache_dcache(phys_addr_t paddr, unsigned long vaddr, int len); 36 void __inv_icache_page(phys_addr_t paddr, unsigned long vaddr); 37 void __flush_dcache_page(phys_addr_t paddr, unsigned long vaddr); 108 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 112 __sync_icache_dcache((unsigned long)(dst), vaddr, len); \ 115 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
| D | page.h | 16 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument 96 #define __pa(vaddr) ((unsigned long)vaddr) argument
|
| /linux-4.4.14/sound/pci/asihpi/ |
| D | hpios.c | 50 p_mem_area->vaddr = in hpios_locked_mem_alloc() 54 if (p_mem_area->vaddr) { in hpios_locked_mem_alloc() 57 p_mem_area->vaddr); in hpios_locked_mem_alloc() 73 p_mem_area->vaddr, p_mem_area->dma_handle); in hpios_locked_mem_free() 77 p_mem_area->vaddr); in hpios_locked_mem_free()
|
| /linux-4.4.14/arch/ia64/kernel/ |
| D | crash_dump.c | 37 void *vaddr; in copy_oldmem_page() local 41 vaddr = __va(pfn<<PAGE_SHIFT); in copy_oldmem_page() 43 if (copy_to_user(buf, (vaddr + offset), csize)) { in copy_oldmem_page() 47 memcpy(buf, (vaddr + offset), csize); in copy_oldmem_page()
|
| /linux-4.4.14/arch/metag/kernel/ |
| D | dma.c | 225 unsigned long vaddr = c->vm_start; in dma_alloc_coherent() local 226 pte_t *pte = consistent_pte + CONSISTENT_OFFSET(vaddr); in dma_alloc_coherent() 241 set_pte_at(&init_mm, vaddr, in dma_alloc_coherent() 247 vaddr += PAGE_SIZE; in dma_alloc_coherent() 272 void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument 282 c = metag_vm_region_find(&consistent_head, (unsigned long)vaddr); in dma_free_coherent() 329 __func__, vaddr); in dma_free_coherent() 426 void dma_sync_for_device(void *vaddr, size_t size, int dma_direction) in dma_sync_for_device() argument 444 flush_dcache_region(vaddr, size); in dma_sync_for_device() 452 writeback_dcache_region(vaddr, size); in dma_sync_for_device() [all …]
|
| D | tcm.c | 51 unsigned long vaddr; in tcm_alloc() local 58 vaddr = gen_pool_alloc(pool->pool, len); in tcm_alloc() 59 if (!vaddr) in tcm_alloc() 62 return vaddr; in tcm_alloc()
|
| /linux-4.4.14/arch/sparc/prom/ |
| D | misc_64.c | 203 unsigned long tte_data, unsigned long vaddr) in tlb_load() argument 212 args[5] = vaddr; in tlb_load() 224 unsigned long vaddr) in prom_itlb_load() argument 226 return tlb_load("SUNW,itlb-load", index, tte_data, vaddr); in prom_itlb_load() 231 unsigned long vaddr) in prom_dtlb_load() argument 233 return tlb_load("SUNW,dtlb-load", index, tte_data, vaddr); in prom_dtlb_load() 237 unsigned long vaddr, unsigned long paddr) in prom_map() argument 249 args[7] = vaddr; in prom_map() 262 void prom_unmap(unsigned long size, unsigned long vaddr) in prom_unmap() argument 272 args[6] = vaddr; in prom_unmap()
|
| /linux-4.4.14/drivers/staging/rdma/ipath/ |
| D | ipath_keys.c | 143 isge->vaddr = (void *) sge->addr; in ipath_lkey_ok() 176 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in ipath_lkey_ok() 200 u32 len, u64 vaddr, u32 rkey, int acc) in ipath_rkey_ok() argument 223 sge->vaddr = (void *) vaddr; in ipath_rkey_ok() 239 off = vaddr - mr->iova; in ipath_rkey_ok() 240 if (unlikely(vaddr < mr->iova || off + len > mr->length || in ipath_rkey_ok() 258 sge->vaddr = mr->map[m]->segs[n].vaddr + off; in ipath_rkey_ok()
|
| /linux-4.4.14/arch/microblaze/include/asm/ |
| D | page.h | 81 # define clear_user_page(pgaddr, vaddr, page) memset((pgaddr), 0, PAGE_SIZE) argument 82 # define copy_user_page(vto, vfrom, vaddr, topg) \ argument 147 # define virt_to_pfn(vaddr) (phys_to_pfn((__pa(vaddr)))) argument 157 # define virt_to_page(vaddr) (pfn_to_page(virt_to_pfn(vaddr))) argument 175 #define virt_addr_valid(vaddr) (pfn_valid(virt_to_pfn(vaddr))) argument
|
| /linux-4.4.14/arch/xtensa/kernel/ |
| D | pci-dma.c | 28 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument 33 __flush_invalidate_dcache_range((unsigned long)vaddr, size); in dma_cache_sync() 37 __invalidate_dcache_range((unsigned long)vaddr, size); in dma_cache_sync() 41 __flush_dcache_range((unsigned long)vaddr, size); in dma_cache_sync() 63 void *vaddr = kmap_atomic(page); in do_cache_op() local 65 fn((unsigned long)vaddr + off, sz); in do_cache_op() 66 kunmap_atomic(vaddr); in do_cache_op() 173 static void xtensa_dma_free(struct device *hwdev, size_t size, void *vaddr, in xtensa_dma_free() argument 176 unsigned long addr = (unsigned long)vaddr + in xtensa_dma_free()
|
| /linux-4.4.14/arch/openrisc/include/asm/ |
| D | fixmap.h | 81 static inline unsigned long virt_to_fix(const unsigned long vaddr) in virt_to_fix() argument 83 BUG_ON(vaddr >= FIXADDR_TOP || vaddr < FIXADDR_START); in virt_to_fix() 84 return __virt_to_fix(vaddr); in virt_to_fix()
|
| D | page.h | 43 #define get_user_page(vaddr) __get_free_page(GFP_KERNEL) argument 49 #define clear_user_page(page, vaddr, pg) clear_page(page) argument 50 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument
|
| /linux-4.4.14/arch/nios2/mm/ |
| D | cacheflush.c | 230 void copy_user_page(void *vto, void *vfrom, unsigned long vaddr, in copy_user_page() argument 233 __flush_dcache(vaddr, vaddr + PAGE_SIZE); in copy_user_page() 234 __flush_icache(vaddr, vaddr + PAGE_SIZE); in copy_user_page() 240 void clear_user_page(void *addr, unsigned long vaddr, struct page *page) in clear_user_page() argument 242 __flush_dcache(vaddr, vaddr + PAGE_SIZE); in clear_user_page() 243 __flush_icache(vaddr, vaddr + PAGE_SIZE); in clear_user_page()
|
| D | tlb.c | 77 unsigned long vaddr = CONFIG_NIOS2_IO_REGION_BASE + in flush_tlb_one_pid() local 81 vaddr, way, (pid_misc >> TLBMISC_PID_SHIFT)); in flush_tlb_one_pid() 83 WRCTL(CTL_PTEADDR, (vaddr >> 12) << 2); in flush_tlb_one_pid() 139 unsigned long vaddr = CONFIG_NIOS2_IO_REGION_BASE + in flush_tlb_one() local 144 vaddr, way, (pid_misc >> TLBMISC_PID_SHIFT)); in flush_tlb_one() 148 WRCTL(CTL_PTEADDR, (vaddr >> 12) << 2); in flush_tlb_one() 247 unsigned long vaddr = CONFIG_NIOS2_IO_REGION_BASE; in flush_tlb_all() local 260 WRCTL(CTL_PTEADDR, ((vaddr) >> PAGE_SHIFT) << 2); in flush_tlb_all() 263 vaddr += 1UL << 12; in flush_tlb_all()
|
| /linux-4.4.14/arch/mn10300/mm/ |
| D | pgtable.c | 35 void set_pmd_pfn(unsigned long vaddr, unsigned long pfn, pgprot_t flags) in set_pmd_pfn() argument 41 if (vaddr & (PMD_SIZE-1)) { /* vaddr is misaligned */ in set_pmd_pfn() 49 pgd = swapper_pg_dir + pgd_index(vaddr); in set_pmd_pfn() 54 pud = pud_offset(pgd, vaddr); in set_pmd_pfn() 55 pmd = pmd_offset(pud, vaddr); in set_pmd_pfn() 61 local_flush_tlb_one(vaddr); in set_pmd_pfn()
|
| /linux-4.4.14/drivers/staging/android/ion/ |
| D | ion_test.c | 68 void *vaddr = vmap(&page, 1, VM_MAP, pgprot); in ion_handle_test_dma() local 72 if (!vaddr) { in ion_handle_test_dma() 78 ret = copy_from_user(vaddr + offset, ptr, to_copy); in ion_handle_test_dma() 80 ret = copy_to_user(ptr, vaddr + offset, to_copy); in ion_handle_test_dma() 82 vunmap(vaddr); in ion_handle_test_dma() 118 void *vaddr = dma_buf_kmap(dma_buf, page_offset); in ion_handle_test_kernel() local 120 if (!vaddr) in ion_handle_test_kernel() 126 ret = copy_from_user(vaddr + copy_offset, ptr, to_copy); in ion_handle_test_kernel() 128 ret = copy_to_user(ptr, vaddr + copy_offset, to_copy); in ion_handle_test_kernel() 130 dma_buf_kunmap(dma_buf, page_offset, vaddr); in ion_handle_test_kernel()
|
| /linux-4.4.14/arch/hexagon/mm/ |
| D | vm_tlb.c | 49 void flush_tlb_one(unsigned long vaddr) in flush_tlb_one() argument 51 __vmclrmap((void *)vaddr, PAGE_SIZE); in flush_tlb_one() 78 void flush_tlb_page(struct vm_area_struct *vma, unsigned long vaddr) in flush_tlb_page() argument 83 __vmclrmap((void *)vaddr, PAGE_SIZE); in flush_tlb_page()
|
| /linux-4.4.14/arch/s390/kvm/ |
| D | gaccess.c | 543 union vaddress vaddr = {.addr = gva}; in guest_translate() local 559 if (vaddr.rfx01 > asce.tl) in guest_translate() 561 ptr += vaddr.rfx * 8; in guest_translate() 564 if (vaddr.rfx) in guest_translate() 566 if (vaddr.rsx01 > asce.tl) in guest_translate() 568 ptr += vaddr.rsx * 8; in guest_translate() 571 if (vaddr.rfx || vaddr.rsx) in guest_translate() 573 if (vaddr.rtx01 > asce.tl) in guest_translate() 575 ptr += vaddr.rtx * 8; in guest_translate() 578 if (vaddr.rfx || vaddr.rsx || vaddr.rtx) in guest_translate() [all …]
|
| /linux-4.4.14/arch/arm64/include/asm/ |
| D | page.h | 50 #define clear_user_page(addr,vaddr,pg) __cpu_clear_user_page(addr, vaddr) argument 51 #define copy_user_page(to,from,vaddr,pg) __cpu_copy_user_page(to, from, vaddr) argument
|
| /linux-4.4.14/drivers/net/wireless/ath/ath6kl/ |
| D | target.h | 333 #define AR6003_VTOP(vaddr) ((vaddr) & 0x001fffff) argument 334 #define AR6004_VTOP(vaddr) (vaddr) argument 336 #define TARG_VTOP(target_type, vaddr) \ argument 337 (((target_type) == TARGET_TYPE_AR6003) ? AR6003_VTOP(vaddr) : \ 338 (((target_type) == TARGET_TYPE_AR6004) ? AR6004_VTOP(vaddr) : 0))
|
| /linux-4.4.14/arch/powerpc/platforms/512x/ |
| D | mpc512x_shared.c | 226 void __iomem *vaddr; in mpc512x_init_diu() local 250 vaddr = ioremap(desc, sizeof(struct diu_ad)); in mpc512x_init_diu() 251 if (!vaddr) { in mpc512x_init_diu() 255 memcpy(&diu_shared_fb.ad0, vaddr, sizeof(struct diu_ad)); in mpc512x_init_diu() 261 pix_fmt = in_le32(vaddr); in mpc512x_init_diu() 263 diu_shared_fb.fb_phys = in_le32(vaddr + 4); in mpc512x_init_diu() 266 iounmap(vaddr); in mpc512x_init_diu() 269 vaddr = ioremap(desc, sizeof(diu_shared_fb.gamma)); in mpc512x_init_diu() 270 if (!vaddr) { in mpc512x_init_diu() 275 memcpy(&diu_shared_fb.gamma, vaddr, sizeof(diu_shared_fb.gamma)); in mpc512x_init_diu() [all …]
|
| /linux-4.4.14/arch/arm/probes/uprobes/ |
| D | core.c | 33 unsigned long vaddr) in set_swbp() argument 35 return uprobe_write_opcode(mm, vaddr, in set_swbp() 116 void arch_uprobe_copy_ixol(struct page *page, unsigned long vaddr, in arch_uprobe_copy_ixol() argument 120 void *dst = xol_page_kaddr + (vaddr & ~PAGE_MASK); in arch_uprobe_copy_ixol() 128 flush_uprobe_xol_access(page, vaddr, dst, len); in arch_uprobe_copy_ixol() 157 regs->ARM_pc = utask->vaddr + 4; in arch_uprobe_post_xol() 178 instruction_pointer_set(regs, utask->vaddr); in arch_uprobe_abort_xol()
|
| /linux-4.4.14/drivers/staging/rdma/hfi1/ |
| D | keys.c | 203 isge->vaddr = (void *) sge->addr; in hfi1_lkey_ok() 249 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in hfi1_lkey_ok() 275 u32 len, u64 vaddr, u32 rkey, int acc) in hfi1_rkey_ok() argument 300 sge->vaddr = (void *) vaddr; in hfi1_rkey_ok() 313 off = vaddr - mr->iova; in hfi1_rkey_ok() 314 if (unlikely(vaddr < mr->iova || off + len > mr->length || in hfi1_rkey_ok() 346 sge->vaddr = mr->map[m]->segs[n].vaddr + off; in hfi1_rkey_ok()
|
| /linux-4.4.14/drivers/acpi/apei/ |
| D | ghes.c | 149 unsigned long vaddr; in ghes_ioremap_pfn_nmi() local 151 vaddr = (unsigned long)GHES_IOREMAP_NMI_PAGE(ghes_ioremap_area->addr); in ghes_ioremap_pfn_nmi() 152 ioremap_page_range(vaddr, vaddr + PAGE_SIZE, in ghes_ioremap_pfn_nmi() 155 return (void __iomem *)vaddr; in ghes_ioremap_pfn_nmi() 160 unsigned long vaddr, paddr; in ghes_ioremap_pfn_irq() local 163 vaddr = (unsigned long)GHES_IOREMAP_IRQ_PAGE(ghes_ioremap_area->addr); in ghes_ioremap_pfn_irq() 168 ioremap_page_range(vaddr, vaddr + PAGE_SIZE, paddr, prot); in ghes_ioremap_pfn_irq() 170 return (void __iomem *)vaddr; in ghes_ioremap_pfn_irq() 175 unsigned long vaddr = (unsigned long __force)vaddr_ptr; in ghes_iounmap_nmi() local 178 BUG_ON(vaddr != (unsigned long)GHES_IOREMAP_NMI_PAGE(base)); in ghes_iounmap_nmi() [all …]
|
| /linux-4.4.14/drivers/infiniband/hw/qib/ |
| D | qib_keys.c | 180 isge->vaddr = (void *) sge->addr; in qib_lkey_ok() 227 isge->vaddr = mr->map[m]->segs[n].vaddr + off; in qib_lkey_ok() 253 u32 len, u64 vaddr, u32 rkey, int acc) in qib_rkey_ok() argument 279 sge->vaddr = (void *) vaddr; in qib_rkey_ok() 292 off = vaddr - mr->iova; in qib_rkey_ok() 293 if (unlikely(vaddr < mr->iova || off + len > mr->length || in qib_rkey_ok() 326 sge->vaddr = mr->map[m]->segs[n].vaddr + off; in qib_rkey_ok() 380 mrg->map[m]->segs[n].vaddr = (void *) page_list[i]; in qib_reg_mr()
|
| /linux-4.4.14/arch/alpha/mm/ |
| D | init.c | 193 unsigned long vaddr; in callback_init() local 205 vaddr = (unsigned long)console_remap_vm.addr; in callback_init() 211 crb->map[i].va = vaddr; in callback_init() 216 if (pmd != pmd_offset(pgd, vaddr)) { in callback_init() 218 pmd = pmd_offset(pgd, vaddr); in callback_init() 222 set_pte(pte_offset_kernel(pmd, vaddr), in callback_init() 225 vaddr += PAGE_SIZE; in callback_init()
|
| /linux-4.4.14/arch/x86/xen/ |
| D | grant-table.c | 121 void *vaddr; in xlated_setup_gnttab_pages() local 147 vaddr = vmap(pages, nr_grant_frames, 0, PAGE_KERNEL); in xlated_setup_gnttab_pages() 148 if (!vaddr) { in xlated_setup_gnttab_pages() 160 xen_auto_xlat_grant_frames.vaddr = vaddr; in xlated_setup_gnttab_pages()
|
| D | mmu.c | 127 unsigned long arbitrary_virt_to_mfn(void *vaddr) in arbitrary_virt_to_mfn() argument 129 xmaddr_t maddr = arbitrary_virt_to_machine(vaddr); in arbitrary_virt_to_mfn() 134 xmaddr_t arbitrary_virt_to_machine(void *vaddr) in arbitrary_virt_to_machine() argument 136 unsigned long address = (unsigned long)vaddr; in arbitrary_virt_to_machine() 145 if (virt_addr_valid(vaddr)) in arbitrary_virt_to_machine() 146 return virt_to_machine(vaddr); in arbitrary_virt_to_machine() 157 void make_lowmem_page_readonly(void *vaddr) in make_lowmem_page_readonly() argument 160 unsigned long address = (unsigned long)vaddr; in make_lowmem_page_readonly() 173 void make_lowmem_page_readwrite(void *vaddr) in make_lowmem_page_readwrite() argument 176 unsigned long address = (unsigned long)vaddr; in make_lowmem_page_readwrite() [all …]
|
| /linux-4.4.14/drivers/gpu/drm/ |
| D | drm_gem_cma_helper.c | 110 cma_obj->vaddr = dma_alloc_writecombine(drm->dev, size, in drm_gem_cma_create() 112 if (!cma_obj->vaddr) { in drm_gem_cma_create() 192 if (cma_obj->vaddr) { in drm_gem_cma_free_object() 194 cma_obj->vaddr, cma_obj->paddr); in drm_gem_cma_free_object() 326 cma_obj->vaddr, cma_obj->paddr, in drm_gem_cma_mmap_obj() 385 off, &cma_obj->paddr, cma_obj->vaddr, obj->size); in drm_gem_cma_describe() 414 ret = dma_get_sgtable(obj->dev->dev, sgt, cma_obj->vaddr, in drm_gem_cma_prime_get_sg_table() 513 return cma_obj->vaddr; in drm_gem_cma_prime_vmap() 528 void drm_gem_cma_prime_vunmap(struct drm_gem_object *obj, void *vaddr) in drm_gem_cma_prime_vunmap() argument
|
| D | drm_pci.c | 60 dmah->vaddr = dma_alloc_coherent(&dev->pdev->dev, size, &dmah->busaddr, GFP_KERNEL | __GFP_COMP); in drm_pci_alloc() 62 if (dmah->vaddr == NULL) { in drm_pci_alloc() 67 memset(dmah->vaddr, 0, size); in drm_pci_alloc() 71 for (addr = (unsigned long)dmah->vaddr, sz = size; in drm_pci_alloc() 91 if (dmah->vaddr) { in __drm_legacy_pci_free() 94 for (addr = (unsigned long)dmah->vaddr, sz = dmah->size; in __drm_legacy_pci_free() 98 dma_free_coherent(&dev->pdev->dev, dmah->size, dmah->vaddr, in __drm_legacy_pci_free()
|
| /linux-4.4.14/include/asm-generic/ |
| D | fixmap.h | 35 static inline unsigned long virt_to_fix(const unsigned long vaddr) in virt_to_fix() argument 37 BUG_ON(vaddr >= FIXADDR_TOP || vaddr < FIXADDR_START); in virt_to_fix() 38 return __virt_to_fix(vaddr); in virt_to_fix()
|
| D | dma-coherent.h | 11 int dma_release_from_coherent(struct device *dev, int order, void *vaddr); 28 #define dma_release_from_coherent(dev, order, vaddr) (0) argument 29 #define dma_mmap_from_coherent(dev, vma, vaddr, order, ret) (0) argument
|
| D | cacheflush.h | 26 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 29 flush_icache_user_range(vma, page, vaddr, len); \ 31 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
| /linux-4.4.14/drivers/media/platform/coda/ |
| D | coda-jpeg.c | 172 coda_memcpy_parabuf(ctx->parabuf.vaddr, huff + i); in coda_jpeg_write_tables() 176 coda_memcpy_parabuf(ctx->parabuf.vaddr, qmat + i); in coda_jpeg_write_tables() 183 void *vaddr = vb2_plane_vaddr(&vb->vb2_buf, 0); in coda_jpeg_check_buffer() local 184 u16 soi = be16_to_cpup((__be16 *)vaddr); in coda_jpeg_check_buffer() 185 u16 eoi = be16_to_cpup((__be16 *)(vaddr + in coda_jpeg_check_buffer()
|
| /linux-4.4.14/drivers/net/wireless/orinoco/ |
| D | airport.c | 29 void __iomem *vaddr; member 101 if (card->vaddr) in airport_detach() 102 iounmap(card->vaddr); in airport_detach() 103 card->vaddr = NULL; in airport_detach() 184 card->vaddr = ioremap(phys_addr, AIRPORT_IO_LEN); in airport_attach() 185 if (!card->vaddr) { in airport_attach() 190 hermes_struct_init(hw, card->vaddr, HERMES_16BIT_REGSPACING); in airport_attach()
|
| /linux-4.4.14/arch/cris/include/asm/ |
| D | page.h | 15 #define clear_user_page(page, vaddr, pg) clear_page(page) argument 16 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument 18 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument 19 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
|
| /linux-4.4.14/drivers/vfio/ |
| D | vfio_iommu_type1.c | 75 unsigned long vaddr; /* Process virtual addr */ member 231 static int vaddr_get_pfn(unsigned long vaddr, int prot, unsigned long *pfn) in vaddr_get_pfn() argument 237 if (get_user_pages_fast(vaddr, 1, !!(prot & IOMMU_WRITE), page) == 1) { in vaddr_get_pfn() 244 vma = find_vma_intersection(current->mm, vaddr, vaddr + 1); in vaddr_get_pfn() 247 *pfn = ((vaddr - vma->vm_start) >> PAGE_SHIFT) + vma->vm_pgoff; in vaddr_get_pfn() 262 static long vfio_pin_pages(unsigned long vaddr, long npage, in vfio_pin_pages() argument 273 ret = vaddr_get_pfn(vaddr, prot, pfn_base); in vfio_pin_pages() 293 for (i = 1, vaddr += PAGE_SIZE; i < npage; i++, vaddr += PAGE_SIZE) { in vfio_pin_pages() 296 ret = vaddr_get_pfn(vaddr, prot, &pfn); in vfio_pin_pages() 565 unsigned long vaddr = map->vaddr; in vfio_dma_do_map() local [all …]
|
| /linux-4.4.14/arch/powerpc/include/asm/ |
| D | dma-mapping.h | 30 void *vaddr, dma_addr_t dma_handle, 48 extern void __dma_free_coherent(size_t size, void *vaddr); 49 extern void __dma_sync(void *vaddr, size_t size, int direction); 160 static inline void dma_cache_sync(struct device *dev, void *vaddr, size_t size, in dma_cache_sync() argument 164 __dma_sync(vaddr, size, (int)direction); in dma_cache_sync()
|
| D | cacheflush.h | 60 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 63 flush_icache_user_range(vma, page, vaddr, len); \ 65 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
| /linux-4.4.14/arch/avr32/mm/ |
| D | dma-coherent.c | 16 void dma_cache_sync(struct device *dev, void *vaddr, size_t size, int direction) in dma_cache_sync() argument 21 if (PXSEG(vaddr) == P2SEG) in dma_cache_sync() 26 invalidate_dcache_region(vaddr, size); in dma_cache_sync() 29 clean_dcache_region(vaddr, size); in dma_cache_sync() 32 flush_dcache_region(vaddr, size); in dma_cache_sync()
|
| /linux-4.4.14/drivers/i2c/busses/ |
| D | i2c-ibm_iic.c | 85 volatile struct iic_regs __iomem *iic = dev->vaddr; in dump_iic_regs() 129 out_8(&dev->vaddr->intmsk, enable ? INTRMSK_EIMTC : 0); in iic_interrupt_mode() 137 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_dev_init() 182 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_dev_reset() 243 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_smbus_quick() 329 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_handler() 347 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_xfer_result() 380 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_abort_xfer() 412 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_wait_for_tc() 463 volatile struct iic_regs __iomem *iic = dev->vaddr; in iic_xfer_bytes() [all …]
|
| /linux-4.4.14/arch/m32r/include/asm/ |
| D | page.h | 16 #define clear_user_page(page, vaddr, pg) clear_page(page) argument 17 #define copy_user_page(to, from, vaddr, pg) copy_page(to, from) argument 19 #define __alloc_zeroed_user_highpage(movableflags, vma, vaddr) \ argument 20 alloc_page_vma(GFP_HIGHUSER | __GFP_ZERO | movableflags, vma, vaddr)
|
| /linux-4.4.14/arch/cris/arch-v32/drivers/pci/ |
| D | dma.c | 43 void *vaddr, dma_addr_t dma_handle) in dma_free_coherent() argument 47 if (!dma_release_from_coherent(dev, order, vaddr)) in dma_free_coherent() 48 free_pages((unsigned long)vaddr, order); in dma_free_coherent()
|
| /linux-4.4.14/arch/alpha/boot/ |
| D | bootpz.c | 79 unsigned long vaddr, kaddr; in check_range() local 86 for (vaddr = vstart; vaddr <= vend; vaddr += PAGE_SIZE) in check_range() 88 kaddr = (find_pa(vaddr) | PAGE_OFFSET); in check_range() 94 vaddr, kaddr, kstart, kend); in check_range()
|
| /linux-4.4.14/fs/minix/ |
| D | minix.h | 124 static inline int minix_find_first_zero_bit(const void *vaddr, unsigned size) in minix_find_first_zero_bit() argument 126 const unsigned short *p = vaddr, *addr = vaddr; in minix_find_first_zero_bit() 149 static inline int minix_test_bit(int nr, const void *vaddr) in minix_test_bit() argument 151 const unsigned short *p = vaddr; in minix_test_bit()
|
| /linux-4.4.14/arch/alpha/kernel/ |
| D | core_irongate.c | 318 unsigned long vaddr; in irongate_ioremap() local 384 for(baddr = addr, vaddr = (unsigned long)area->addr; in irongate_ioremap() 386 baddr += PAGE_SIZE, vaddr += PAGE_SIZE) in irongate_ioremap() 391 if (__alpha_remap_area_pages(vaddr, in irongate_ioremap() 401 vaddr = (unsigned long)area->addr + (addr & ~PAGE_MASK); in irongate_ioremap() 404 addr, size, vaddr); in irongate_ioremap() 406 return (void __iomem *)vaddr; in irongate_ioremap()
|
| /linux-4.4.14/arch/ia64/include/asm/ |
| D | cacheflush.h | 47 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 49 flush_icache_user_range(vma, page, vaddr, len); \ 51 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
| /linux-4.4.14/drivers/tty/serial/8250/ |
| D | 8250_acorn.c | 38 void __iomem *vaddr; member 57 info->vaddr = ecardm_iomap(ec, type->type, 0, 0); in serial_card_probe() 58 if (!info->vaddr) { in serial_card_probe() 74 uart.port.membase = info->vaddr + type->offset[i]; in serial_card_probe()
|
| /linux-4.4.14/drivers/scsi/sym53c8xx_2/ |
| D | sym_malloc.c | 226 void *vaddr; in ___get_dma_mem_cluster() local 232 vaddr = sym_m_get_dma_mem_cluster(mp, vbp); in ___get_dma_mem_cluster() 233 if (vaddr) { in ___get_dma_mem_cluster() 234 int hc = VTOB_HASH_CODE(vaddr); in ___get_dma_mem_cluster() 239 return vaddr; in ___get_dma_mem_cluster() 252 while (*vbpp && (*vbpp)->vaddr != m) in ___free_dma_mem_cluster() 370 while (vp && vp->vaddr != a) in __vtobus()
|
| /linux-4.4.14/arch/alpha/include/asm/ |
| D | cacheflush.h | 70 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument 72 flush_icache_user_range(vma, page, vaddr, len); \ 74 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
| /linux-4.4.14/drivers/media/pci/cx23885/ |
| D | cx23885-alsa.c | 89 buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT); in cx23885_alsa_dma_init() 90 if (NULL == buf->vaddr) { in cx23885_alsa_dma_init() 96 (unsigned long)buf->vaddr, in cx23885_alsa_dma_init() 99 memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT); in cx23885_alsa_dma_init() 108 pg = vmalloc_to_page(buf->vaddr + i * PAGE_SIZE); in cx23885_alsa_dma_init() 119 vfree(buf->vaddr); in cx23885_alsa_dma_init() 120 buf->vaddr = NULL; in cx23885_alsa_dma_init() 154 vfree(buf->vaddr); in cx23885_alsa_dma_free() 155 buf->vaddr = NULL; in cx23885_alsa_dma_free() 416 substream->runtime->dma_area = chip->buf->vaddr; in snd_cx23885_hw_params()
|
| /linux-4.4.14/drivers/video/fbdev/matrox/ |
| D | matroxfb_base.h | 125 void __iomem* vaddr; member 129 return readb(va.vaddr + offs); in mga_readb() 133 writeb(value, va.vaddr + offs); in mga_writeb() 137 writew(value, va.vaddr + offs); in mga_writew() 141 return readl(va.vaddr + offs); in mga_readl() 145 writel(value, va.vaddr + offs); in mga_writel() 157 iowrite32_rep(va.vaddr, src, len >> 2); in mga_memcpy_toio() 159 u_int32_t __iomem* addr = va.vaddr; in mga_memcpy_toio() 180 va->vaddr += offs; in vaddr_add() 184 return va.vaddr; in vaddr_va()
|
| /linux-4.4.14/drivers/parisc/ |
| D | iommu-helpers.h | 30 unsigned long vaddr; in iommu_fill_pdir() local 69 vaddr = (unsigned long)sg_virt(startsg); in iommu_fill_pdir() 78 vaddr, hint); in iommu_fill_pdir() 79 vaddr += IOVP_SIZE; in iommu_fill_pdir()
|
| /linux-4.4.14/arch/unicore32/include/asm/ |
| D | cacheflush.h | 19 #define CACHE_COLOUR(vaddr) ((vaddr & (SHMLBA - 1)) >> PAGE_SHIFT) argument 120 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
|
| /linux-4.4.14/fs/pstore/ |
| D | ram_core.c | 391 void *vaddr; in persistent_ram_vmap() local 412 vaddr = vmap(pages, page_count, VM_MAP, prot); in persistent_ram_vmap() 415 return vaddr; in persistent_ram_vmap() 447 prz->vaddr = persistent_ram_vmap(start, size, memtype); in persistent_ram_buffer_map() 449 prz->vaddr = persistent_ram_iomap(start, size, memtype); in persistent_ram_buffer_map() 451 if (!prz->vaddr) { in persistent_ram_buffer_map() 457 prz->buffer = prz->vaddr + offset_in_page(start); in persistent_ram_buffer_map() 501 if (prz->vaddr) { in persistent_ram_free() 503 vunmap(prz->vaddr); in persistent_ram_free() 505 iounmap(prz->vaddr); in persistent_ram_free() [all …]
|
| /linux-4.4.14/drivers/net/wireless/ath/ath10k/ |
| D | htt_tx.c | 112 htt->txbuf.vaddr = dma_alloc_coherent(ar->dev, size, in ath10k_htt_tx_alloc() 115 if (!htt->txbuf.vaddr) { in ath10k_htt_tx_alloc() 125 htt->frag_desc.vaddr = dma_alloc_coherent(ar->dev, size, in ath10k_htt_tx_alloc() 128 if (!htt->frag_desc.vaddr) { in ath10k_htt_tx_alloc() 140 dma_free_coherent(htt->ar->dev, size, htt->txbuf.vaddr, in ath10k_htt_tx_alloc() 170 if (htt->txbuf.vaddr) { in ath10k_htt_tx_free() 173 dma_free_coherent(htt->ar->dev, size, htt->txbuf.vaddr, in ath10k_htt_tx_free() 177 if (htt->frag_desc.vaddr) { in ath10k_htt_tx_free() 180 dma_free_coherent(htt->ar->dev, size, htt->frag_desc.vaddr, in ath10k_htt_tx_free() 361 fw_idx = __le32_to_cpu(*htt->rx_ring.alloc_idx.vaddr); in ath10k_htt_send_rx_ring_cfg_ll() [all …]
|
| /linux-4.4.14/kernel/trace/ |
| D | trace_uprobe.c | 33 unsigned long vaddr[]; member 127 void __user *vaddr = (void __force __user *) addr; \ 129 if (copy_from_user(&retval, vaddr, sizeof(type))) \ 165 void __user *vaddr = (void __force __user *) addr; in FETCH_FUNC_NAME() local 167 len = strnlen_user(vaddr, MAX_STRING_SIZE); in FETCH_FUNC_NAME() 180 udd = (void *) current->utask->vaddr; in translate_user_vaddr() 190 void *vaddr = (void *)translate_user_vaddr(offset); \ 192 FETCH_FUNC_NAME(memory, type)(regs, vaddr, dest); \ 814 entry->vaddr[0] = func; in __uprobe_trace_func() 815 entry->vaddr[1] = instruction_pointer(regs); in __uprobe_trace_func() [all …]
|
| /linux-4.4.14/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
| D | gk20a.c | 58 u32 *vaddr; member 182 vunmap(obj->vaddr); in gk20a_instmem_vaddr_gc() 183 obj->vaddr = NULL; in gk20a_instmem_vaddr_gc() 204 if (node->vaddr) { in gk20a_instobj_acquire() 214 node->vaddr = imem->cpu_map(memory); in gk20a_instobj_acquire() 216 if (!node->vaddr) { in gk20a_instobj_acquire() 229 return node->vaddr; in gk20a_instobj_acquire() 256 return node->vaddr[offset / 4]; in gk20a_instobj_rd32() 264 node->vaddr[offset / 4] = data; in gk20a_instobj_wr32() 287 if (!node->vaddr) in gk20a_instobj_dtor() [all …]
|
| /linux-4.4.14/drivers/gpio/ |
| D | gpio-ts5500.c | 64 #define TS5500_DIO_IN_OUT(vaddr, vbit, caddr, cbit) \ argument 66 .value_addr = vaddr, \ 99 #define TS5500_DIO_GROUP(vaddr, vbitfrom, caddr, cbit) \ argument 100 TS5500_DIO_IN_OUT(vaddr, vbitfrom + 0, caddr, cbit), \ 101 TS5500_DIO_IN_OUT(vaddr, vbitfrom + 1, caddr, cbit), \ 102 TS5500_DIO_IN_OUT(vaddr, vbitfrom + 2, caddr, cbit), \ 103 TS5500_DIO_IN_OUT(vaddr, vbitfrom + 3, caddr, cbit)
|
| /linux-4.4.14/arch/m32r/mm/ |
| D | fault.c | 348 void update_mmu_cache(struct vm_area_struct *vma, unsigned long vaddr, in update_mmu_cache() argument 363 vaddr = (vaddr & PAGE_MASK) | get_asid(); in update_mmu_cache() 370 if (*entry1++ == vaddr) { in update_mmu_cache() 378 if (*entry2++ == vaddr) { in update_mmu_cache() 405 : "r" (vaddr), "r" (pte_data), "i" (MMU_REG_BASE), in update_mmu_cache() 449 *entry1++ = vaddr; /* Set TLB tag */ in update_mmu_cache()
|
| /linux-4.4.14/drivers/hwtracing/coresight/ |
| D | coresight-tmc.c | 127 void __iomem *vaddr; member 199 memset(drvdata->vaddr, 0, drvdata->size); in tmc_etr_enable_hw() 336 drvdata->buf = drvdata->vaddr + rwp - drvdata->paddr; in tmc_etr_dump_hw() 338 drvdata->buf = drvdata->vaddr; in tmc_etr_dump_hw() 516 if (bufp == (char *)(drvdata->vaddr + drvdata->size)) in tmc_read() 517 bufp = drvdata->vaddr; in tmc_read() 518 else if (bufp > (char *)(drvdata->vaddr + drvdata->size)) in tmc_read() 520 if ((bufp + len) > (char *)(drvdata->vaddr + drvdata->size)) in tmc_read() 521 len = (char *)(drvdata->vaddr + drvdata->size) - bufp; in tmc_read() 710 drvdata->vaddr = dma_alloc_coherent(dev, drvdata->size, in tmc_probe() [all …]
|
| /linux-4.4.14/arch/arm/plat-samsung/ |
| D | pm-debug.c | 59 unsigned long vaddr; in s3c_pm_uart_base() local 61 debug_ll_addr(&paddr, &vaddr); in s3c_pm_uart_base() 63 return (void __iomem *)vaddr; in s3c_pm_uart_base()
|
| /linux-4.4.14/crypto/ |
| D | scatterwalk.c | 85 u8 *vaddr; in scatterwalk_copychunks() local 90 vaddr = scatterwalk_map(walk); in scatterwalk_copychunks() 91 memcpy_dir(buf, vaddr, len_this_page, out); in scatterwalk_copychunks() 92 scatterwalk_unmap(vaddr); in scatterwalk_copychunks()
|
| /linux-4.4.14/arch/arm64/mm/ |
| D | copypage.c | 25 void __cpu_copy_user_page(void *kto, const void *kfrom, unsigned long vaddr) in __cpu_copy_user_page() argument 32 void __cpu_clear_user_page(void *kaddr, unsigned long vaddr) in __cpu_clear_user_page() argument
|
| /linux-4.4.14/drivers/media/platform/exynos4-is/ |
| D | fimc-is.c | 241 buf = is->memory.vaddr + is->setfile.base; in fimc_is_load_setfile() 246 pr_debug("mem vaddr: %p, setfile buf: %p\n", is->memory.vaddr, buf); in fimc_is_load_setfile() 317 memcpy(is->memory.vaddr, is->fw.f_w->data, is->fw.f_w->size); in fimc_is_start_firmware() 337 is->memory.vaddr = dma_alloc_coherent(dev, FIMC_IS_CPU_MEM_SIZE, in fimc_is_alloc_cpu_memory() 339 if (is->memory.vaddr == NULL) in fimc_is_alloc_cpu_memory() 343 memset(is->memory.vaddr, 0, is->memory.size); in fimc_is_alloc_cpu_memory() 350 dma_free_coherent(dev, is->memory.size, is->memory.vaddr, in fimc_is_alloc_cpu_memory() 355 is->is_p_region = (struct is_region *)(is->memory.vaddr + in fimc_is_alloc_cpu_memory() 361 is->is_shared_region = (struct is_share_region *)(is->memory.vaddr + in fimc_is_alloc_cpu_memory() 370 if (is->memory.vaddr == NULL) in fimc_is_free_cpu_memory() [all …]
|