Lines Matching refs:page
41 static inline void flush_kernel_dcache_page(struct page *page) in flush_kernel_dcache_page() argument
43 flush_kernel_dcache_page_addr(page_address(page)); in flush_kernel_dcache_page()
63 struct page *page = vmalloc_to_page(cursor); in invalidate_kernel_vmap_range() local
65 if (test_and_clear_bit(PG_dcache_dirty, &page->flags)) in invalidate_kernel_vmap_range()
66 flush_kernel_dcache_page(page); in invalidate_kernel_vmap_range()
75 extern void flush_dcache_page(struct page *page);
82 #define flush_icache_page(vma,page) do { \ argument
83 flush_kernel_dcache_page(page); \
84 flush_kernel_icache_page(page_address(page)); \
92 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ argument
94 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
99 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ argument
101 flush_cache_page(vma, vaddr, page_to_pfn(page)); \
114 flush_anon_page(struct vm_area_struct *vma, struct page *page, unsigned long vmaddr) in flush_anon_page() argument
116 if (PageAnon(page)) { in flush_anon_page()
119 flush_dcache_page_asm(page_to_phys(page), vmaddr); in flush_anon_page()
132 static inline void *kmap(struct page *page) in kmap() argument
135 return page_address(page); in kmap()
138 static inline void kunmap(struct page *page) in kunmap() argument
140 flush_kernel_dcache_page_addr(page_address(page)); in kunmap()
143 static inline void *kmap_atomic(struct page *page) in kmap_atomic() argument
146 return page_address(page); in kmap_atomic()
155 #define kmap_atomic_prot(page, prot) kmap_atomic(page) argument