Lines Matching refs:page

66 	struct page *page;  in setup_zero_pages()  local
77 page = virt_to_page((void *)empty_zero_page); in setup_zero_pages()
78 split_page(page, order); in setup_zero_pages()
79 for (i = 0; i < (1 << order); i++, page++) in setup_zero_pages()
80 mark_page_reserved(page); in setup_zero_pages()
85 static void *__kmap_pgprot(struct page *page, unsigned long addr, pgprot_t prot) in __kmap_pgprot() argument
93 BUG_ON(Page_dcache_dirty(page)); in __kmap_pgprot()
100 pte = mk_pte(page, prot); in __kmap_pgprot()
129 void *kmap_coherent(struct page *page, unsigned long addr) in kmap_coherent() argument
131 return __kmap_pgprot(page, addr, PAGE_KERNEL); in kmap_coherent()
134 void *kmap_noncoherent(struct page *page, unsigned long addr) in kmap_noncoherent() argument
136 return __kmap_pgprot(page, addr, PAGE_KERNEL_NC); in kmap_noncoherent()
161 void copy_user_highpage(struct page *to, struct page *from, in copy_user_highpage()
186 struct page *page, unsigned long vaddr, void *dst, const void *src, in copy_to_user_page() argument
190 page_mapped(page) && !Page_dcache_dirty(page)) { in copy_to_user_page()
191 void *vto = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_to_user_page()
197 SetPageDcacheDirty(page); in copy_to_user_page()
200 flush_cache_page(vma, vaddr, page_to_pfn(page)); in copy_to_user_page()
204 struct page *page, unsigned long vaddr, void *dst, const void *src, in copy_from_user_page() argument
208 page_mapped(page) && !Page_dcache_dirty(page)) { in copy_from_user_page()
209 void *vfrom = kmap_coherent(page, vaddr) + (vaddr & ~PAGE_MASK); in copy_from_user_page()
215 SetPageDcacheDirty(page); in copy_from_user_page()
441 struct page *page = pfn_to_page(tmp); in mem_init_free_highmem() local
444 SetPageReserved(page); in mem_init_free_highmem()
446 free_highmem_page(page); in mem_init_free_highmem()
484 struct page *page = pfn_to_page(pfn); in free_init_pages() local
488 free_reserved_page(page); in free_init_pages()