Lines Matching refs:page
113 struct page *page, void *shadow) in page_cache_tree_delete() argument
121 VM_BUG_ON(!PageLocked(page)); in page_cache_tree_delete()
123 __radix_tree_lookup(&mapping->page_tree, page->index, &node, &slot); in page_cache_tree_delete()
145 index = page->index; in page_cache_tree_delete()
181 void __delete_from_page_cache(struct page *page, void *shadow, in __delete_from_page_cache() argument
184 struct address_space *mapping = page->mapping; in __delete_from_page_cache()
186 trace_mm_filemap_delete_from_page_cache(page); in __delete_from_page_cache()
192 if (PageUptodate(page) && PageMappedToDisk(page)) in __delete_from_page_cache()
193 cleancache_put_page(page); in __delete_from_page_cache()
195 cleancache_invalidate_page(mapping, page); in __delete_from_page_cache()
197 page_cache_tree_delete(mapping, page, shadow); in __delete_from_page_cache()
199 page->mapping = NULL; in __delete_from_page_cache()
203 if (!PageHuge(page)) in __delete_from_page_cache()
204 __dec_zone_page_state(page, NR_FILE_PAGES); in __delete_from_page_cache()
205 if (PageSwapBacked(page)) in __delete_from_page_cache()
206 __dec_zone_page_state(page, NR_SHMEM); in __delete_from_page_cache()
207 BUG_ON(page_mapped(page)); in __delete_from_page_cache()
217 if (WARN_ON_ONCE(PageDirty(page))) in __delete_from_page_cache()
218 account_page_cleaned(page, mapping, memcg, in __delete_from_page_cache()
230 void delete_from_page_cache(struct page *page) in delete_from_page_cache() argument
232 struct address_space *mapping = page->mapping; in delete_from_page_cache()
236 void (*freepage)(struct page *); in delete_from_page_cache()
238 BUG_ON(!PageLocked(page)); in delete_from_page_cache()
242 memcg = mem_cgroup_begin_page_stat(page); in delete_from_page_cache()
244 __delete_from_page_cache(page, NULL, memcg); in delete_from_page_cache()
249 freepage(page); in delete_from_page_cache()
250 page_cache_release(page); in delete_from_page_cache()
354 struct page *page = pvec.pages[i]; in __filemap_fdatawait_range() local
357 if (page->index > end) in __filemap_fdatawait_range()
360 wait_on_page_writeback(page); in __filemap_fdatawait_range()
361 if (TestClearPageError(page)) in __filemap_fdatawait_range()
516 int replace_page_cache_page(struct page *old, struct page *new, gfp_t gfp_mask) in replace_page_cache_page()
527 void (*freepage)(struct page *); in replace_page_cache_page()
566 struct page *page, void **shadowp) in page_cache_tree_insert() argument
572 error = __radix_tree_create(&mapping->page_tree, page->index, in page_cache_tree_insert()
588 radix_tree_replace_slot(slot, page); in page_cache_tree_insert()
607 static int __add_to_page_cache_locked(struct page *page, in __add_to_page_cache_locked() argument
612 int huge = PageHuge(page); in __add_to_page_cache_locked()
616 VM_BUG_ON_PAGE(!PageLocked(page), page); in __add_to_page_cache_locked()
617 VM_BUG_ON_PAGE(PageSwapBacked(page), page); in __add_to_page_cache_locked()
620 error = mem_cgroup_try_charge(page, current->mm, in __add_to_page_cache_locked()
629 mem_cgroup_cancel_charge(page, memcg); in __add_to_page_cache_locked()
633 page_cache_get(page); in __add_to_page_cache_locked()
634 page->mapping = mapping; in __add_to_page_cache_locked()
635 page->index = offset; in __add_to_page_cache_locked()
638 error = page_cache_tree_insert(mapping, page, shadowp); in __add_to_page_cache_locked()
645 __inc_zone_page_state(page, NR_FILE_PAGES); in __add_to_page_cache_locked()
648 mem_cgroup_commit_charge(page, memcg, false); in __add_to_page_cache_locked()
649 trace_mm_filemap_add_to_page_cache(page); in __add_to_page_cache_locked()
652 page->mapping = NULL; in __add_to_page_cache_locked()
656 mem_cgroup_cancel_charge(page, memcg); in __add_to_page_cache_locked()
657 page_cache_release(page); in __add_to_page_cache_locked()
671 int add_to_page_cache_locked(struct page *page, struct address_space *mapping, in add_to_page_cache_locked() argument
674 return __add_to_page_cache_locked(page, mapping, offset, in add_to_page_cache_locked()
679 int add_to_page_cache_lru(struct page *page, struct address_space *mapping, in add_to_page_cache_lru() argument
685 __set_page_locked(page); in add_to_page_cache_lru()
686 ret = __add_to_page_cache_locked(page, mapping, offset, in add_to_page_cache_lru()
689 __clear_page_locked(page); in add_to_page_cache_lru()
697 SetPageActive(page); in add_to_page_cache_lru()
698 workingset_activation(page); in add_to_page_cache_lru()
700 ClearPageActive(page); in add_to_page_cache_lru()
701 lru_cache_add(page); in add_to_page_cache_lru()
708 struct page *__page_cache_alloc(gfp_t gfp) in __page_cache_alloc()
711 struct page *page; in __page_cache_alloc() local
718 page = __alloc_pages_node(n, gfp, 0); in __page_cache_alloc()
719 } while (!page && read_mems_allowed_retry(cpuset_mems_cookie)); in __page_cache_alloc()
721 return page; in __page_cache_alloc()
738 wait_queue_head_t *page_waitqueue(struct page *page) in page_waitqueue() argument
740 const struct zone *zone = page_zone(page); in page_waitqueue()
742 return &zone->wait_table[hash_ptr(page, zone->wait_table_bits)]; in page_waitqueue()
746 void wait_on_page_bit(struct page *page, int bit_nr) in wait_on_page_bit() argument
748 DEFINE_WAIT_BIT(wait, &page->flags, bit_nr); in wait_on_page_bit()
750 if (test_bit(bit_nr, &page->flags)) in wait_on_page_bit()
751 __wait_on_bit(page_waitqueue(page), &wait, bit_wait_io, in wait_on_page_bit()
756 int wait_on_page_bit_killable(struct page *page, int bit_nr) in wait_on_page_bit_killable() argument
758 DEFINE_WAIT_BIT(wait, &page->flags, bit_nr); in wait_on_page_bit_killable()
760 if (!test_bit(bit_nr, &page->flags)) in wait_on_page_bit_killable()
763 return __wait_on_bit(page_waitqueue(page), &wait, in wait_on_page_bit_killable()
767 int wait_on_page_bit_killable_timeout(struct page *page, in wait_on_page_bit_killable_timeout() argument
770 DEFINE_WAIT_BIT(wait, &page->flags, bit_nr); in wait_on_page_bit_killable_timeout()
773 if (!test_bit(bit_nr, &page->flags)) in wait_on_page_bit_killable_timeout()
775 return __wait_on_bit(page_waitqueue(page), &wait, in wait_on_page_bit_killable_timeout()
787 void add_page_wait_queue(struct page *page, wait_queue_t *waiter) in add_page_wait_queue() argument
789 wait_queue_head_t *q = page_waitqueue(page); in add_page_wait_queue()
810 void unlock_page(struct page *page) in unlock_page() argument
812 VM_BUG_ON_PAGE(!PageLocked(page), page); in unlock_page()
813 clear_bit_unlock(PG_locked, &page->flags); in unlock_page()
815 wake_up_page(page, PG_locked); in unlock_page()
823 void end_page_writeback(struct page *page) in end_page_writeback() argument
832 if (PageReclaim(page)) { in end_page_writeback()
833 ClearPageReclaim(page); in end_page_writeback()
834 rotate_reclaimable_page(page); in end_page_writeback()
837 if (!test_clear_page_writeback(page)) in end_page_writeback()
841 wake_up_page(page, PG_writeback); in end_page_writeback()
849 void page_endio(struct page *page, int rw, int err) in page_endio() argument
853 SetPageUptodate(page); in page_endio()
855 ClearPageUptodate(page); in page_endio()
856 SetPageError(page); in page_endio()
858 unlock_page(page); in page_endio()
861 SetPageError(page); in page_endio()
862 if (page->mapping) in page_endio()
863 mapping_set_error(page->mapping, err); in page_endio()
865 end_page_writeback(page); in page_endio()
874 void __lock_page(struct page *page) in __lock_page() argument
876 DEFINE_WAIT_BIT(wait, &page->flags, PG_locked); in __lock_page()
878 __wait_on_bit_lock(page_waitqueue(page), &wait, bit_wait_io, in __lock_page()
883 int __lock_page_killable(struct page *page) in __lock_page_killable() argument
885 DEFINE_WAIT_BIT(wait, &page->flags, PG_locked); in __lock_page_killable()
887 return __wait_on_bit_lock(page_waitqueue(page), &wait, in __lock_page_killable()
903 int __lock_page_or_retry(struct page *page, struct mm_struct *mm, in __lock_page_or_retry() argument
916 wait_on_page_locked_killable(page); in __lock_page_or_retry()
918 wait_on_page_locked(page); in __lock_page_or_retry()
924 ret = __lock_page_killable(page); in __lock_page_or_retry()
930 __lock_page(page); in __lock_page_or_retry()
962 struct page *page; in page_cache_next_hole() local
964 page = radix_tree_lookup(&mapping->page_tree, index); in page_cache_next_hole()
965 if (!page || radix_tree_exceptional_entry(page)) in page_cache_next_hole()
1003 struct page *page; in page_cache_prev_hole() local
1005 page = radix_tree_lookup(&mapping->page_tree, index); in page_cache_prev_hole()
1006 if (!page || radix_tree_exceptional_entry(page)) in page_cache_prev_hole()
1030 struct page *find_get_entry(struct address_space *mapping, pgoff_t offset) in find_get_entry()
1033 struct page *page; in find_get_entry() local
1037 page = NULL; in find_get_entry()
1040 page = radix_tree_deref_slot(pagep); in find_get_entry()
1041 if (unlikely(!page)) in find_get_entry()
1043 if (radix_tree_exception(page)) { in find_get_entry()
1044 if (radix_tree_deref_retry(page)) in find_get_entry()
1053 if (!page_cache_get_speculative(page)) in find_get_entry()
1061 if (unlikely(page != *pagep)) { in find_get_entry()
1062 page_cache_release(page); in find_get_entry()
1069 return page; in find_get_entry()
1089 struct page *find_lock_entry(struct address_space *mapping, pgoff_t offset) in find_lock_entry()
1091 struct page *page; in find_lock_entry() local
1094 page = find_get_entry(mapping, offset); in find_lock_entry()
1095 if (page && !radix_tree_exception(page)) { in find_lock_entry()
1096 lock_page(page); in find_lock_entry()
1098 if (unlikely(page->mapping != mapping)) { in find_lock_entry()
1099 unlock_page(page); in find_lock_entry()
1100 page_cache_release(page); in find_lock_entry()
1103 VM_BUG_ON_PAGE(page->index != offset, page); in find_lock_entry()
1105 return page; in find_lock_entry()
1132 struct page *pagecache_get_page(struct address_space *mapping, pgoff_t offset, in pagecache_get_page()
1135 struct page *page; in pagecache_get_page() local
1138 page = find_get_entry(mapping, offset); in pagecache_get_page()
1139 if (radix_tree_exceptional_entry(page)) in pagecache_get_page()
1140 page = NULL; in pagecache_get_page()
1141 if (!page) in pagecache_get_page()
1146 if (!trylock_page(page)) { in pagecache_get_page()
1147 page_cache_release(page); in pagecache_get_page()
1151 lock_page(page); in pagecache_get_page()
1155 if (unlikely(page->mapping != mapping)) { in pagecache_get_page()
1156 unlock_page(page); in pagecache_get_page()
1157 page_cache_release(page); in pagecache_get_page()
1160 VM_BUG_ON_PAGE(page->index != offset, page); in pagecache_get_page()
1163 if (page && (fgp_flags & FGP_ACCESSED)) in pagecache_get_page()
1164 mark_page_accessed(page); in pagecache_get_page()
1167 if (!page && (fgp_flags & FGP_CREAT)) { in pagecache_get_page()
1174 page = __page_cache_alloc(gfp_mask); in pagecache_get_page()
1175 if (!page) in pagecache_get_page()
1183 __SetPageReferenced(page); in pagecache_get_page()
1185 err = add_to_page_cache_lru(page, mapping, offset, in pagecache_get_page()
1188 page_cache_release(page); in pagecache_get_page()
1189 page = NULL; in pagecache_get_page()
1195 return page; in pagecache_get_page()
1224 struct page **entries, pgoff_t *indices) in find_get_entries()
1236 struct page *page; in find_get_entries() local
1238 page = radix_tree_deref_slot(slot); in find_get_entries()
1239 if (unlikely(!page)) in find_get_entries()
1241 if (radix_tree_exception(page)) { in find_get_entries()
1242 if (radix_tree_deref_retry(page)) in find_get_entries()
1251 if (!page_cache_get_speculative(page)) in find_get_entries()
1255 if (unlikely(page != *slot)) { in find_get_entries()
1256 page_cache_release(page); in find_get_entries()
1261 entries[ret] = page; in find_get_entries()
1286 unsigned int nr_pages, struct page **pages) in find_get_pages()
1298 struct page *page; in find_get_pages() local
1300 page = radix_tree_deref_slot(slot); in find_get_pages()
1301 if (unlikely(!page)) in find_get_pages()
1304 if (radix_tree_exception(page)) { in find_get_pages()
1305 if (radix_tree_deref_retry(page)) { in find_get_pages()
1322 if (!page_cache_get_speculative(page)) in find_get_pages()
1326 if (unlikely(page != *slot)) { in find_get_pages()
1327 page_cache_release(page); in find_get_pages()
1331 pages[ret] = page; in find_get_pages()
1353 unsigned int nr_pages, struct page **pages) in find_get_pages_contig()
1365 struct page *page; in find_get_pages_contig() local
1367 page = radix_tree_deref_slot(slot); in find_get_pages_contig()
1369 if (unlikely(!page)) in find_get_pages_contig()
1372 if (radix_tree_exception(page)) { in find_get_pages_contig()
1373 if (radix_tree_deref_retry(page)) { in find_get_pages_contig()
1389 if (!page_cache_get_speculative(page)) in find_get_pages_contig()
1393 if (unlikely(page != *slot)) { in find_get_pages_contig()
1394 page_cache_release(page); in find_get_pages_contig()
1403 if (page->mapping == NULL || page->index != iter.index) { in find_get_pages_contig()
1404 page_cache_release(page); in find_get_pages_contig()
1408 pages[ret] = page; in find_get_pages_contig()
1429 int tag, unsigned int nr_pages, struct page **pages) in find_get_pages_tag()
1442 struct page *page; in find_get_pages_tag() local
1444 page = radix_tree_deref_slot(slot); in find_get_pages_tag()
1445 if (unlikely(!page)) in find_get_pages_tag()
1448 if (radix_tree_exception(page)) { in find_get_pages_tag()
1449 if (radix_tree_deref_retry(page)) { in find_get_pages_tag()
1471 if (!page_cache_get_speculative(page)) in find_get_pages_tag()
1475 if (unlikely(page != *slot)) { in find_get_pages_tag()
1476 page_cache_release(page); in find_get_pages_tag()
1480 pages[ret] = page; in find_get_pages_tag()
1548 struct page *page; in do_generic_file_read() local
1555 page = find_get_page(mapping, index); in do_generic_file_read()
1556 if (!page) { in do_generic_file_read()
1560 page = find_get_page(mapping, index); in do_generic_file_read()
1561 if (unlikely(page == NULL)) in do_generic_file_read()
1564 if (PageReadahead(page)) { in do_generic_file_read()
1566 ra, filp, page, in do_generic_file_read()
1569 if (!PageUptodate(page)) { in do_generic_file_read()
1573 if (!trylock_page(page)) in do_generic_file_read()
1576 if (!page->mapping) in do_generic_file_read()
1578 if (!mapping->a_ops->is_partially_uptodate(page, in do_generic_file_read()
1581 unlock_page(page); in do_generic_file_read()
1596 page_cache_release(page); in do_generic_file_read()
1605 page_cache_release(page); in do_generic_file_read()
1616 flush_dcache_page(page); in do_generic_file_read()
1623 mark_page_accessed(page); in do_generic_file_read()
1631 ret = copy_page_to_iter(page, offset, nr, iter); in do_generic_file_read()
1637 page_cache_release(page); in do_generic_file_read()
1649 error = lock_page_killable(page); in do_generic_file_read()
1655 if (!page->mapping) { in do_generic_file_read()
1656 unlock_page(page); in do_generic_file_read()
1657 page_cache_release(page); in do_generic_file_read()
1662 if (PageUptodate(page)) { in do_generic_file_read()
1663 unlock_page(page); in do_generic_file_read()
1673 ClearPageError(page); in do_generic_file_read()
1675 error = mapping->a_ops->readpage(filp, page); in do_generic_file_read()
1679 page_cache_release(page); in do_generic_file_read()
1686 if (!PageUptodate(page)) { in do_generic_file_read()
1687 error = lock_page_killable(page); in do_generic_file_read()
1690 if (!PageUptodate(page)) { in do_generic_file_read()
1691 if (page->mapping == NULL) { in do_generic_file_read()
1695 unlock_page(page); in do_generic_file_read()
1696 page_cache_release(page); in do_generic_file_read()
1699 unlock_page(page); in do_generic_file_read()
1704 unlock_page(page); in do_generic_file_read()
1711 page_cache_release(page); in do_generic_file_read()
1719 page = page_cache_alloc_cold(mapping); in do_generic_file_read()
1720 if (!page) { in do_generic_file_read()
1724 error = add_to_page_cache_lru(page, mapping, index, in do_generic_file_read()
1727 page_cache_release(page); in do_generic_file_read()
1818 struct page *page; in page_cache_read() local
1822 page = page_cache_alloc_cold(mapping); in page_cache_read()
1823 if (!page) in page_cache_read()
1826 ret = add_to_page_cache_lru(page, mapping, offset, in page_cache_read()
1829 ret = mapping->a_ops->readpage(file, page); in page_cache_read()
1833 page_cache_release(page); in page_cache_read()
1892 struct page *page, in do_async_mmap_readahead() argument
1902 if (PageReadahead(page)) in do_async_mmap_readahead()
1904 page, offset, ra->ra_pages); in do_async_mmap_readahead()
1939 struct page *page; in filemap_fault() local
1950 page = find_get_page(mapping, offset); in filemap_fault()
1951 if (likely(page) && !(vmf->flags & FAULT_FLAG_TRIED)) { in filemap_fault()
1956 do_async_mmap_readahead(vma, ra, file, page, offset); in filemap_fault()
1957 } else if (!page) { in filemap_fault()
1964 page = find_get_page(mapping, offset); in filemap_fault()
1965 if (!page) in filemap_fault()
1969 if (!lock_page_or_retry(page, vma->vm_mm, vmf->flags)) { in filemap_fault()
1970 page_cache_release(page); in filemap_fault()
1975 if (unlikely(page->mapping != mapping)) { in filemap_fault()
1976 unlock_page(page); in filemap_fault()
1977 put_page(page); in filemap_fault()
1980 VM_BUG_ON_PAGE(page->index != offset, page); in filemap_fault()
1986 if (unlikely(!PageUptodate(page))) in filemap_fault()
1995 unlock_page(page); in filemap_fault()
1996 page_cache_release(page); in filemap_fault()
2000 vmf->page = page; in filemap_fault()
2034 ClearPageError(page); in filemap_fault()
2035 error = mapping->a_ops->readpage(file, page); in filemap_fault()
2037 wait_on_page_locked(page); in filemap_fault()
2038 if (!PageUptodate(page)) in filemap_fault()
2041 page_cache_release(page); in filemap_fault()
2059 struct page *page; in filemap_map_pages() local
2069 page = radix_tree_deref_slot(slot); in filemap_map_pages()
2070 if (unlikely(!page)) in filemap_map_pages()
2072 if (radix_tree_exception(page)) { in filemap_map_pages()
2073 if (radix_tree_deref_retry(page)) in filemap_map_pages()
2079 if (!page_cache_get_speculative(page)) in filemap_map_pages()
2083 if (unlikely(page != *slot)) { in filemap_map_pages()
2084 page_cache_release(page); in filemap_map_pages()
2088 if (!PageUptodate(page) || in filemap_map_pages()
2089 PageReadahead(page) || in filemap_map_pages()
2090 PageHWPoison(page)) in filemap_map_pages()
2092 if (!trylock_page(page)) in filemap_map_pages()
2095 if (page->mapping != mapping || !PageUptodate(page)) in filemap_map_pages()
2099 if (page->index >= size >> PAGE_CACHE_SHIFT) in filemap_map_pages()
2102 pte = vmf->pte + page->index - vmf->pgoff; in filemap_map_pages()
2108 addr = address + (page->index - vmf->pgoff) * PAGE_SIZE; in filemap_map_pages()
2109 do_set_pte(vma, addr, page, pte, false, false); in filemap_map_pages()
2110 unlock_page(page); in filemap_map_pages()
2113 unlock_page(page); in filemap_map_pages()
2115 page_cache_release(page); in filemap_map_pages()
2126 struct page *page = vmf->page; in filemap_page_mkwrite() local
2132 lock_page(page); in filemap_page_mkwrite()
2133 if (page->mapping != inode->i_mapping) { in filemap_page_mkwrite()
2134 unlock_page(page); in filemap_page_mkwrite()
2143 set_page_dirty(page); in filemap_page_mkwrite()
2144 wait_for_stable_page(page); in filemap_page_mkwrite()
2193 static struct page *wait_on_page_read(struct page *page) in wait_on_page_read() argument
2195 if (!IS_ERR(page)) { in wait_on_page_read()
2196 wait_on_page_locked(page); in wait_on_page_read()
2197 if (!PageUptodate(page)) { in wait_on_page_read()
2198 page_cache_release(page); in wait_on_page_read()
2199 page = ERR_PTR(-EIO); in wait_on_page_read()
2202 return page; in wait_on_page_read()
2205 static struct page *__read_cache_page(struct address_space *mapping, in __read_cache_page()
2207 int (*filler)(void *, struct page *), in __read_cache_page() argument
2211 struct page *page; in __read_cache_page() local
2214 page = find_get_page(mapping, index); in __read_cache_page()
2215 if (!page) { in __read_cache_page()
2216 page = __page_cache_alloc(gfp | __GFP_COLD); in __read_cache_page()
2217 if (!page) in __read_cache_page()
2219 err = add_to_page_cache_lru(page, mapping, index, gfp); in __read_cache_page()
2221 page_cache_release(page); in __read_cache_page()
2227 err = filler(data, page); in __read_cache_page()
2229 page_cache_release(page); in __read_cache_page()
2230 page = ERR_PTR(err); in __read_cache_page()
2232 page = wait_on_page_read(page); in __read_cache_page()
2235 return page; in __read_cache_page()
2238 static struct page *do_read_cache_page(struct address_space *mapping, in do_read_cache_page()
2240 int (*filler)(void *, struct page *), in do_read_cache_page() argument
2245 struct page *page; in do_read_cache_page() local
2249 page = __read_cache_page(mapping, index, filler, data, gfp); in do_read_cache_page()
2250 if (IS_ERR(page)) in do_read_cache_page()
2251 return page; in do_read_cache_page()
2252 if (PageUptodate(page)) in do_read_cache_page()
2255 lock_page(page); in do_read_cache_page()
2256 if (!page->mapping) { in do_read_cache_page()
2257 unlock_page(page); in do_read_cache_page()
2258 page_cache_release(page); in do_read_cache_page()
2261 if (PageUptodate(page)) { in do_read_cache_page()
2262 unlock_page(page); in do_read_cache_page()
2265 err = filler(data, page); in do_read_cache_page()
2267 page_cache_release(page); in do_read_cache_page()
2270 page = wait_on_page_read(page); in do_read_cache_page()
2271 if (IS_ERR(page)) in do_read_cache_page()
2272 return page; in do_read_cache_page()
2275 mark_page_accessed(page); in do_read_cache_page()
2276 return page; in do_read_cache_page()
2291 struct page *read_cache_page(struct address_space *mapping, in read_cache_page()
2293 int (*filler)(void *, struct page *), in read_cache_page() argument
2311 struct page *read_cache_page_gfp(struct address_space *mapping, in read_cache_page_gfp()
2379 struct page **pagep, void **fsdata) in pagecache_write_begin()
2390 struct page *page, void *fsdata) in pagecache_write_end() argument
2394 return aops->write_end(file, mapping, pos, len, copied, page, fsdata); in pagecache_write_end()
2470 struct page *grab_cache_page_write_begin(struct address_space *mapping, in grab_cache_page_write_begin()
2473 struct page *page; in grab_cache_page_write_begin() local
2479 page = pagecache_get_page(mapping, index, fgp_flags, in grab_cache_page_write_begin()
2481 if (page) in grab_cache_page_write_begin()
2482 wait_for_stable_page(page); in grab_cache_page_write_begin()
2484 return page; in grab_cache_page_write_begin()
2504 struct page *page; in generic_perform_write() local
2536 &page, &fsdata); in generic_perform_write()
2541 flush_dcache_page(page); in generic_perform_write()
2543 copied = iov_iter_copy_from_user_atomic(page, i, offset, bytes); in generic_perform_write()
2544 flush_dcache_page(page); in generic_perform_write()
2547 page, fsdata); in generic_perform_write()
2719 int try_to_release_page(struct page *page, gfp_t gfp_mask) in try_to_release_page() argument
2721 struct address_space * const mapping = page->mapping; in try_to_release_page()
2723 BUG_ON(!PageLocked(page)); in try_to_release_page()
2724 if (PageWriteback(page)) in try_to_release_page()
2728 return mapping->a_ops->releasepage(page, gfp_mask); in try_to_release_page()
2729 return try_to_free_buffers(page); in try_to_release_page()