Lines Matching refs:page

112 				   struct page *page, void *shadow)  in page_cache_tree_delete()  argument
120 VM_BUG_ON(!PageLocked(page)); in page_cache_tree_delete()
122 __radix_tree_lookup(&mapping->page_tree, page->index, &node, &slot); in page_cache_tree_delete()
144 index = page->index; in page_cache_tree_delete()
179 void __delete_from_page_cache(struct page *page, void *shadow) in __delete_from_page_cache() argument
181 struct address_space *mapping = page->mapping; in __delete_from_page_cache()
183 trace_mm_filemap_delete_from_page_cache(page); in __delete_from_page_cache()
189 if (PageUptodate(page) && PageMappedToDisk(page)) in __delete_from_page_cache()
190 cleancache_put_page(page); in __delete_from_page_cache()
192 cleancache_invalidate_page(mapping, page); in __delete_from_page_cache()
194 page_cache_tree_delete(mapping, page, shadow); in __delete_from_page_cache()
196 page->mapping = NULL; in __delete_from_page_cache()
199 __dec_zone_page_state(page, NR_FILE_PAGES); in __delete_from_page_cache()
200 if (PageSwapBacked(page)) in __delete_from_page_cache()
201 __dec_zone_page_state(page, NR_SHMEM); in __delete_from_page_cache()
202 BUG_ON(page_mapped(page)); in __delete_from_page_cache()
212 if (WARN_ON_ONCE(PageDirty(page))) in __delete_from_page_cache()
213 account_page_cleaned(page, mapping); in __delete_from_page_cache()
224 void delete_from_page_cache(struct page *page) in delete_from_page_cache() argument
226 struct address_space *mapping = page->mapping; in delete_from_page_cache()
227 void (*freepage)(struct page *); in delete_from_page_cache()
229 BUG_ON(!PageLocked(page)); in delete_from_page_cache()
233 __delete_from_page_cache(page, NULL); in delete_from_page_cache()
237 freepage(page); in delete_from_page_cache()
238 page_cache_release(page); in delete_from_page_cache()
349 struct page *page = pvec.pages[i]; in filemap_fdatawait_range() local
352 if (page->index > end) in filemap_fdatawait_range()
355 wait_on_page_writeback(page); in filemap_fdatawait_range()
356 if (TestClearPageError(page)) in filemap_fdatawait_range()
461 int replace_page_cache_page(struct page *old, struct page *new, gfp_t gfp_mask) in replace_page_cache_page()
472 void (*freepage)(struct page *); in replace_page_cache_page()
502 struct page *page, void **shadowp) in page_cache_tree_insert() argument
508 error = __radix_tree_create(&mapping->page_tree, page->index, in page_cache_tree_insert()
524 radix_tree_replace_slot(slot, page); in page_cache_tree_insert()
543 static int __add_to_page_cache_locked(struct page *page, in __add_to_page_cache_locked() argument
548 int huge = PageHuge(page); in __add_to_page_cache_locked()
552 VM_BUG_ON_PAGE(!PageLocked(page), page); in __add_to_page_cache_locked()
553 VM_BUG_ON_PAGE(PageSwapBacked(page), page); in __add_to_page_cache_locked()
556 error = mem_cgroup_try_charge(page, current->mm, in __add_to_page_cache_locked()
565 mem_cgroup_cancel_charge(page, memcg); in __add_to_page_cache_locked()
569 page_cache_get(page); in __add_to_page_cache_locked()
570 page->mapping = mapping; in __add_to_page_cache_locked()
571 page->index = offset; in __add_to_page_cache_locked()
574 error = page_cache_tree_insert(mapping, page, shadowp); in __add_to_page_cache_locked()
578 __inc_zone_page_state(page, NR_FILE_PAGES); in __add_to_page_cache_locked()
581 mem_cgroup_commit_charge(page, memcg, false); in __add_to_page_cache_locked()
582 trace_mm_filemap_add_to_page_cache(page); in __add_to_page_cache_locked()
585 page->mapping = NULL; in __add_to_page_cache_locked()
589 mem_cgroup_cancel_charge(page, memcg); in __add_to_page_cache_locked()
590 page_cache_release(page); in __add_to_page_cache_locked()
604 int add_to_page_cache_locked(struct page *page, struct address_space *mapping, in add_to_page_cache_locked() argument
607 return __add_to_page_cache_locked(page, mapping, offset, in add_to_page_cache_locked()
612 int add_to_page_cache_lru(struct page *page, struct address_space *mapping, in add_to_page_cache_lru() argument
618 __set_page_locked(page); in add_to_page_cache_lru()
619 ret = __add_to_page_cache_locked(page, mapping, offset, in add_to_page_cache_lru()
622 __clear_page_locked(page); in add_to_page_cache_lru()
630 SetPageActive(page); in add_to_page_cache_lru()
631 workingset_activation(page); in add_to_page_cache_lru()
633 ClearPageActive(page); in add_to_page_cache_lru()
634 lru_cache_add(page); in add_to_page_cache_lru()
641 struct page *__page_cache_alloc(gfp_t gfp) in __page_cache_alloc()
644 struct page *page; in __page_cache_alloc() local
651 page = alloc_pages_exact_node(n, gfp, 0); in __page_cache_alloc()
652 } while (!page && read_mems_allowed_retry(cpuset_mems_cookie)); in __page_cache_alloc()
654 return page; in __page_cache_alloc()
671 wait_queue_head_t *page_waitqueue(struct page *page) in page_waitqueue() argument
673 const struct zone *zone = page_zone(page); in page_waitqueue()
675 return &zone->wait_table[hash_ptr(page, zone->wait_table_bits)]; in page_waitqueue()
679 void wait_on_page_bit(struct page *page, int bit_nr) in wait_on_page_bit() argument
681 DEFINE_WAIT_BIT(wait, &page->flags, bit_nr); in wait_on_page_bit()
683 if (test_bit(bit_nr, &page->flags)) in wait_on_page_bit()
684 __wait_on_bit(page_waitqueue(page), &wait, bit_wait_io, in wait_on_page_bit()
689 int wait_on_page_bit_killable(struct page *page, int bit_nr) in wait_on_page_bit_killable() argument
691 DEFINE_WAIT_BIT(wait, &page->flags, bit_nr); in wait_on_page_bit_killable()
693 if (!test_bit(bit_nr, &page->flags)) in wait_on_page_bit_killable()
696 return __wait_on_bit(page_waitqueue(page), &wait, in wait_on_page_bit_killable()
700 int wait_on_page_bit_killable_timeout(struct page *page, in wait_on_page_bit_killable_timeout() argument
703 DEFINE_WAIT_BIT(wait, &page->flags, bit_nr); in wait_on_page_bit_killable_timeout()
706 if (!test_bit(bit_nr, &page->flags)) in wait_on_page_bit_killable_timeout()
708 return __wait_on_bit(page_waitqueue(page), &wait, in wait_on_page_bit_killable_timeout()
720 void add_page_wait_queue(struct page *page, wait_queue_t *waiter) in add_page_wait_queue() argument
722 wait_queue_head_t *q = page_waitqueue(page); in add_page_wait_queue()
743 void unlock_page(struct page *page) in unlock_page() argument
745 VM_BUG_ON_PAGE(!PageLocked(page), page); in unlock_page()
746 clear_bit_unlock(PG_locked, &page->flags); in unlock_page()
748 wake_up_page(page, PG_locked); in unlock_page()
756 void end_page_writeback(struct page *page) in end_page_writeback() argument
765 if (PageReclaim(page)) { in end_page_writeback()
766 ClearPageReclaim(page); in end_page_writeback()
767 rotate_reclaimable_page(page); in end_page_writeback()
770 if (!test_clear_page_writeback(page)) in end_page_writeback()
774 wake_up_page(page, PG_writeback); in end_page_writeback()
782 void page_endio(struct page *page, int rw, int err) in page_endio() argument
786 SetPageUptodate(page); in page_endio()
788 ClearPageUptodate(page); in page_endio()
789 SetPageError(page); in page_endio()
791 unlock_page(page); in page_endio()
794 SetPageError(page); in page_endio()
795 if (page->mapping) in page_endio()
796 mapping_set_error(page->mapping, err); in page_endio()
798 end_page_writeback(page); in page_endio()
807 void __lock_page(struct page *page) in __lock_page() argument
809 DEFINE_WAIT_BIT(wait, &page->flags, PG_locked); in __lock_page()
811 __wait_on_bit_lock(page_waitqueue(page), &wait, bit_wait_io, in __lock_page()
816 int __lock_page_killable(struct page *page) in __lock_page_killable() argument
818 DEFINE_WAIT_BIT(wait, &page->flags, PG_locked); in __lock_page_killable()
820 return __wait_on_bit_lock(page_waitqueue(page), &wait, in __lock_page_killable()
836 int __lock_page_or_retry(struct page *page, struct mm_struct *mm, in __lock_page_or_retry() argument
849 wait_on_page_locked_killable(page); in __lock_page_or_retry()
851 wait_on_page_locked(page); in __lock_page_or_retry()
857 ret = __lock_page_killable(page); in __lock_page_or_retry()
863 __lock_page(page); in __lock_page_or_retry()
895 struct page *page; in page_cache_next_hole() local
897 page = radix_tree_lookup(&mapping->page_tree, index); in page_cache_next_hole()
898 if (!page || radix_tree_exceptional_entry(page)) in page_cache_next_hole()
936 struct page *page; in page_cache_prev_hole() local
938 page = radix_tree_lookup(&mapping->page_tree, index); in page_cache_prev_hole()
939 if (!page || radix_tree_exceptional_entry(page)) in page_cache_prev_hole()
963 struct page *find_get_entry(struct address_space *mapping, pgoff_t offset) in find_get_entry()
966 struct page *page; in find_get_entry() local
970 page = NULL; in find_get_entry()
973 page = radix_tree_deref_slot(pagep); in find_get_entry()
974 if (unlikely(!page)) in find_get_entry()
976 if (radix_tree_exception(page)) { in find_get_entry()
977 if (radix_tree_deref_retry(page)) in find_get_entry()
986 if (!page_cache_get_speculative(page)) in find_get_entry()
994 if (unlikely(page != *pagep)) { in find_get_entry()
995 page_cache_release(page); in find_get_entry()
1002 return page; in find_get_entry()
1022 struct page *find_lock_entry(struct address_space *mapping, pgoff_t offset) in find_lock_entry()
1024 struct page *page; in find_lock_entry() local
1027 page = find_get_entry(mapping, offset); in find_lock_entry()
1028 if (page && !radix_tree_exception(page)) { in find_lock_entry()
1029 lock_page(page); in find_lock_entry()
1031 if (unlikely(page->mapping != mapping)) { in find_lock_entry()
1032 unlock_page(page); in find_lock_entry()
1033 page_cache_release(page); in find_lock_entry()
1036 VM_BUG_ON_PAGE(page->index != offset, page); in find_lock_entry()
1038 return page; in find_lock_entry()
1065 struct page *pagecache_get_page(struct address_space *mapping, pgoff_t offset, in pagecache_get_page()
1068 struct page *page; in pagecache_get_page() local
1071 page = find_get_entry(mapping, offset); in pagecache_get_page()
1072 if (radix_tree_exceptional_entry(page)) in pagecache_get_page()
1073 page = NULL; in pagecache_get_page()
1074 if (!page) in pagecache_get_page()
1079 if (!trylock_page(page)) { in pagecache_get_page()
1080 page_cache_release(page); in pagecache_get_page()
1084 lock_page(page); in pagecache_get_page()
1088 if (unlikely(page->mapping != mapping)) { in pagecache_get_page()
1089 unlock_page(page); in pagecache_get_page()
1090 page_cache_release(page); in pagecache_get_page()
1093 VM_BUG_ON_PAGE(page->index != offset, page); in pagecache_get_page()
1096 if (page && (fgp_flags & FGP_ACCESSED)) in pagecache_get_page()
1097 mark_page_accessed(page); in pagecache_get_page()
1100 if (!page && (fgp_flags & FGP_CREAT)) { in pagecache_get_page()
1107 page = __page_cache_alloc(gfp_mask); in pagecache_get_page()
1108 if (!page) in pagecache_get_page()
1116 __SetPageReferenced(page); in pagecache_get_page()
1118 err = add_to_page_cache_lru(page, mapping, offset, in pagecache_get_page()
1121 page_cache_release(page); in pagecache_get_page()
1122 page = NULL; in pagecache_get_page()
1128 return page; in pagecache_get_page()
1157 struct page **entries, pgoff_t *indices) in find_get_entries()
1169 struct page *page; in find_get_entries() local
1171 page = radix_tree_deref_slot(slot); in find_get_entries()
1172 if (unlikely(!page)) in find_get_entries()
1174 if (radix_tree_exception(page)) { in find_get_entries()
1175 if (radix_tree_deref_retry(page)) in find_get_entries()
1184 if (!page_cache_get_speculative(page)) in find_get_entries()
1188 if (unlikely(page != *slot)) { in find_get_entries()
1189 page_cache_release(page); in find_get_entries()
1194 entries[ret] = page; in find_get_entries()
1219 unsigned int nr_pages, struct page **pages) in find_get_pages()
1231 struct page *page; in find_get_pages() local
1233 page = radix_tree_deref_slot(slot); in find_get_pages()
1234 if (unlikely(!page)) in find_get_pages()
1237 if (radix_tree_exception(page)) { in find_get_pages()
1238 if (radix_tree_deref_retry(page)) { in find_get_pages()
1255 if (!page_cache_get_speculative(page)) in find_get_pages()
1259 if (unlikely(page != *slot)) { in find_get_pages()
1260 page_cache_release(page); in find_get_pages()
1264 pages[ret] = page; in find_get_pages()
1286 unsigned int nr_pages, struct page **pages) in find_get_pages_contig()
1298 struct page *page; in find_get_pages_contig() local
1300 page = radix_tree_deref_slot(slot); in find_get_pages_contig()
1302 if (unlikely(!page)) in find_get_pages_contig()
1305 if (radix_tree_exception(page)) { in find_get_pages_contig()
1306 if (radix_tree_deref_retry(page)) { in find_get_pages_contig()
1322 if (!page_cache_get_speculative(page)) in find_get_pages_contig()
1326 if (unlikely(page != *slot)) { in find_get_pages_contig()
1327 page_cache_release(page); in find_get_pages_contig()
1336 if (page->mapping == NULL || page->index != iter.index) { in find_get_pages_contig()
1337 page_cache_release(page); in find_get_pages_contig()
1341 pages[ret] = page; in find_get_pages_contig()
1362 int tag, unsigned int nr_pages, struct page **pages) in find_get_pages_tag()
1375 struct page *page; in find_get_pages_tag() local
1377 page = radix_tree_deref_slot(slot); in find_get_pages_tag()
1378 if (unlikely(!page)) in find_get_pages_tag()
1381 if (radix_tree_exception(page)) { in find_get_pages_tag()
1382 if (radix_tree_deref_retry(page)) { in find_get_pages_tag()
1404 if (!page_cache_get_speculative(page)) in find_get_pages_tag()
1408 if (unlikely(page != *slot)) { in find_get_pages_tag()
1409 page_cache_release(page); in find_get_pages_tag()
1413 pages[ret] = page; in find_get_pages_tag()
1481 struct page *page; in do_generic_file_read() local
1488 page = find_get_page(mapping, index); in do_generic_file_read()
1489 if (!page) { in do_generic_file_read()
1493 page = find_get_page(mapping, index); in do_generic_file_read()
1494 if (unlikely(page == NULL)) in do_generic_file_read()
1497 if (PageReadahead(page)) { in do_generic_file_read()
1499 ra, filp, page, in do_generic_file_read()
1502 if (!PageUptodate(page)) { in do_generic_file_read()
1506 if (!trylock_page(page)) in do_generic_file_read()
1509 if (!page->mapping) in do_generic_file_read()
1511 if (!mapping->a_ops->is_partially_uptodate(page, in do_generic_file_read()
1514 unlock_page(page); in do_generic_file_read()
1529 page_cache_release(page); in do_generic_file_read()
1538 page_cache_release(page); in do_generic_file_read()
1549 flush_dcache_page(page); in do_generic_file_read()
1556 mark_page_accessed(page); in do_generic_file_read()
1564 ret = copy_page_to_iter(page, offset, nr, iter); in do_generic_file_read()
1570 page_cache_release(page); in do_generic_file_read()
1582 error = lock_page_killable(page); in do_generic_file_read()
1588 if (!page->mapping) { in do_generic_file_read()
1589 unlock_page(page); in do_generic_file_read()
1590 page_cache_release(page); in do_generic_file_read()
1595 if (PageUptodate(page)) { in do_generic_file_read()
1596 unlock_page(page); in do_generic_file_read()
1606 ClearPageError(page); in do_generic_file_read()
1608 error = mapping->a_ops->readpage(filp, page); in do_generic_file_read()
1612 page_cache_release(page); in do_generic_file_read()
1619 if (!PageUptodate(page)) { in do_generic_file_read()
1620 error = lock_page_killable(page); in do_generic_file_read()
1623 if (!PageUptodate(page)) { in do_generic_file_read()
1624 if (page->mapping == NULL) { in do_generic_file_read()
1628 unlock_page(page); in do_generic_file_read()
1629 page_cache_release(page); in do_generic_file_read()
1632 unlock_page(page); in do_generic_file_read()
1637 unlock_page(page); in do_generic_file_read()
1644 page_cache_release(page); in do_generic_file_read()
1652 page = page_cache_alloc_cold(mapping); in do_generic_file_read()
1653 if (!page) { in do_generic_file_read()
1657 error = add_to_page_cache_lru(page, mapping, in do_generic_file_read()
1660 page_cache_release(page); in do_generic_file_read()
1751 struct page *page; in page_cache_read() local
1755 page = page_cache_alloc_cold(mapping); in page_cache_read()
1756 if (!page) in page_cache_read()
1759 ret = add_to_page_cache_lru(page, mapping, offset, GFP_KERNEL); in page_cache_read()
1761 ret = mapping->a_ops->readpage(file, page); in page_cache_read()
1765 page_cache_release(page); in page_cache_read()
1826 struct page *page, in do_async_mmap_readahead() argument
1836 if (PageReadahead(page)) in do_async_mmap_readahead()
1838 page, offset, ra->ra_pages); in do_async_mmap_readahead()
1873 struct page *page; in filemap_fault() local
1884 page = find_get_page(mapping, offset); in filemap_fault()
1885 if (likely(page) && !(vmf->flags & FAULT_FLAG_TRIED)) { in filemap_fault()
1890 do_async_mmap_readahead(vma, ra, file, page, offset); in filemap_fault()
1891 } else if (!page) { in filemap_fault()
1898 page = find_get_page(mapping, offset); in filemap_fault()
1899 if (!page) in filemap_fault()
1903 if (!lock_page_or_retry(page, vma->vm_mm, vmf->flags)) { in filemap_fault()
1904 page_cache_release(page); in filemap_fault()
1909 if (unlikely(page->mapping != mapping)) { in filemap_fault()
1910 unlock_page(page); in filemap_fault()
1911 put_page(page); in filemap_fault()
1914 VM_BUG_ON_PAGE(page->index != offset, page); in filemap_fault()
1920 if (unlikely(!PageUptodate(page))) in filemap_fault()
1929 unlock_page(page); in filemap_fault()
1930 page_cache_release(page); in filemap_fault()
1934 vmf->page = page; in filemap_fault()
1968 ClearPageError(page); in filemap_fault()
1969 error = mapping->a_ops->readpage(file, page); in filemap_fault()
1971 wait_on_page_locked(page); in filemap_fault()
1972 if (!PageUptodate(page)) in filemap_fault()
1975 page_cache_release(page); in filemap_fault()
1993 struct page *page; in filemap_map_pages() local
2003 page = radix_tree_deref_slot(slot); in filemap_map_pages()
2004 if (unlikely(!page)) in filemap_map_pages()
2006 if (radix_tree_exception(page)) { in filemap_map_pages()
2007 if (radix_tree_deref_retry(page)) in filemap_map_pages()
2013 if (!page_cache_get_speculative(page)) in filemap_map_pages()
2017 if (unlikely(page != *slot)) { in filemap_map_pages()
2018 page_cache_release(page); in filemap_map_pages()
2022 if (!PageUptodate(page) || in filemap_map_pages()
2023 PageReadahead(page) || in filemap_map_pages()
2024 PageHWPoison(page)) in filemap_map_pages()
2026 if (!trylock_page(page)) in filemap_map_pages()
2029 if (page->mapping != mapping || !PageUptodate(page)) in filemap_map_pages()
2033 if (page->index >= size >> PAGE_CACHE_SHIFT) in filemap_map_pages()
2036 pte = vmf->pte + page->index - vmf->pgoff; in filemap_map_pages()
2042 addr = address + (page->index - vmf->pgoff) * PAGE_SIZE; in filemap_map_pages()
2043 do_set_pte(vma, addr, page, pte, false, false); in filemap_map_pages()
2044 unlock_page(page); in filemap_map_pages()
2047 unlock_page(page); in filemap_map_pages()
2049 page_cache_release(page); in filemap_map_pages()
2060 struct page *page = vmf->page; in filemap_page_mkwrite() local
2066 lock_page(page); in filemap_page_mkwrite()
2067 if (page->mapping != inode->i_mapping) { in filemap_page_mkwrite()
2068 unlock_page(page); in filemap_page_mkwrite()
2077 set_page_dirty(page); in filemap_page_mkwrite()
2078 wait_for_stable_page(page); in filemap_page_mkwrite()
2127 static struct page *wait_on_page_read(struct page *page) in wait_on_page_read() argument
2129 if (!IS_ERR(page)) { in wait_on_page_read()
2130 wait_on_page_locked(page); in wait_on_page_read()
2131 if (!PageUptodate(page)) { in wait_on_page_read()
2132 page_cache_release(page); in wait_on_page_read()
2133 page = ERR_PTR(-EIO); in wait_on_page_read()
2136 return page; in wait_on_page_read()
2139 static struct page *__read_cache_page(struct address_space *mapping, in __read_cache_page()
2141 int (*filler)(void *, struct page *), in __read_cache_page() argument
2145 struct page *page; in __read_cache_page() local
2148 page = find_get_page(mapping, index); in __read_cache_page()
2149 if (!page) { in __read_cache_page()
2150 page = __page_cache_alloc(gfp | __GFP_COLD); in __read_cache_page()
2151 if (!page) in __read_cache_page()
2153 err = add_to_page_cache_lru(page, mapping, index, gfp); in __read_cache_page()
2155 page_cache_release(page); in __read_cache_page()
2161 err = filler(data, page); in __read_cache_page()
2163 page_cache_release(page); in __read_cache_page()
2164 page = ERR_PTR(err); in __read_cache_page()
2166 page = wait_on_page_read(page); in __read_cache_page()
2169 return page; in __read_cache_page()
2172 static struct page *do_read_cache_page(struct address_space *mapping, in do_read_cache_page()
2174 int (*filler)(void *, struct page *), in do_read_cache_page() argument
2179 struct page *page; in do_read_cache_page() local
2183 page = __read_cache_page(mapping, index, filler, data, gfp); in do_read_cache_page()
2184 if (IS_ERR(page)) in do_read_cache_page()
2185 return page; in do_read_cache_page()
2186 if (PageUptodate(page)) in do_read_cache_page()
2189 lock_page(page); in do_read_cache_page()
2190 if (!page->mapping) { in do_read_cache_page()
2191 unlock_page(page); in do_read_cache_page()
2192 page_cache_release(page); in do_read_cache_page()
2195 if (PageUptodate(page)) { in do_read_cache_page()
2196 unlock_page(page); in do_read_cache_page()
2199 err = filler(data, page); in do_read_cache_page()
2201 page_cache_release(page); in do_read_cache_page()
2204 page = wait_on_page_read(page); in do_read_cache_page()
2205 if (IS_ERR(page)) in do_read_cache_page()
2206 return page; in do_read_cache_page()
2209 mark_page_accessed(page); in do_read_cache_page()
2210 return page; in do_read_cache_page()
2225 struct page *read_cache_page(struct address_space *mapping, in read_cache_page()
2227 int (*filler)(void *, struct page *), in read_cache_page() argument
2245 struct page *read_cache_page_gfp(struct address_space *mapping, in read_cache_page_gfp()
2313 struct page **pagep, void **fsdata) in pagecache_write_begin()
2324 struct page *page, void *fsdata) in pagecache_write_end() argument
2328 return aops->write_end(file, mapping, pos, len, copied, page, fsdata); in pagecache_write_end()
2404 struct page *grab_cache_page_write_begin(struct address_space *mapping, in grab_cache_page_write_begin()
2407 struct page *page; in grab_cache_page_write_begin() local
2413 page = pagecache_get_page(mapping, index, fgp_flags, in grab_cache_page_write_begin()
2415 if (page) in grab_cache_page_write_begin()
2416 wait_for_stable_page(page); in grab_cache_page_write_begin()
2418 return page; in grab_cache_page_write_begin()
2438 struct page *page; in generic_perform_write() local
2470 &page, &fsdata); in generic_perform_write()
2475 flush_dcache_page(page); in generic_perform_write()
2477 copied = iov_iter_copy_from_user_atomic(page, i, offset, bytes); in generic_perform_write()
2478 flush_dcache_page(page); in generic_perform_write()
2481 page, fsdata); in generic_perform_write()
2653 int try_to_release_page(struct page *page, gfp_t gfp_mask) in try_to_release_page() argument
2655 struct address_space * const mapping = page->mapping; in try_to_release_page()
2657 BUG_ON(!PageLocked(page)); in try_to_release_page()
2658 if (PageWriteback(page)) in try_to_release_page()
2662 return mapping->a_ops->releasepage(page, gfp_mask); in try_to_release_page()
2663 return try_to_free_buffers(page); in try_to_release_page()