Lines Matching refs:mapping
354 set_bit(AS_EIO, &page->mapping->flags); in end_buffer_async_write()
570 int sync_mapping_buffers(struct address_space *mapping) in sync_mapping_buffers() argument
572 struct address_space *buffer_mapping = mapping->private_data; in sync_mapping_buffers()
574 if (buffer_mapping == NULL || list_empty(&mapping->private_list)) in sync_mapping_buffers()
578 &mapping->private_list); in sync_mapping_buffers()
601 struct address_space *mapping = inode->i_mapping; in mark_buffer_dirty_inode() local
602 struct address_space *buffer_mapping = bh->b_page->mapping; in mark_buffer_dirty_inode()
605 if (!mapping->private_data) { in mark_buffer_dirty_inode()
606 mapping->private_data = buffer_mapping; in mark_buffer_dirty_inode()
608 BUG_ON(mapping->private_data != buffer_mapping); in mark_buffer_dirty_inode()
613 &mapping->private_list); in mark_buffer_dirty_inode()
614 bh->b_assoc_map = mapping; in mark_buffer_dirty_inode()
628 struct address_space *mapping, int warn) in __set_page_dirty() argument
632 spin_lock_irqsave(&mapping->tree_lock, flags); in __set_page_dirty()
633 if (page->mapping) { /* Race with truncate? */ in __set_page_dirty()
635 account_page_dirtied(page, mapping); in __set_page_dirty()
636 radix_tree_tag_set(&mapping->page_tree, in __set_page_dirty()
639 spin_unlock_irqrestore(&mapping->tree_lock, flags); in __set_page_dirty()
640 __mark_inode_dirty(mapping->host, I_DIRTY_PAGES); in __set_page_dirty()
671 struct address_space *mapping = page_mapping(page); in __set_page_dirty_buffers() local
673 if (unlikely(!mapping)) in __set_page_dirty_buffers()
676 spin_lock(&mapping->private_lock); in __set_page_dirty_buffers()
687 spin_unlock(&mapping->private_lock); in __set_page_dirty_buffers()
690 __set_page_dirty(page, mapping, 1); in __set_page_dirty_buffers()
718 struct address_space *mapping; in fsync_buffers_list() local
728 mapping = bh->b_assoc_map; in fsync_buffers_list()
735 bh->b_assoc_map = mapping; in fsync_buffers_list()
767 mapping = bh->b_assoc_map; in fsync_buffers_list()
774 &mapping->private_list); in fsync_buffers_list()
775 bh->b_assoc_map = mapping; in fsync_buffers_list()
805 struct address_space *mapping = &inode->i_data; in invalidate_inode_buffers() local
806 struct list_head *list = &mapping->private_list; in invalidate_inode_buffers()
807 struct address_space *buffer_mapping = mapping->private_data; in invalidate_inode_buffers()
828 struct address_space *mapping = &inode->i_data; in remove_inode_buffers() local
829 struct list_head *list = &mapping->private_list; in remove_inode_buffers()
830 struct address_space *buffer_mapping = mapping->private_data; in remove_inode_buffers()
1162 struct address_space *mapping = page_mapping(page); in mark_buffer_dirty() local
1163 if (mapping) in mark_buffer_dirty()
1164 __set_page_dirty(page, mapping, 0); in mark_buffer_dirty()
1195 struct address_space *buffer_mapping = bh->b_page->mapping; in __bforget()
1578 spin_lock(&page->mapping->private_lock); in create_empty_buffers()
1590 spin_unlock(&page->mapping->private_lock); in create_empty_buffers()
1825 mapping_set_error(page->mapping, err); in __block_write_full_page()
1887 struct inode *inode = page->mapping->host; in __block_write_begin()
2008 int block_write_begin(struct address_space *mapping, loff_t pos, unsigned len, in block_write_begin() argument
2015 page = grab_cache_page_write_begin(mapping, index, flags); in block_write_begin()
2031 int block_write_end(struct file *file, struct address_space *mapping, in block_write_end() argument
2035 struct inode *inode = mapping->host; in block_write_end()
2067 int generic_write_end(struct file *file, struct address_space *mapping, in generic_write_end() argument
2071 struct inode *inode = mapping->host; in generic_write_end()
2075 copied = block_write_end(file, mapping, pos, len, copied, page, fsdata); in generic_write_end()
2161 struct inode *inode = page->mapping->host; in block_read_full_page()
2251 struct address_space *mapping = inode->i_mapping; in generic_cont_expand_simple() local
2260 err = pagecache_write_begin(NULL, mapping, size, 0, in generic_cont_expand_simple()
2266 err = pagecache_write_end(NULL, mapping, size, 0, 0, page, fsdata); in generic_cont_expand_simple()
2274 static int cont_expand_zero(struct file *file, struct address_space *mapping, in cont_expand_zero() argument
2277 struct inode *inode = mapping->host; in cont_expand_zero()
2297 err = pagecache_write_begin(file, mapping, curpos, len, in cont_expand_zero()
2303 err = pagecache_write_end(file, mapping, curpos, len, len, in cont_expand_zero()
2310 balance_dirty_pages_ratelimited(mapping); in cont_expand_zero()
2331 err = pagecache_write_begin(file, mapping, curpos, len, in cont_expand_zero()
2337 err = pagecache_write_end(file, mapping, curpos, len, len, in cont_expand_zero()
2352 int cont_write_begin(struct file *file, struct address_space *mapping, in cont_write_begin() argument
2357 struct inode *inode = mapping->host; in cont_write_begin()
2362 err = cont_expand_zero(file, mapping, pos, bytes); in cont_write_begin()
2372 return block_write_begin(mapping, pos, len, flags, pagep, get_block); in cont_write_begin()
2378 struct inode *inode = page->mapping->host; in block_commit_write()
2413 if ((page->mapping != inode->i_mapping) || in __block_page_mkwrite()
2482 spin_lock(&page->mapping->private_lock); in attach_nobh_buffers()
2492 spin_unlock(&page->mapping->private_lock); in attach_nobh_buffers()
2500 int nobh_write_begin(struct address_space *mapping, in nobh_write_begin() argument
2505 struct inode *inode = mapping->host; in nobh_write_begin()
2523 page = grab_cache_page_write_begin(mapping, index, flags); in nobh_write_begin()
2641 int nobh_write_end(struct file *file, struct address_space *mapping, in nobh_write_end() argument
2645 struct inode *inode = page->mapping->host; in nobh_write_end()
2653 return generic_write_end(file, mapping, pos, len, in nobh_write_end()
2684 struct inode * const inode = page->mapping->host; in nobh_writepage()
2704 if (page->mapping->a_ops->invalidatepage) in nobh_writepage()
2705 page->mapping->a_ops->invalidatepage(page, offset); in nobh_writepage()
2728 int nobh_truncate_page(struct address_space *mapping, in nobh_truncate_page() argument
2736 struct inode *inode = mapping->host; in nobh_truncate_page()
2751 page = grab_cache_page(mapping, index); in nobh_truncate_page()
2760 return block_truncate_page(mapping, from, get_block); in nobh_truncate_page()
2781 err = mapping->a_ops->readpage(NULL, page); in nobh_truncate_page()
2806 int block_truncate_page(struct address_space *mapping, in block_truncate_page() argument
2814 struct inode *inode = mapping->host; in block_truncate_page()
2829 page = grab_cache_page(mapping, index); in block_truncate_page()
2888 struct inode * const inode = page->mapping->host; in block_write_full_page()
2924 sector_t generic_block_bmap(struct address_space *mapping, sector_t block, in generic_block_bmap() argument
2928 struct inode *inode = mapping->host; in generic_block_bmap()
3193 if (buffer_write_io_error(bh) && page->mapping) in drop_buffers()
3194 set_bit(AS_EIO, &page->mapping->flags); in drop_buffers()
3216 struct address_space * const mapping = page->mapping; in try_to_free_buffers() local
3224 if (mapping == NULL) { /* can this still happen? */ in try_to_free_buffers()
3229 spin_lock(&mapping->private_lock); in try_to_free_buffers()
3247 account_page_cleaned(page, mapping); in try_to_free_buffers()
3248 spin_unlock(&mapping->private_lock); in try_to_free_buffers()