Lines Matching refs:end

68 		       state->start, state->end, state->state,  in btrfs_leak_debug_check()
85 #define btrfs_debug_check_extent_io_range(tree, start, end) \ argument
86 __btrfs_debug_check_extent_io_range(__func__, (tree), (start), (end))
88 struct extent_io_tree *tree, u64 start, u64 end) in __btrfs_debug_check_extent_io_range() argument
98 if (end >= PAGE_SIZE && (end % 2) == 0 && end != isize - 1) { in __btrfs_debug_check_extent_io_range()
101 caller, btrfs_ino(inode), isize, start, end); in __btrfs_debug_check_extent_io_range()
115 u64 end; member
261 else if (offset > entry->end) in tree_insert()
293 else if (offset > entry->end) in __etree_search()
306 while (prev && offset > prev_entry->end) { in __etree_search()
375 if (other->end == state->start - 1 && in merge_state()
387 if (other->start == state->end + 1 && in merge_state()
390 state->end = other->end; in merge_state()
426 struct extent_state *state, u64 start, u64 end, in insert_state() argument
433 if (end < start) in insert_state()
435 end, start); in insert_state()
437 state->end = end; in insert_state()
441 node = tree_insert(&tree->state, NULL, end, &state->rb_node, p, parent); in insert_state()
447 found->start, found->end, start, end); in insert_state()
483 prealloc->end = split - 1; in split_state()
487 node = tree_insert(&tree->state, &orig->rb_node, prealloc->end, in split_state()
520 u64 range = state->end - state->start + 1; in clear_state_bit()
572 int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, in clear_extent_bit() argument
585 btrfs_debug_check_extent_io_range(tree, start, end); in clear_extent_bit()
618 cached->start <= start && cached->end > start) { in clear_extent_bit()
636 if (state->start > end) in clear_extent_bit()
638 WARN_ON(state->end < start); in clear_extent_bit()
639 last_end = state->end; in clear_extent_bit()
673 if (state->end <= end) { in clear_extent_bit()
685 if (state->start <= end && state->end > end) { in clear_extent_bit()
688 err = split_state(tree, state, prealloc, end + 1); in clear_extent_bit()
706 if (start <= end && state && !need_resched()) in clear_extent_bit()
718 if (start > end) in clear_extent_bit()
744 static void wait_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, in wait_extent_bit() argument
750 btrfs_debug_check_extent_io_range(tree, start, end); in wait_extent_bit()
766 if (state->start > end) in wait_extent_bit()
776 start = state->end + 1; in wait_extent_bit()
778 if (start > end) in wait_extent_bit()
798 u64 range = state->end - state->start + 1; in set_state_bits()
835 __set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, in __set_extent_bit() argument
849 btrfs_debug_check_extent_io_range(tree, start, end); in __set_extent_bit()
861 if (state->start <= start && state->end > start && in __set_extent_bit()
875 err = insert_state(tree, prealloc, start, end, in __set_extent_bit()
887 last_end = state->end; in __set_extent_bit()
895 if (state->start == start && state->end <= end) { in __set_extent_bit()
909 if (start < end && state && state->start == start && in __set_extent_bit()
947 if (state->end <= end) { in __set_extent_bit()
955 if (start < end && state && state->start == start && in __set_extent_bit()
970 if (end < last_start) in __set_extent_bit()
971 this_end = end; in __set_extent_bit()
998 if (state->start <= end && state->end > end) { in __set_extent_bit()
1007 err = split_state(tree, state, prealloc, end + 1); in __set_extent_bit()
1028 if (start > end) in __set_extent_bit()
1036 int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, in set_extent_bit() argument
1040 return __set_extent_bit(tree, start, end, bits, 0, failed_start, in set_extent_bit()
1062 int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, in convert_extent_bit() argument
1076 btrfs_debug_check_extent_io_range(tree, start, end); in convert_extent_bit()
1095 if (state->start <= start && state->end > start && in convert_extent_bit()
1113 err = insert_state(tree, prealloc, start, end, in convert_extent_bit()
1124 last_end = state->end; in convert_extent_bit()
1132 if (state->start == start && state->end <= end) { in convert_extent_bit()
1139 if (start < end && state && state->start == start && in convert_extent_bit()
1173 if (state->end <= end) { in convert_extent_bit()
1180 if (start < end && state && state->start == start && in convert_extent_bit()
1195 if (end < last_start) in convert_extent_bit()
1196 this_end = end; in convert_extent_bit()
1225 if (state->start <= end && state->end > end) { in convert_extent_bit()
1232 err = split_state(tree, state, prealloc, end + 1); in convert_extent_bit()
1253 if (start > end) in convert_extent_bit()
1263 int set_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end, in set_extent_dirty() argument
1266 return set_extent_bit(tree, start, end, EXTENT_DIRTY, NULL, in set_extent_dirty()
1270 int set_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, in set_extent_bits() argument
1273 return set_extent_bit(tree, start, end, bits, NULL, in set_extent_bits()
1277 int clear_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, in clear_extent_bits() argument
1280 return clear_extent_bit(tree, start, end, bits, 0, 0, NULL, mask); in clear_extent_bits()
1283 int set_extent_delalloc(struct extent_io_tree *tree, u64 start, u64 end, in set_extent_delalloc() argument
1286 return set_extent_bit(tree, start, end, in set_extent_delalloc()
1291 int set_extent_defrag(struct extent_io_tree *tree, u64 start, u64 end, in set_extent_defrag() argument
1294 return set_extent_bit(tree, start, end, in set_extent_defrag()
1299 int clear_extent_dirty(struct extent_io_tree *tree, u64 start, u64 end, in clear_extent_dirty() argument
1302 return clear_extent_bit(tree, start, end, in clear_extent_dirty()
1307 int set_extent_new(struct extent_io_tree *tree, u64 start, u64 end, in set_extent_new() argument
1310 return set_extent_bit(tree, start, end, EXTENT_NEW, NULL, in set_extent_new()
1314 int set_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, in set_extent_uptodate() argument
1317 return set_extent_bit(tree, start, end, EXTENT_UPTODATE, NULL, in set_extent_uptodate()
1321 int clear_extent_uptodate(struct extent_io_tree *tree, u64 start, u64 end, in clear_extent_uptodate() argument
1324 return clear_extent_bit(tree, start, end, EXTENT_UPTODATE, 0, 0, in clear_extent_uptodate()
1332 int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, in lock_extent_bits() argument
1339 err = __set_extent_bit(tree, start, end, EXTENT_LOCKED | bits, in lock_extent_bits()
1343 wait_extent_bit(tree, failed_start, end, EXTENT_LOCKED); in lock_extent_bits()
1347 WARN_ON(start > end); in lock_extent_bits()
1352 int lock_extent(struct extent_io_tree *tree, u64 start, u64 end) in lock_extent() argument
1354 return lock_extent_bits(tree, start, end, 0, NULL); in lock_extent()
1357 int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end) in try_lock_extent() argument
1362 err = __set_extent_bit(tree, start, end, EXTENT_LOCKED, EXTENT_LOCKED, in try_lock_extent()
1373 int unlock_extent_cached(struct extent_io_tree *tree, u64 start, u64 end, in unlock_extent_cached() argument
1376 return clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, cached, in unlock_extent_cached()
1380 int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end) in unlock_extent() argument
1382 return clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, NULL, in unlock_extent()
1386 int extent_range_clear_dirty_for_io(struct inode *inode, u64 start, u64 end) in extent_range_clear_dirty_for_io() argument
1389 unsigned long end_index = end >> PAGE_CACHE_SHIFT; in extent_range_clear_dirty_for_io()
1402 int extent_range_redirty_for_io(struct inode *inode, u64 start, u64 end) in extent_range_redirty_for_io() argument
1405 unsigned long end_index = end >> PAGE_CACHE_SHIFT; in extent_range_redirty_for_io()
1422 static int set_range_writeback(struct extent_io_tree *tree, u64 start, u64 end) in set_range_writeback() argument
1425 unsigned long end_index = end >> PAGE_CACHE_SHIFT; in set_range_writeback()
1459 if (state->end >= start && (state->state & bits)) in find_first_extent_bit_state()
1488 if (state->end == start - 1 && extent_state_in_tree(state)) { in find_first_extent_bit()
1510 *end_ret = state->end; in find_first_extent_bit()
1525 u64 *start, u64 *end, u64 max_bytes, in find_delalloc_range() argument
1543 *end = (u64)-1; in find_delalloc_range()
1555 *end = state->end; in find_delalloc_range()
1564 *end = state->end; in find_delalloc_range()
1565 cur_start = state->end + 1; in find_delalloc_range()
1567 total_bytes += state->end - state->start + 1; in find_delalloc_range()
1580 u64 start, u64 end) in __unlock_for_delalloc() argument
1585 unsigned long end_index = end >> PAGE_CACHE_SHIFT; in __unlock_for_delalloc()
1678 u64 *end, u64 max_bytes) in find_lock_delalloc_range() argument
1695 *end = delalloc_end; in find_lock_delalloc_range()
1750 *end = delalloc_end; in find_lock_delalloc_range()
1755 int extent_clear_unlock_delalloc(struct inode *inode, u64 start, u64 end, in extent_clear_unlock_delalloc() argument
1764 unsigned long end_index = end >> PAGE_CACHE_SHIFT; in extent_clear_unlock_delalloc()
1768 clear_extent_bit(tree, start, end, clear_bits, 1, 0, NULL, GFP_NOFS); in extent_clear_unlock_delalloc()
1845 if (state->end >= cur_start && (state->state & bits) == bits) { in count_range_bits()
1846 total_bytes += min(search_end, state->end) + 1 - in count_range_bits()
1854 last = state->end; in count_range_bits()
1931 int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end, in test_range_bit() argument
1940 cached->end > start) in test_range_bit()
1944 while (node && start <= end) { in test_range_bit()
1952 if (state->start > end) in test_range_bit()
1964 if (state->end == (u64)-1) in test_range_bit()
1967 start = state->end + 1; in test_range_bit()
1968 if (start > end) in test_range_bit()
1988 u64 end = start + PAGE_CACHE_SIZE - 1; in check_page_uptodate() local
1989 if (test_range_bit(tree, start, end, EXTENT_UPTODATE, 1, NULL)) in check_page_uptodate()
2153 state->end >= failrec->start + failrec->len - 1) { in clean_io_failure()
2175 void btrfs_free_io_failure_record(struct inode *inode, u64 start, u64 end) in btrfs_free_io_failure_record() argument
2187 if (state->start > end) in btrfs_free_io_failure_record()
2190 ASSERT(state->end <= end); in btrfs_free_io_failure_record()
2203 int btrfs_get_io_failure_record(struct inode *inode, u64 start, u64 end, in btrfs_get_io_failure_record() argument
2222 failrec->len = end - start + 1; in btrfs_get_io_failure_record()
2261 ret = set_extent_bits(failure_tree, start, end, in btrfs_get_io_failure_record()
2268 ret = set_extent_bits(tree, start, end, EXTENT_DAMAGED, in btrfs_get_io_failure_record()
2398 struct page *page, u64 start, u64 end, in bio_readpage_error() argument
2410 ret = btrfs_get_io_failure_record(inode, start, end, &failrec); in bio_readpage_error()
2451 int end_extent_writepage(struct page *page, int err, u64 start, u64 end) in end_extent_writepage() argument
2461 end, NULL, uptodate); in end_extent_writepage()
2488 u64 end; in end_bio_extent_writepage() local
2512 end = start + bvec->bv_offset + bvec->bv_len - 1; in end_bio_extent_writepage()
2514 if (end_extent_writepage(page, err, start, end)) in end_bio_extent_writepage()
2528 u64 end = start + len - 1; in endio_readpage_release_extent() local
2531 set_extent_uptodate(tree, start, end, &cached, GFP_ATOMIC); in endio_readpage_release_extent()
2532 unlock_extent_cached(tree, start, end, &cached, GFP_ATOMIC); in endio_readpage_release_extent()
2554 u64 end; in end_bio_extent_readpage() local
2592 end = start + bvec->bv_offset + bvec->bv_len - 1; in end_bio_extent_readpage()
2599 page, start, end, in end_bio_extent_readpage()
2626 ret = bio_readpage_error(bio, offset, page, start, end, in end_bio_extent_readpage()
2664 end - start + 1, 0); in end_bio_extent_readpage()
2667 extent_len = end + 1 - start; in end_bio_extent_readpage()
2669 extent_len += end + 1 - start; in end_bio_extent_readpage()
2674 extent_len = end + 1 - start; in end_bio_extent_readpage()
2917 u64 end; in __do_readpage() local
2937 end = page_end; in __do_readpage()
2941 unlock_extent(tree, start, end); in __do_readpage()
2958 while (cur <= end) { in __do_readpage()
2980 end - cur + 1, get_extent, em_cached); in __do_readpage()
2984 unlock_extent(tree, cur, end); in __do_readpage()
2989 BUG_ON(end < cur); in __do_readpage()
2997 iosize = min(extent_map_end(em) - cur, end - cur + 1); in __do_readpage()
2998 cur_end = min(extent_map_end(em) - 1, end); in __do_readpage()
3127 u64 start, u64 end, in __do_contiguous_readpages() argument
3140 lock_extent(tree, start, end); in __do_contiguous_readpages()
3142 end - start + 1); in __do_contiguous_readpages()
3145 unlock_extent(tree, start, end); in __do_contiguous_readpages()
3166 u64 end = 0; in __extent_readpages() local
3173 if (!end) { in __extent_readpages()
3175 end = start + PAGE_CACHE_SIZE - 1; in __extent_readpages()
3177 } else if (end + 1 == page_start) { in __extent_readpages()
3178 end += PAGE_CACHE_SIZE; in __extent_readpages()
3182 end, get_extent, em_cached, in __extent_readpages()
3186 end = start + PAGE_CACHE_SIZE - 1; in __extent_readpages()
3191 if (end) in __extent_readpages()
3194 end, get_extent, em_cached, bio, in __extent_readpages()
3208 u64 end = start + PAGE_CACHE_SIZE - 1; in __extent_read_full_page() local
3212 lock_extent(tree, start, end); in __extent_read_full_page()
3216 unlock_extent(tree, start, end); in __extent_read_full_page()
3374 u64 end; in __extent_writepage_io() local
3412 end = page_end; in __extent_writepage_io()
3422 while (cur <= end) { in __extent_writepage_io()
3431 end - cur + 1, 1); in __extent_writepage_io()
3441 BUG_ON(end < cur); in __extent_writepage_io()
3442 iosize = min(em_end - cur, end - cur + 1); in __extent_writepage_io()
3494 page->index, cur, end); in __extent_writepage_io()
3848 pgoff_t end; /* Inclusive */ in btree_write_cache_pages() local
3855 end = -1; in btree_write_cache_pages()
3858 end = wbc->range_end >> PAGE_CACHE_SHIFT; in btree_write_cache_pages()
3867 tag_pages_for_writeback(mapping, index, end); in btree_write_cache_pages()
3868 while (!done && !nr_to_write_done && (index <= end) && in btree_write_cache_pages()
3870 min(end - index, (pgoff_t)PAGEVEC_SIZE-1) + 1))) { in btree_write_cache_pages()
3880 if (!wbc->range_cyclic && page->index > end) { in btree_write_cache_pages()
3980 pgoff_t end; /* Inclusive */ in extent_write_cache_pages() local
3999 end = -1; in extent_write_cache_pages()
4002 end = wbc->range_end >> PAGE_CACHE_SHIFT; in extent_write_cache_pages()
4011 tag_pages_for_writeback(mapping, index, end); in extent_write_cache_pages()
4012 while (!done && !nr_to_write_done && (index <= end) && in extent_write_cache_pages()
4014 min(end - index, (pgoff_t)PAGEVEC_SIZE-1) + 1))) { in extent_write_cache_pages()
4038 if (!wbc->range_cyclic && page->index > end) { in extent_write_cache_pages()
4130 u64 start, u64 end, get_extent_t *get_extent, in extent_write_locked_range() argument
4136 unsigned long nr_pages = (end - start + PAGE_CACHE_SIZE) >> in extent_write_locked_range()
4151 .range_end = end + 1, in extent_write_locked_range()
4154 while (start <= end) { in extent_write_locked_range()
4250 u64 end = start + PAGE_CACHE_SIZE - 1; in extent_invalidatepage() local
4254 if (start > end) in extent_invalidatepage()
4257 lock_extent_bits(tree, start, end, 0, &cached_state); in extent_invalidatepage()
4259 clear_extent_bit(tree, start, end, in extent_invalidatepage()
4276 u64 end = start + PAGE_CACHE_SIZE - 1; in try_release_extent_state() local
4279 if (test_range_bit(tree, start, end, in try_release_extent_state()
4289 ret = clear_extent_bit(tree, start, end, in try_release_extent_state()
4315 u64 end = start + PAGE_CACHE_SIZE - 1; in try_release_extent_mapping() local
4320 while (start <= end) { in try_release_extent_mapping()
4321 len = end - start + 1; in try_release_extent_mapping()
4409 int end = 0; in extent_fiemap() local
4479 while (!end) { in extent_fiemap()
4512 end = 1; in extent_fiemap()
4515 end = 1; in extent_fiemap()
4553 end = 1; in extent_fiemap()
4565 end = 1; in extent_fiemap()