tbh               335 fs/ntfs/aops.c 		struct buffer_head *tbh;
tbh               339 fs/ntfs/aops.c 			tbh = arr[i];
tbh               340 fs/ntfs/aops.c 			lock_buffer(tbh);
tbh               341 fs/ntfs/aops.c 			tbh->b_end_io = ntfs_end_buffer_async_read;
tbh               342 fs/ntfs/aops.c 			set_buffer_async_read(tbh);
tbh               346 fs/ntfs/aops.c 			tbh = arr[i];
tbh               347 fs/ntfs/aops.c 			if (likely(!buffer_uptodate(tbh)))
tbh               348 fs/ntfs/aops.c 				submit_bh(REQ_OP_READ, 0, tbh);
tbh               350 fs/ntfs/aops.c 				ntfs_end_buffer_async_read(tbh, 1);
tbh               912 fs/ntfs/aops.c 	struct buffer_head *bh, *head, *tbh, *rec_start_bh;
tbh              1115 fs/ntfs/aops.c 		tbh = bhs[i];
tbh              1116 fs/ntfs/aops.c 		ofs = bh_offset(tbh);
tbh              1180 fs/ntfs/aops.c 		tbh = bhs[i];
tbh              1181 fs/ntfs/aops.c 		if (!tbh)
tbh              1183 fs/ntfs/aops.c 		if (!trylock_buffer(tbh))
tbh              1186 fs/ntfs/aops.c 		clear_buffer_dirty(tbh);
tbh              1187 fs/ntfs/aops.c 		BUG_ON(!buffer_uptodate(tbh));
tbh              1188 fs/ntfs/aops.c 		BUG_ON(!buffer_mapped(tbh));
tbh              1189 fs/ntfs/aops.c 		get_bh(tbh);
tbh              1190 fs/ntfs/aops.c 		tbh->b_end_io = end_buffer_write_sync;
tbh              1191 fs/ntfs/aops.c 		submit_bh(REQ_OP_WRITE, 0, tbh);
tbh              1199 fs/ntfs/aops.c 		tbh = bhs[i];
tbh              1200 fs/ntfs/aops.c 		if (!tbh)
tbh              1202 fs/ntfs/aops.c 		wait_on_buffer(tbh);
tbh              1203 fs/ntfs/aops.c 		if (unlikely(!buffer_uptodate(tbh))) {
tbh              1209 fs/ntfs/aops.c 					page->index, bh_offset(tbh));
tbh              1216 fs/ntfs/aops.c 			set_buffer_uptodate(tbh);
tbh              1232 fs/ntfs/aops.c 			tbh = bhs[i];
tbh              1234 fs/ntfs/aops.c 			if (!tbh)
tbh              1236 fs/ntfs/aops.c 			ofs = bh_offset(tbh);
tbh              1251 fs/ntfs/aops.c 			tbh = bhs[i];
tbh              1252 fs/ntfs/aops.c 			if (!tbh)
tbh              1255 fs/ntfs/aops.c 					bh_offset(tbh)));
tbh               651 fs/ntfs/compress.c 		struct buffer_head *tbh = bhs[i];
tbh               653 fs/ntfs/compress.c 		if (!trylock_buffer(tbh))
tbh               655 fs/ntfs/compress.c 		if (unlikely(buffer_uptodate(tbh))) {
tbh               656 fs/ntfs/compress.c 			unlock_buffer(tbh);
tbh               659 fs/ntfs/compress.c 		get_bh(tbh);
tbh               660 fs/ntfs/compress.c 		tbh->b_end_io = end_buffer_read_sync;
tbh               661 fs/ntfs/compress.c 		submit_bh(REQ_OP_READ, 0, tbh);
tbh               666 fs/ntfs/compress.c 		struct buffer_head *tbh = bhs[i];
tbh               668 fs/ntfs/compress.c 		if (buffer_uptodate(tbh))
tbh               670 fs/ntfs/compress.c 		wait_on_buffer(tbh);
tbh               680 fs/ntfs/compress.c 		if (unlikely(!buffer_uptodate(tbh))) {
tbh               684 fs/ntfs/compress.c 			get_bh(tbh);
tbh               686 fs/ntfs/compress.c 			put_bh(tbh);
tbh               687 fs/ntfs/compress.c 			if (unlikely(!buffer_uptodate(tbh)))
tbh               578 fs/ntfs/mft.c  			struct buffer_head *tbh = bhs[i_bhs];
tbh               580 fs/ntfs/mft.c  			if (!trylock_buffer(tbh))
tbh               582 fs/ntfs/mft.c  			BUG_ON(!buffer_uptodate(tbh));
tbh               583 fs/ntfs/mft.c  			clear_buffer_dirty(tbh);
tbh               584 fs/ntfs/mft.c  			get_bh(tbh);
tbh               585 fs/ntfs/mft.c  			tbh->b_end_io = end_buffer_write_sync;
tbh               586 fs/ntfs/mft.c  			submit_bh(REQ_OP_WRITE, 0, tbh);
tbh               590 fs/ntfs/mft.c  			struct buffer_head *tbh = bhs[i_bhs];
tbh               592 fs/ntfs/mft.c  			wait_on_buffer(tbh);
tbh               593 fs/ntfs/mft.c  			if (unlikely(!buffer_uptodate(tbh))) {
tbh               599 fs/ntfs/mft.c  				set_buffer_uptodate(tbh);
tbh               775 fs/ntfs/mft.c  		struct buffer_head *tbh = bhs[i_bhs];
tbh               777 fs/ntfs/mft.c  		if (!trylock_buffer(tbh))
tbh               779 fs/ntfs/mft.c  		BUG_ON(!buffer_uptodate(tbh));
tbh               780 fs/ntfs/mft.c  		clear_buffer_dirty(tbh);
tbh               781 fs/ntfs/mft.c  		get_bh(tbh);
tbh               782 fs/ntfs/mft.c  		tbh->b_end_io = end_buffer_write_sync;
tbh               783 fs/ntfs/mft.c  		submit_bh(REQ_OP_WRITE, 0, tbh);
tbh               790 fs/ntfs/mft.c  		struct buffer_head *tbh = bhs[i_bhs];
tbh               792 fs/ntfs/mft.c  		wait_on_buffer(tbh);
tbh               793 fs/ntfs/mft.c  		if (unlikely(!buffer_uptodate(tbh))) {
tbh               800 fs/ntfs/mft.c  				set_buffer_uptodate(tbh);
tbh               977 fs/reiserfs/journal.c 	struct buffer_head *tbh = NULL;
tbh              1055 fs/reiserfs/journal.c 		tbh = journal_find_get_block(s, bn);
tbh              1056 fs/reiserfs/journal.c 		if (tbh) {
tbh              1057 fs/reiserfs/journal.c 			if (buffer_dirty(tbh)) {
tbh              1059 fs/reiserfs/journal.c 			    ll_rw_block(REQ_OP_WRITE, 0, 1, &tbh);
tbh              1062 fs/reiserfs/journal.c 			put_bh(tbh) ;
tbh              1070 fs/reiserfs/journal.c 		tbh = journal_find_get_block(s, bn);
tbh              1073 fs/reiserfs/journal.c 		__wait_on_buffer(tbh);
tbh              1080 fs/reiserfs/journal.c 		if (buffer_dirty(tbh)) {
tbh              1082 fs/reiserfs/journal.c 			sync_dirty_buffer(tbh);
tbh              1085 fs/reiserfs/journal.c 		if (unlikely(!buffer_uptodate(tbh))) {
tbh              1093 fs/reiserfs/journal.c 		put_bh(tbh);
tbh              1095 fs/reiserfs/journal.c 		put_bh(tbh);