start_vcn         481 fs/ntfs/compress.c 	VCN start_vcn = (((s64)index << PAGE_SHIFT) & ~cb_size_mask) >>
start_vcn         490 fs/ntfs/compress.c 	unsigned int nr_cbs = (end_vcn - start_vcn) << vol->cluster_size_bits
start_vcn         497 fs/ntfs/compress.c 	unsigned int nr_pages = (end_vcn - start_vcn) <<
start_vcn         535 fs/ntfs/compress.c 	offset = start_vcn << vol->cluster_size_bits >> PAGE_SHIFT;
start_vcn         596 fs/ntfs/compress.c 	for (vcn = start_vcn, start_vcn += cb_clusters; vcn < start_vcn;
start_vcn         730 fs/ntfs/compress.c 	if (vcn == start_vcn - cb_clusters) {
start_vcn         773 fs/ntfs/compress.c 	} else if (vcn == start_vcn) {
start_vcn         132 fs/ntfs/lcnalloc.c runlist_element *ntfs_cluster_alloc(ntfs_volume *vol, const VCN start_vcn,
start_vcn         150 fs/ntfs/lcnalloc.c 			"0x%llx, zone %s_ZONE.", (unsigned long long)start_vcn,
start_vcn         157 fs/ntfs/lcnalloc.c 	BUG_ON(start_vcn < 0);
start_vcn         393 fs/ntfs/lcnalloc.c 					rl[rlpos].vcn = start_vcn;
start_vcn         835 fs/ntfs/lcnalloc.c s64 __ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn, s64 count,
start_vcn         846 fs/ntfs/lcnalloc.c 			"0x%llx.%s", ni->mft_no, (unsigned long long)start_vcn,
start_vcn         852 fs/ntfs/lcnalloc.c 	BUG_ON(start_vcn < 0);
start_vcn         866 fs/ntfs/lcnalloc.c 	rl = ntfs_attr_find_vcn_nolock(ni, start_vcn, ctx);
start_vcn         883 fs/ntfs/lcnalloc.c 	delta = start_vcn - rl->vcn;
start_vcn         988 fs/ntfs/lcnalloc.c 	delta = __ntfs_cluster_free(ni, start_vcn, total_freed, ctx, true);
start_vcn          30 fs/ntfs/lcnalloc.h 		const VCN start_vcn, const s64 count, const LCN start_lcn,
start_vcn          34 fs/ntfs/lcnalloc.h extern s64 __ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn,
start_vcn          93 fs/ntfs/lcnalloc.h static inline s64 ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn,
start_vcn          96 fs/ntfs/lcnalloc.h 	return __ntfs_cluster_free(ni, start_vcn, count, ctx, false);