start_vcn 481 fs/ntfs/compress.c VCN start_vcn = (((s64)index << PAGE_SHIFT) & ~cb_size_mask) >> start_vcn 490 fs/ntfs/compress.c unsigned int nr_cbs = (end_vcn - start_vcn) << vol->cluster_size_bits start_vcn 497 fs/ntfs/compress.c unsigned int nr_pages = (end_vcn - start_vcn) << start_vcn 535 fs/ntfs/compress.c offset = start_vcn << vol->cluster_size_bits >> PAGE_SHIFT; start_vcn 596 fs/ntfs/compress.c for (vcn = start_vcn, start_vcn += cb_clusters; vcn < start_vcn; start_vcn 730 fs/ntfs/compress.c if (vcn == start_vcn - cb_clusters) { start_vcn 773 fs/ntfs/compress.c } else if (vcn == start_vcn) { start_vcn 132 fs/ntfs/lcnalloc.c runlist_element *ntfs_cluster_alloc(ntfs_volume *vol, const VCN start_vcn, start_vcn 150 fs/ntfs/lcnalloc.c "0x%llx, zone %s_ZONE.", (unsigned long long)start_vcn, start_vcn 157 fs/ntfs/lcnalloc.c BUG_ON(start_vcn < 0); start_vcn 393 fs/ntfs/lcnalloc.c rl[rlpos].vcn = start_vcn; start_vcn 835 fs/ntfs/lcnalloc.c s64 __ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn, s64 count, start_vcn 846 fs/ntfs/lcnalloc.c "0x%llx.%s", ni->mft_no, (unsigned long long)start_vcn, start_vcn 852 fs/ntfs/lcnalloc.c BUG_ON(start_vcn < 0); start_vcn 866 fs/ntfs/lcnalloc.c rl = ntfs_attr_find_vcn_nolock(ni, start_vcn, ctx); start_vcn 883 fs/ntfs/lcnalloc.c delta = start_vcn - rl->vcn; start_vcn 988 fs/ntfs/lcnalloc.c delta = __ntfs_cluster_free(ni, start_vcn, total_freed, ctx, true); start_vcn 30 fs/ntfs/lcnalloc.h const VCN start_vcn, const s64 count, const LCN start_lcn, start_vcn 34 fs/ntfs/lcnalloc.h extern s64 __ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn, start_vcn 93 fs/ntfs/lcnalloc.h static inline s64 ntfs_cluster_free(ntfs_inode *ni, const VCN start_vcn, start_vcn 96 fs/ntfs/lcnalloc.h return __ntfs_cluster_free(ni, start_vcn, count, ctx, false);