H A D | attrib.c | 61 struct runs_tree *run, const CLST *vcn) in attr_load_runs() 72 if (vcn && (evcn < *vcn || *vcn < svcn)) in attr_load_runs() 82 vcn ? *vcn : svcn, Add2Ptr(attr, run_off), in attr_load_runs() 94 CLST vcn, CLST len, CLST *done, bool trim) in run_deallocate_ex() 97 CLST vcn_next, vcn0 = vcn, lcn, clen, dn = 0; in run_deallocate_ex() 103 if (!run_lookup_entry(run, vcn, &lcn, &clen, &idx)) { in run_deallocate_ex() 131 vcn_next = vcn in run_deallocate_ex() 60 attr_load_runs(struct ATTRIB *attr, struct ntfs_inode *ni, struct runs_tree *run, const CLST *vcn) attr_load_runs() argument 93 run_deallocate_ex(struct ntfs_sb_info *sbi, struct runs_tree *run, CLST vcn, CLST len, CLST *done, bool trim) run_deallocate_ex() argument 149 attr_allocate_clusters(struct ntfs_sb_info *sbi, struct runs_tree *run, CLST vcn, CLST lcn, CLST len, CLST *pre_alloc, enum ALLOCATE_OPT opt, CLST *alen, const size_t fr, CLST *new_lcn, CLST *new_len) attr_allocate_clusters() argument 414 CLST alen, vcn, lcn, new_alen, old_alen, svcn, evcn; attr_set_size() local 882 attr_data_get_block(struct ntfs_inode *ni, CLST vcn, CLST clen, CLST *lcn, CLST *len, bool *new, bool zero) attr_data_get_block() argument 1299 attr_load_runs_vcn(struct ntfs_inode *ni, enum ATTR_TYPE type, const __le16 *name, u8 name_len, struct runs_tree *run, CLST vcn) attr_load_runs_vcn() argument 1348 CLST vcn; attr_load_runs_range() local 1522 CLST clen, lcn, vcn, alen, slen, vcn_next; attr_is_frame_compressed() local 1631 CLST vcn, end, clst_data; attr_allocate_frame() local 1856 CLST vcn, end; attr_collapse_range() local 2113 CLST svcn, evcn1, vcn, len, end, alen, hole, next_svcn; attr_punch_hole() local 2331 CLST vcn, svcn, evcn1, len, next_svcn; attr_insert_range() local [all...] |