/kernel/linux/linux-6.6/fs/ntfs3/ |
H A D | attrib.c | 64 CLST svcn = le64_to_cpu(attr->nres.svcn); in attr_load_runs() local 69 if (svcn >= evcn + 1 || run_is_mapped_full(run, svcn, evcn)) in attr_load_runs() 72 if (vcn && (evcn < *vcn || *vcn < svcn)) in attr_load_runs() 81 err = run_unpack_ex(run, ni->mi.sbi, ni->mi.rno, svcn, evcn, in attr_load_runs() 82 vcn ? *vcn : svcn, Add2Ptr(attr, run_off), in attr_load_runs() 414 CLST alen, vcn, lcn, new_alen, old_alen, svcn, evcn; in attr_set_size() local 470 svcn = le64_to_cpu(attr_b->nres.svcn); in attr_set_size() 892 CLST hint, svcn, to_alloc, evcn1, next_svcn, asize, end, vcn0, alen; attr_data_get_block() local 1305 CLST svcn, evcn; attr_load_runs_vcn() local 1630 CLST svcn, evcn1, next_svcn, len; attr_allocate_frame() local 1855 CLST svcn, evcn1, len, dealloc, alen; attr_collapse_range() local 2113 CLST svcn, evcn1, vcn, len, end, alen, hole, next_svcn; attr_punch_hole() local 2331 CLST vcn, svcn, evcn1, len, next_svcn; attr_insert_range() local [all...] |
H A D | frecord.c | 227 if (attr->nres.svcn) in ni_find_attr() 229 } else if (le64_to_cpu(attr->nres.svcn) > *vcn || in ni_find_attr() 327 if (le64_to_cpu(attr->nres.svcn) <= vcn && in ni_load_attr() 459 CLST svcn, struct ATTR_LIST_ENTRY **ins_le) in ni_ins_new_attr() 469 err = al_add_le(ni, type, name, name_len, svcn, cpu_to_le16(-1), in ni_ins_new_attr() 531 CLST svcn, evcn = 0, svcn_p, evcn_p, next_svcn; in ni_repack() 541 svcn = le64_to_cpu(attr->nres.svcn); in ni_repack() 542 if (svcn != le64_to_cpu(le->vcn)) { in ni_repack() 547 if (!svcn) { in ni_repack() 456 ni_ins_new_attr(struct ntfs_inode *ni, struct mft_inode *mi, struct ATTR_LIST_ENTRY *le, enum ATTR_TYPE type, const __le16 *name, u8 name_len, u32 asize, u16 name_off, CLST svcn, struct ATTR_LIST_ENTRY **ins_le) ni_ins_new_attr() argument 939 ni_ins_attr_ext(struct ntfs_inode *ni, struct ATTR_LIST_ENTRY *le, enum ATTR_TYPE type, const __le16 *name, u8 name_len, u32 asize, CLST svcn, u16 name_off, bool force_ext, struct ATTRIB **ins_attr, struct mft_inode **ins_mi, struct ATTR_LIST_ENTRY **ins_le) ni_ins_attr_ext() argument 1090 ni_insert_attr(struct ntfs_inode *ni, enum ATTR_TYPE type, const __le16 *name, u8 name_len, u32 asize, u16 name_off, CLST svcn, struct ATTRIB **ins_attr, struct mft_inode **ins_mi, struct ATTR_LIST_ENTRY **ins_le) ni_insert_attr() argument 1248 CLST mft_min, mft_new, svcn, evcn, plen; ni_expand_mft_list() local 1439 ni_insert_nonresident(struct ntfs_inode *ni, enum ATTR_TYPE type, const __le16 *name, u8 name_len, const struct runs_tree *run, CLST svcn, CLST len, __le16 flags, struct ATTRIB **new_attr, struct mft_inode **mi, struct ATTR_LIST_ENTRY **le) ni_insert_nonresident() argument 1575 CLST svcn, evcn; ni_delete_all() local 2280 CLST svcn, evcn; ni_decompress_file() local [all...] |
H A D | run.c | 165 * Return: True if range [svcn - evcn] is mapped. 167 bool run_is_mapped_full(const struct runs_tree *run, CLST svcn, CLST evcn) in run_is_mapped_full() argument 173 if (!run_lookup(run, svcn, &i)) in run_is_mapped_full() 817 int run_pack(const struct runs_tree *run, CLST svcn, CLST len, u8 *run_buf, in run_pack() argument 822 CLST evcn1 = svcn + len; in run_pack() 834 /* Check all required entries [svcn, encv1) available. */ in run_pack() 835 if (!run_lookup(run, svcn, &i)) in run_pack() 849 len = svcn - r->vcn; in run_pack() 850 vcn = svcn; in run_pack() 921 CLST svcn, CLS in run_unpack() 920 run_unpack(struct runs_tree *run, struct ntfs_sb_info *sbi, CLST ino, CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, int run_buf_size) run_unpack() argument 1050 run_unpack_ex(struct runs_tree *run, struct ntfs_sb_info *sbi, CLST ino, CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, int run_buf_size) run_unpack_ex() argument [all...] |
H A D | attrlist.c | 62 } else if (attr->nres.svcn) { in ntfs_load_attr_list() 172 CLST svcn = attr_svcn(attr); in al_find_le() local 175 &svcn); in al_find_le() 287 u8 name_len, CLST svcn, __le16 id, const struct MFT_REF *ref, in al_add_le() 309 le = al_find_le_to_insert(ni, type, name, name_len, svcn); in al_add_le() 334 le->vcn = cpu_to_le64(svcn); in al_add_le() 286 al_add_le(struct ntfs_inode *ni, enum ATTR_TYPE type, const __le16 *name, u8 name_len, CLST svcn, __le16 id, const struct MFT_REF *ref, struct ATTR_LIST_ENTRY **new_le) al_add_le() argument
|
H A D | record.c | 305 if (le64_to_cpu(attr->nres.svcn) > le64_to_cpu(attr->nres.evcn) + 1) in mi_enum_attr() 320 if (!attr->nres.svcn && is_attr_ext(attr)) { in mi_enum_attr() 624 CLST svcn = le64_to_cpu(attr->nres.svcn); in mi_pack_runs() local 638 err = run_pack(run, svcn, len, Add2Ptr(attr, run_off), run_size + dsize, in mi_pack_runs() 650 attr->nres.evcn = cpu_to_le64(svcn + plen - 1); in mi_pack_runs()
|
H A D | ntfs_fs.h | 470 u8 name_len, CLST svcn, __le16 id, const struct MFT_REF *ref, 537 const struct runs_tree *run, CLST svcn, CLST len, 802 bool run_is_mapped_full(const struct runs_tree *run, CLST svcn, CLST evcn); 804 int run_pack(const struct runs_tree *run, CLST svcn, CLST len, u8 *run_buf, 807 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf, 812 CLST svcn, CLST evcn, CLST vcn, const u8 *run_buf,
|
H A D | fslog.c | 2675 u64 dsize, svcn, evcn; in check_attr() local 2699 svcn = le64_to_cpu(attr->nres.svcn); in check_attr() 2703 if (svcn > evcn + 1 || run_off >= asize || in check_attr() 2712 if (run_unpack(NULL, sbi, 0, svcn, evcn, svcn, in check_attr() 3336 if (run_get_highest_vcn(le64_to_cpu(attr->nres.svcn), in do_action() 4738 CLST svcn = le64_to_cpu(attr->nres.svcn); in log_replay() local 4746 err = run_unpack(&oa->run0, sbi, inode->i_ino, svcn, in log_replay() [all...] |
H A D | ntfs.h | 338 __le64 svcn; // 0x10: Starting VCN of this segment. member 442 return attr->non_res ? le64_to_cpu(attr->nres.svcn) : 0; in attr_svcn()
|
H A D | inode.c | 398 t64 = le64_to_cpu(attr->nres.svcn); in ntfs_read_mft()
|