/kernel/linux/linux-6.6/fs/ntfs3/ |
H A D | run.c | 23 CLST lcn; /* Logical cluster number. */ member 124 if (n->lcn != SPARSE_LCN) in run_consolidate() 125 n->lcn += dl; in run_consolidate() 133 if ((n->lcn == SPARSE_LCN) != (r->lcn == SPARSE_LCN)) { in run_consolidate() 141 * of a next run lcn does not match in run_consolidate() 144 if (n->lcn != SPARSE_LCN && n->lcn != r->lcn + r->len) in run_consolidate() 192 bool run_lookup_entry(const struct runs_tree *run, CLST vcn, CLST *lcn, in run_lookup_entry() argument 316 run_add_entry(struct runs_tree *run, CLST vcn, CLST lcn, CLST len, bool is_mft) run_add_entry() argument 595 run_get_entry(const struct runs_tree *run, size_t index, CLST *vcn, CLST *lcn, CLST *len) run_get_entry() argument 820 CLST next_vcn, vcn, lcn; run_pack() local 924 u64 prev_lcn, vcn64, lcn, next_vcn; run_unpack() local 1055 CLST next_vcn, lcn, len; run_unpack_ex() local [all...] |
H A D | fsntfs.c | 352 int ntfs_look_for_free_space(struct ntfs_sb_info *sbi, CLST lcn, CLST len, in ntfs_look_for_free_space() argument 380 lcn = wnd_zone_bit(wnd); in ntfs_look_for_free_space() 383 wnd_zone_set(wnd, lcn + alen, zlen - alen); in ntfs_look_for_free_space() 385 err = wnd_set_used(wnd, lcn, alen); in ntfs_look_for_free_space() 389 alcn = lcn; in ntfs_look_for_free_space() 396 if (!lcn) in ntfs_look_for_free_space() 397 lcn = sbi->used.next_free_lcn; in ntfs_look_for_free_space() 399 if (lcn >= wnd->nbits) in ntfs_look_for_free_space() 400 lcn = 0; in ntfs_look_for_free_space() 402 alen = wnd_find(wnd, len, lcn, BITMAP_FIND_MARK_AS_USE in ntfs_look_for_free_space() 815 CLST lcn, vcn, len; ntfs_refresh_zone() local 1133 CLST lcn, clen, vcn = vbo >> cluster_bits, vcn_next; ntfs_sb_write_run() local 1180 CLST lcn; ntfs_bread_run() local 1201 CLST lcn, clen; ntfs_read_run_nb() local 1330 CLST lcn, clen; ntfs_get_bh() local 1507 CLST lcn, clen, vcn, vcn_next; ntfs_bio_pages() local 1603 CLST lcn, clen; ntfs_bio_fill_1() local 1670 CLST lcn, len; ntfs_vbo_to_lbo() local 2471 ntfs_unmap_and_discard(struct ntfs_sb_info *sbi, CLST lcn, CLST len) ntfs_unmap_and_discard() argument 2478 mark_as_free_ex(struct ntfs_sb_info *sbi, CLST lcn, CLST len, bool trim) mark_as_free_ex() argument 2551 CLST lcn, len; run_deallocate() local [all...] |
H A D | attrib.c | 97 CLST vcn_next, vcn0 = vcn, lcn, clen, dn = 0; in run_deallocate_ex() local 103 if (!run_lookup_entry(run, vcn, &lcn, &clen, &idx)) { in run_deallocate_ex() 119 if (lcn != SPARSE_LCN) { in run_deallocate_ex() 121 /* mark bitmap range [lcn + clen) as free and trim clusters. */ in run_deallocate_ex() 122 mark_as_free_ex(sbi, lcn, clen, trim); in run_deallocate_ex() 132 if (!run_get_entry(run, ++idx, &vcn, &lcn, &clen) || in run_deallocate_ex() 150 CLST vcn, CLST lcn, CLST len, CLST *pre_alloc, in attr_allocate_clusters() 159 err = ntfs_look_for_free_space(sbi, lcn, len + pre, &lcn, &flen, in attr_allocate_clusters() 175 *new_lcn = lcn; in attr_allocate_clusters() 149 attr_allocate_clusters(struct ntfs_sb_info *sbi, struct runs_tree *run, CLST vcn, CLST lcn, CLST len, CLST *pre_alloc, enum ALLOCATE_OPT opt, CLST *alen, const size_t fr, CLST *new_lcn, CLST *new_len) attr_allocate_clusters() argument 414 CLST alen, vcn, lcn, new_alen, old_alen, svcn, evcn; attr_set_size() local 882 attr_data_get_block(struct ntfs_inode *ni, CLST vcn, CLST clen, CLST *lcn, CLST *len, bool *new, bool zero) attr_data_get_block() argument 1350 CLST lcn, clen; attr_load_runs_range() local 1522 CLST clen, lcn, vcn, alen, slen, vcn_next; attr_is_frame_compressed() local [all...] |
H A D | file.c | 120 CLST vcn, lcn, clen; in ntfs_extend_initialized_size() local 126 err = attr_data_get_block(ni, vcn, 1, &lcn, &clen, NULL, in ntfs_extend_initialized_size() 131 if (lcn == SPARSE_LCN) { in ntfs_extend_initialized_size() 288 CLST lcn, len; in ntfs_file_mmap() local 294 err = attr_data_get_block(ni, vcn, 1, &lcn, in ntfs_file_mmap() 610 CLST lcn, clen; in ntfs_fallocate() local 622 &lcn, &clen, &new, in ntfs_fallocate() 632 &lcn, &clen, &new, in ntfs_fallocate() 889 CLST lcn, clen; in ntfs_compress_write() local 895 err = attr_data_get_block(ni, frame << NTFS_LZNT_CUNIT, 1, &lcn, in ntfs_compress_write() [all...] |
H A D | bitmap.c | 510 CLST lcn, clen; in wnd_rescan() local 546 &lcn, &clen, NULL)) { in wnd_rescan() 551 lbo = ((u64)lcn << cluster_bits) + off; in wnd_rescan() 683 CLST lcn, clen; in wnd_map() local 692 if (!run_lookup_entry(&wnd->run, vbo >> sbi->cluster_bits, &lcn, &clen, in wnd_map() 697 lbo = ((u64)lcn << sbi->cluster_bits) + (vbo & sbi->cluster_mask); in wnd_map() 1416 void wnd_zone_set(struct wnd_bitmap *wnd, size_t lcn, size_t len) in wnd_zone_set() argument 1424 wnd_remove_free_ext(wnd, lcn, len); in wnd_zone_set() 1426 wnd->zone_bit = lcn; in wnd_zone_set() 1427 wnd->zone_end = lcn in wnd_zone_set() 1436 CLST len = 0, lcn = 0, done = 0; ntfs_trim_fs() local [all...] |
/kernel/linux/linux-5.10/fs/ntfs/ |
H A D | runlist.c | 148 if ((dst->lcn == LCN_RL_NOT_MAPPED) && (src->lcn == LCN_RL_NOT_MAPPED)) in ntfs_are_rl_mergeable() 154 if ((dst->lcn >= 0) && (src->lcn >= 0) && in ntfs_are_rl_mergeable() 155 ((dst->lcn + dst->length) == src->lcn)) in ntfs_are_rl_mergeable() 158 if ((dst->lcn == LCN_HOLE) && (src->lcn == LCN_HOLE)) in ntfs_are_rl_mergeable() 241 if (dst[marker].lcn == LCN_ENOENT) in ntfs_rl_append() 327 if (dst[marker].lcn in ntfs_rl_insert() 739 LCN lcn; /* Current lcn. */ ntfs_mapping_pairs_decompress() local [all...] |
H A D | lcnalloc.c | 33 * Locking: - The volume lcn bitmap must be locked for writing on entry and is 48 if (rl->lcn < 0) in ntfs_cluster_free_from_rl_nolock() 50 err = ntfs_bitmap_clear_run(lcnbmp_vi, rl->lcn, rl->length); in ntfs_cluster_free_from_rl_nolock() 63 * @start_lcn: starting lcn at which to allocate the clusters (or -1 if none) 100 * access to the lcn bitmap in portions of up to 8192 bytes at a time, so we 127 * Locking: - The volume lcn bitmap must be unlocked on entry and is unlocked 129 * - This function takes the volume lcn bitmap lock for writing and 137 LCN zone_start, zone_end, bmp_pos, bmp_initial_pos, last_read_pos, lcn; in ntfs_cluster_alloc() local 284 lcn = bmp_pos & 7; in ntfs_cluster_alloc() 286 ntfs_debug("Before inner while loop: buf_size %i, lcn in ntfs_cluster_alloc() [all...] |
H A D | aops.c | 168 LCN lcn; in ntfs_read_block() local 255 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_block() 257 lcn = LCN_RL_NOT_MAPPED; in ntfs_read_block() 259 if (lcn >= 0) { in ntfs_read_block() 261 bh->b_blocknr = ((lcn << vol->cluster_size_bits) in ntfs_read_block() 273 if (lcn == LCN_HOLE) in ntfs_read_block() 276 if (!is_retry && lcn == LCN_RL_NOT_MAPPED) { in ntfs_read_block() 294 if (err == -ENOENT || lcn == LCN_ENOENT) { in ntfs_read_block() 533 LCN lcn; in ntfs_write_block() local 711 lcn in ntfs_write_block() 1003 LCN lcn; ntfs_write_mst_block() local 1550 LCN lcn; ntfs_bmap() local [all...] |
H A D | debug.c | 135 LCN lcn = (rl + i)->lcn; in ntfs_debug_dump_runlist() local 137 if (lcn < (LCN)0) { in ntfs_debug_dump_runlist() 138 int index = -lcn - 1; in ntfs_debug_dump_runlist() 150 (long long)(rl + i)->lcn, in ntfs_debug_dump_runlist()
|
H A D | attrib.c | 298 * ntfs_attr_vcn_to_lcn_nolock - convert a vcn into a lcn given an ntfs inode 305 * number (lcn). 308 * extent containing the @vcn and the vcn to lcn conversion is retried. 330 LCN lcn; in ntfs_attr_vcn_to_lcn_nolock() local 349 /* Convert vcn to lcn. If that fails map the runlist and retry once. */ in ntfs_attr_vcn_to_lcn_nolock() 350 lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn); in ntfs_attr_vcn_to_lcn_nolock() 351 if (likely(lcn >= LCN_HOLE)) { in ntfs_attr_vcn_to_lcn_nolock() 352 ntfs_debug("Done, lcn 0x%llx.", (long long)lcn); in ntfs_attr_vcn_to_lcn_nolock() 353 return lcn; in ntfs_attr_vcn_to_lcn_nolock() 723 LCN lcn; load_attribute_list() local [all...] |
H A D | file.c | 574 LCN lcn; in ntfs_prepare_pages_for_non_resident_write() local 628 vcn = lcn = -1; in ntfs_prepare_pages_for_non_resident_write() 703 * clusters, @lcn is the device cluster corresponding to @vcn, in ntfs_prepare_pages_for_non_resident_write() 704 * and @lcn_block is the block number corresponding to @lcn. in ntfs_prepare_pages_for_non_resident_write() 709 BUG_ON(lcn < 0); in ntfs_prepare_pages_for_non_resident_write() 838 lcn = ntfs_rl_vcn_to_lcn(rl, bh_cpos); in ntfs_prepare_pages_for_non_resident_write() 839 if (likely(lcn >= 0)) { in ntfs_prepare_pages_for_non_resident_write() 847 lcn_block = lcn << (vol->cluster_size_bits - in ntfs_prepare_pages_for_non_resident_write() 868 lcn = LCN_RL_NOT_MAPPED; in ntfs_prepare_pages_for_non_resident_write() 873 if (unlikely(lcn ! in ntfs_prepare_pages_for_non_resident_write() 1740 LCN lcn; ntfs_perform_write() local [all...] |
H A D | logfile.c | 762 LCN lcn; in ntfs_empty_logfile() local 770 lcn = rl->lcn; in ntfs_empty_logfile() 771 if (unlikely(lcn == LCN_RL_NOT_MAPPED)) { in ntfs_empty_logfile() 776 if (unlikely(!rl->length || lcn < LCN_HOLE)) in ntfs_empty_logfile() 779 if (lcn == LCN_HOLE) in ntfs_empty_logfile() 781 block = lcn << vol->cluster_size_bits >> block_size_bits; in ntfs_empty_logfile() 785 end_block = (lcn + len) << vol->cluster_size_bits >> in ntfs_empty_logfile()
|
H A D | mft.c | 526 LCN lcn; in ntfs_sync_mft_mirror() local 548 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_sync_mft_mirror() 549 /* For $MFTMirr, only lcn >= 0 is a successful remap. */ in ntfs_sync_mft_mirror() 550 if (likely(lcn >= 0)) { in ntfs_sync_mft_mirror() 552 bh->b_blocknr = ((lcn << in ntfs_sync_mft_mirror() 563 (long long)lcn); in ntfs_sync_mft_mirror() 719 LCN lcn; in write_mft_record_nolock() local 736 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in write_mft_record_nolock() 737 /* For $MFT, only lcn >= 0 is a successful remap. */ in write_mft_record_nolock() 738 if (likely(lcn > in write_mft_record_nolock() 1275 LCN lcn; ntfs_mft_bitmap_extend_allocation_nolock() local 1711 LCN lcn; ntfs_mft_data_extend_allocation_nolock() local [all...] |
H A D | compress.c | 479 LCN lcn; in ntfs_read_compressed_block() local 609 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_compressed_block() 611 lcn = LCN_RL_NOT_MAPPED; in ntfs_read_compressed_block() 612 ntfs_debug("Reading vcn = 0x%llx, lcn = 0x%llx.", in ntfs_read_compressed_block() 614 (unsigned long long)lcn); in ntfs_read_compressed_block() 615 if (lcn < 0) { in ntfs_read_compressed_block() 620 if (lcn == LCN_HOLE) in ntfs_read_compressed_block() 622 if (is_retry || lcn != LCN_RL_NOT_MAPPED) in ntfs_read_compressed_block() 634 block = lcn << vol->cluster_size_bits >> block_size_bits; in ntfs_read_compressed_block() 635 /* Read the lcn fro in ntfs_read_compressed_block() [all...] |
/kernel/linux/linux-6.6/fs/ntfs/ |
H A D | runlist.c | 148 if ((dst->lcn == LCN_RL_NOT_MAPPED) && (src->lcn == LCN_RL_NOT_MAPPED)) in ntfs_are_rl_mergeable() 154 if ((dst->lcn >= 0) && (src->lcn >= 0) && in ntfs_are_rl_mergeable() 155 ((dst->lcn + dst->length) == src->lcn)) in ntfs_are_rl_mergeable() 158 if ((dst->lcn == LCN_HOLE) && (src->lcn == LCN_HOLE)) in ntfs_are_rl_mergeable() 241 if (dst[marker].lcn == LCN_ENOENT) in ntfs_rl_append() 327 if (dst[marker].lcn in ntfs_rl_insert() 739 LCN lcn; /* Current lcn. */ ntfs_mapping_pairs_decompress() local [all...] |
H A D | lcnalloc.c | 33 * Locking: - The volume lcn bitmap must be locked for writing on entry and is 48 if (rl->lcn < 0) in ntfs_cluster_free_from_rl_nolock() 50 err = ntfs_bitmap_clear_run(lcnbmp_vi, rl->lcn, rl->length); in ntfs_cluster_free_from_rl_nolock() 63 * @start_lcn: starting lcn at which to allocate the clusters (or -1 if none) 100 * access to the lcn bitmap in portions of up to 8192 bytes at a time, so we 127 * Locking: - The volume lcn bitmap must be unlocked on entry and is unlocked 129 * - This function takes the volume lcn bitmap lock for writing and 137 LCN zone_start, zone_end, bmp_pos, bmp_initial_pos, last_read_pos, lcn; in ntfs_cluster_alloc() local 284 lcn = bmp_pos & 7; in ntfs_cluster_alloc() 286 ntfs_debug("Before inner while loop: buf_size %i, lcn in ntfs_cluster_alloc() [all...] |
H A D | aops.c | 168 LCN lcn; in ntfs_read_block() local 255 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_block() 257 lcn = LCN_RL_NOT_MAPPED; in ntfs_read_block() 259 if (lcn >= 0) { in ntfs_read_block() 261 bh->b_blocknr = ((lcn << vol->cluster_size_bits) in ntfs_read_block() 273 if (lcn == LCN_HOLE) in ntfs_read_block() 276 if (!is_retry && lcn == LCN_RL_NOT_MAPPED) { in ntfs_read_block() 294 if (err == -ENOENT || lcn == LCN_ENOENT) { in ntfs_read_block() 534 LCN lcn; in ntfs_write_block() local 713 lcn in ntfs_write_block() 1005 LCN lcn; ntfs_write_mst_block() local 1553 LCN lcn; ntfs_bmap() local [all...] |
H A D | debug.c | 135 LCN lcn = (rl + i)->lcn; in ntfs_debug_dump_runlist() local 137 if (lcn < (LCN)0) { in ntfs_debug_dump_runlist() 138 int index = -lcn - 1; in ntfs_debug_dump_runlist() 150 (long long)(rl + i)->lcn, in ntfs_debug_dump_runlist()
|
H A D | attrib.c | 298 * ntfs_attr_vcn_to_lcn_nolock - convert a vcn into a lcn given an ntfs inode 305 * number (lcn). 308 * extent containing the @vcn and the vcn to lcn conversion is retried. 330 LCN lcn; in ntfs_attr_vcn_to_lcn_nolock() local 349 /* Convert vcn to lcn. If that fails map the runlist and retry once. */ in ntfs_attr_vcn_to_lcn_nolock() 350 lcn = ntfs_rl_vcn_to_lcn(ni->runlist.rl, vcn); in ntfs_attr_vcn_to_lcn_nolock() 351 if (likely(lcn >= LCN_HOLE)) { in ntfs_attr_vcn_to_lcn_nolock() 352 ntfs_debug("Done, lcn 0x%llx.", (long long)lcn); in ntfs_attr_vcn_to_lcn_nolock() 353 return lcn; in ntfs_attr_vcn_to_lcn_nolock() 723 LCN lcn; load_attribute_list() local [all...] |
H A D | file.c | 567 LCN lcn; in ntfs_prepare_pages_for_non_resident_write() local 621 vcn = lcn = -1; in ntfs_prepare_pages_for_non_resident_write() 696 * clusters, @lcn is the device cluster corresponding to @vcn, in ntfs_prepare_pages_for_non_resident_write() 697 * and @lcn_block is the block number corresponding to @lcn. in ntfs_prepare_pages_for_non_resident_write() 702 BUG_ON(lcn < 0); in ntfs_prepare_pages_for_non_resident_write() 831 lcn = ntfs_rl_vcn_to_lcn(rl, bh_cpos); in ntfs_prepare_pages_for_non_resident_write() 832 if (likely(lcn >= 0)) { in ntfs_prepare_pages_for_non_resident_write() 840 lcn_block = lcn << (vol->cluster_size_bits - in ntfs_prepare_pages_for_non_resident_write() 861 lcn = LCN_RL_NOT_MAPPED; in ntfs_prepare_pages_for_non_resident_write() 866 if (unlikely(lcn ! in ntfs_prepare_pages_for_non_resident_write() 1730 LCN lcn; ntfs_perform_write() local [all...] |
H A D | logfile.c | 761 LCN lcn; in ntfs_empty_logfile() local 769 lcn = rl->lcn; in ntfs_empty_logfile() 770 if (unlikely(lcn == LCN_RL_NOT_MAPPED)) { in ntfs_empty_logfile() 775 if (unlikely(!rl->length || lcn < LCN_HOLE)) in ntfs_empty_logfile() 778 if (lcn == LCN_HOLE) in ntfs_empty_logfile() 780 block = lcn << vol->cluster_size_bits >> block_size_bits; in ntfs_empty_logfile() 784 end_block = (lcn + len) << vol->cluster_size_bits >> in ntfs_empty_logfile()
|
H A D | mft.c | 526 LCN lcn; in ntfs_sync_mft_mirror() local 548 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_sync_mft_mirror() 549 /* For $MFTMirr, only lcn >= 0 is a successful remap. */ in ntfs_sync_mft_mirror() 550 if (likely(lcn >= 0)) { in ntfs_sync_mft_mirror() 552 bh->b_blocknr = ((lcn << in ntfs_sync_mft_mirror() 563 (long long)lcn); in ntfs_sync_mft_mirror() 719 LCN lcn; in write_mft_record_nolock() local 736 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in write_mft_record_nolock() 737 /* For $MFT, only lcn >= 0 is a successful remap. */ in write_mft_record_nolock() 738 if (likely(lcn > in write_mft_record_nolock() 1275 LCN lcn; ntfs_mft_bitmap_extend_allocation_nolock() local 1711 LCN lcn; ntfs_mft_data_extend_allocation_nolock() local [all...] |
H A D | compress.c | 479 LCN lcn; in ntfs_read_compressed_block() local 609 lcn = ntfs_rl_vcn_to_lcn(rl, vcn); in ntfs_read_compressed_block() 611 lcn = LCN_RL_NOT_MAPPED; in ntfs_read_compressed_block() 612 ntfs_debug("Reading vcn = 0x%llx, lcn = 0x%llx.", in ntfs_read_compressed_block() 614 (unsigned long long)lcn); in ntfs_read_compressed_block() 615 if (lcn < 0) { in ntfs_read_compressed_block() 620 if (lcn == LCN_HOLE) in ntfs_read_compressed_block() 622 if (is_retry || lcn != LCN_RL_NOT_MAPPED) in ntfs_read_compressed_block() 634 block = lcn << vol->cluster_size_bits >> block_size_bits; in ntfs_read_compressed_block() 635 /* Read the lcn fro in ntfs_read_compressed_block() [all...] |
/kernel/linux/linux-5.10/fs/erofs/ |
H A D | zmap.c | 109 unsigned long lcn; member 150 unsigned long lcn) in legacy_load_cluster_from_disk() 158 lcn * sizeof(struct z_erofs_vle_decompressed_index); in legacy_load_cluster_from_disk() 167 m->lcn = lcn; in legacy_load_cluster_from_disk() 272 unsigned long lcn) in compacted_load_cluster_from_disk() 285 if (lcn >= totalidx) in compacted_load_cluster_from_disk() 288 m->lcn = lcn; in compacted_load_cluster_from_disk() 300 if (lcn < compacted_4b_initia in compacted_load_cluster_from_disk() 149 legacy_load_cluster_from_disk(struct z_erofs_maprecorder *m, unsigned long lcn) legacy_load_cluster_from_disk() argument 271 compacted_load_cluster_from_disk(struct z_erofs_maprecorder *m, unsigned long lcn) compacted_load_cluster_from_disk() argument 322 z_erofs_load_cluster_from_disk(struct z_erofs_maprecorder *m, unsigned int lcn) z_erofs_load_cluster_from_disk() argument 342 unsigned long lcn = m->lcn; z_erofs_extent_lookback() local [all...] |
/kernel/linux/linux-6.6/fs/erofs/ |
H A D | zmap.c | 15 unsigned long lcn; member 26 unsigned long lcn) in z_erofs_load_full_lcluster() 32 lcn * sizeof(struct z_erofs_lcluster_index); in z_erofs_load_full_lcluster() 42 m->lcn = lcn; in z_erofs_load_full_lcluster() 226 unsigned long lcn, bool lookahead) in z_erofs_load_compact_lcluster() 237 if (lcn >= totalidx) in z_erofs_load_compact_lcluster() 240 m->lcn = lcn; in z_erofs_load_compact_lcluster() 253 if (lcn < compacted_4b_initia in z_erofs_load_compact_lcluster() 25 z_erofs_load_full_lcluster(struct z_erofs_maprecorder *m, unsigned long lcn) z_erofs_load_full_lcluster() argument 225 z_erofs_load_compact_lcluster(struct z_erofs_maprecorder *m, unsigned long lcn, bool lookahead) z_erofs_load_compact_lcluster() argument 276 z_erofs_load_lcluster_from_disk(struct z_erofs_maprecorder *m, unsigned int lcn, bool lookahead) z_erofs_load_lcluster_from_disk() argument 297 unsigned long lcn = m->lcn - lookback_distance; z_erofs_extent_lookback() local 337 unsigned long lcn; z_erofs_get_extent_compressedlen() local 409 u64 lcn = m->lcn, headlcn = map->m_la >> lclusterbits; z_erofs_get_extent_decompressedlen() local [all...] |