Lines Matching refs:sbi

32 void f2fs_update_sit_info(struct f2fs_sb_info *sbi)
34 struct f2fs_stat_info *si = F2FS_STAT(sbi);
42 blks_per_sec = BLKS_PER_SEC(sbi);
44 for (segno = 0; segno < MAIN_SEGS(sbi); segno += sbi->segs_per_sec) {
45 vblocks = get_valid_blocks(sbi, segno, true);
54 dist = div_u64(MAIN_SECS(sbi) * hblks_per_sec * hblks_per_sec, 100);
63 static void update_general_status(struct f2fs_sb_info *sbi)
65 struct f2fs_stat_info *si = F2FS_STAT(sbi);
66 struct f2fs_super_block *raw_super = F2FS_RAW_SUPER(sbi);
76 si->hit_largest = atomic64_read(&sbi->read_hit_largest);
77 si->hit_cached = atomic64_read(&sbi->read_hit_cached);
78 si->hit_rbtree = atomic64_read(&sbi->read_hit_rbtree);
80 si->total_ext = atomic64_read(&sbi->total_hit_ext);
81 si->ext_tree = atomic_read(&sbi->total_ext_tree);
82 si->zombie_tree = atomic_read(&sbi->total_zombie_tree);
83 si->ext_node = atomic_read(&sbi->total_ext_node);
84 si->ndirty_node = get_pages(sbi, F2FS_DIRTY_NODES);
85 si->ndirty_dent = get_pages(sbi, F2FS_DIRTY_DENTS);
86 si->ndirty_meta = get_pages(sbi, F2FS_DIRTY_META);
87 si->ndirty_data = get_pages(sbi, F2FS_DIRTY_DATA);
88 si->ndirty_qdata = get_pages(sbi, F2FS_DIRTY_QDATA);
89 si->ndirty_imeta = get_pages(sbi, F2FS_DIRTY_IMETA);
90 si->ndirty_dirs = sbi->ndirty_inode[DIR_INODE];
91 si->ndirty_files = sbi->ndirty_inode[FILE_INODE];
92 si->nquota_files = sbi->nquota_files;
93 si->ndirty_all = sbi->ndirty_inode[DIRTY_META];
94 si->inmem_pages = get_pages(sbi, F2FS_INMEM_PAGES);
95 si->aw_cnt = sbi->atomic_files;
96 si->vw_cnt = atomic_read(&sbi->vw_cnt);
97 si->max_aw_cnt = atomic_read(&sbi->max_aw_cnt);
98 si->max_vw_cnt = atomic_read(&sbi->max_vw_cnt);
99 si->nr_dio_read = get_pages(sbi, F2FS_DIO_READ);
100 si->nr_dio_write = get_pages(sbi, F2FS_DIO_WRITE);
101 si->nr_wb_cp_data = get_pages(sbi, F2FS_WB_CP_DATA);
102 si->nr_wb_data = get_pages(sbi, F2FS_WB_DATA);
103 si->nr_rd_data = get_pages(sbi, F2FS_RD_DATA);
104 si->nr_rd_node = get_pages(sbi, F2FS_RD_NODE);
105 si->nr_rd_meta = get_pages(sbi, F2FS_RD_META);
106 if (SM_I(sbi)->fcc_info) {
108 atomic_read(&SM_I(sbi)->fcc_info->issued_flush);
110 atomic_read(&SM_I(sbi)->fcc_info->queued_flush);
112 llist_empty(&SM_I(sbi)->fcc_info->issue_list);
114 if (SM_I(sbi)->dcc_info) {
116 atomic_read(&SM_I(sbi)->dcc_info->issued_discard);
118 atomic_read(&SM_I(sbi)->dcc_info->queued_discard);
120 atomic_read(&SM_I(sbi)->dcc_info->discard_cmd_cnt);
121 si->undiscard_blks = SM_I(sbi)->dcc_info->undiscard_blks;
123 si->total_count = (int)sbi->user_block_count / sbi->blocks_per_seg;
124 si->rsvd_segs = reserved_segments(sbi);
125 si->overp_segs = overprovision_segments(sbi);
126 si->valid_count = valid_user_blocks(sbi);
127 si->discard_blks = discard_blocks(sbi);
128 si->valid_node_count = valid_node_count(sbi);
129 si->valid_inode_count = valid_inode_count(sbi);
130 si->inline_xattr = atomic_read(&sbi->inline_xattr);
131 si->inline_inode = atomic_read(&sbi->inline_inode);
132 si->inline_dir = atomic_read(&sbi->inline_dir);
133 si->compr_inode = atomic_read(&sbi->compr_inode);
134 si->compr_blocks = atomic64_read(&sbi->compr_blocks);
135 si->append = sbi->im[APPEND_INO].ino_num;
136 si->update = sbi->im[UPDATE_INO].ino_num;
137 si->orphans = sbi->im[ORPHAN_INO].ino_num;
138 si->utilization = utilization(sbi);
140 si->free_segs = free_segments(sbi);
141 si->free_secs = free_sections(sbi);
142 si->prefree_count = prefree_segments(sbi);
143 si->dirty_count = dirty_segments(sbi);
144 if (sbi->node_inode)
145 si->node_pages = NODE_MAPPING(sbi)->nrpages;
146 if (sbi->meta_inode)
147 si->meta_pages = META_MAPPING(sbi)->nrpages;
148 si->nats = NM_I(sbi)->nat_cnt[TOTAL_NAT];
149 si->dirty_nats = NM_I(sbi)->nat_cnt[DIRTY_NAT];
150 si->sits = MAIN_SEGS(sbi);
151 si->dirty_sits = SIT_I(sbi)->dirty_sentries;
152 si->free_nids = NM_I(sbi)->nid_cnt[FREE_NID];
153 si->avail_nids = NM_I(sbi)->available_nids;
154 si->alloc_nids = NM_I(sbi)->nid_cnt[PREALLOC_NID];
155 si->io_skip_bggc = sbi->io_skip_bggc;
156 si->other_skip_bggc = sbi->other_skip_bggc;
157 si->skipped_atomic_files[BG_GC] = sbi->skipped_atomic_files[BG_GC];
158 si->skipped_atomic_files[FG_GC] = sbi->skipped_atomic_files[FG_GC];
159 si->util_free = (int)(free_user_blocks(sbi) >> sbi->log_blocks_per_seg)
160 * 100 / (int)(sbi->user_block_count >> sbi->log_blocks_per_seg)
162 si->util_valid = (int)(written_block_count(sbi) >>
163 sbi->log_blocks_per_seg)
164 * 100 / (int)(sbi->user_block_count >> sbi->log_blocks_per_seg)
168 struct curseg_info *curseg = CURSEG_I(sbi, i);
170 si->cursec[i] = GET_SEC_FROM_SEG(sbi, curseg->segno);
171 si->curzone[i] = GET_ZONE_FROM_SEC(sbi, si->cursec[i]);
175 si->meta_count[i] = atomic_read(&sbi->meta_count[i]);
183 for (i = 0; i < MAIN_SEGS(sbi); i++) {
184 int blks = get_seg_entry(sbi, i)->valid_blocks;
185 int type = get_seg_entry(sbi, i)->type;
190 if (blks == sbi->blocks_per_seg)
198 si->segment_count[i] = sbi->segment_count[i];
199 si->block_count[i] = sbi->block_count[i];
202 si->inplace_count = atomic_read(&sbi->inplace_count);
208 static void update_mem_info(struct f2fs_sb_info *sbi)
210 struct f2fs_stat_info *si = F2FS_STAT(sbi);
220 si->base_mem += sizeof(struct f2fs_sb_info) + sbi->sb->s_blocksize;
222 si->base_mem += sizeof(*sbi->ckpt);
229 si->base_mem += MAIN_SEGS(sbi) * sizeof(struct seg_entry);
230 si->base_mem += f2fs_bitmap_size(MAIN_SEGS(sbi));
231 si->base_mem += 2 * SIT_VBLOCK_MAP_SIZE * MAIN_SEGS(sbi);
232 si->base_mem += SIT_VBLOCK_MAP_SIZE * MAIN_SEGS(sbi);
234 if (__is_large_section(sbi))
235 si->base_mem += MAIN_SECS(sbi) * sizeof(struct sec_entry);
236 si->base_mem += __bitmap_size(sbi, SIT_BITMAP);
240 si->base_mem += f2fs_bitmap_size(MAIN_SEGS(sbi));
241 si->base_mem += f2fs_bitmap_size(MAIN_SECS(sbi));
249 si->base_mem += NR_DIRTY_TYPE * f2fs_bitmap_size(MAIN_SEGS(sbi));
250 si->base_mem += f2fs_bitmap_size(MAIN_SECS(sbi));
254 si->base_mem += __bitmap_size(sbi, NAT_BITMAP);
255 si->base_mem += (NM_I(sbi)->nat_bits_blocks << F2FS_BLKSIZE_BITS);
256 si->base_mem += NM_I(sbi)->nat_blocks *
258 si->base_mem += NM_I(sbi)->nat_blocks / 8;
259 si->base_mem += NM_I(sbi)->nat_blocks * sizeof(unsigned short);
265 if (sbi->gc_thread)
269 if (SM_I(sbi)->fcc_info)
271 if (SM_I(sbi)->dcc_info) {
274 atomic_read(&SM_I(sbi)->dcc_info->discard_cmd_cnt);
278 si->cache_mem += (NM_I(sbi)->nid_cnt[FREE_NID] +
279 NM_I(sbi)->nid_cnt[PREALLOC_NID]) *
281 si->cache_mem += NM_I(sbi)->nat_cnt[TOTAL_NAT] *
283 si->cache_mem += NM_I(sbi)->nat_cnt[DIRTY_NAT] *
287 si->cache_mem += sbi->im[i].ino_num * sizeof(struct ino_entry);
288 si->cache_mem += atomic_read(&sbi->total_ext_tree) *
290 si->cache_mem += atomic_read(&sbi->total_ext_node) *
294 if (sbi->node_inode) {
295 unsigned npages = NODE_MAPPING(sbi)->nrpages;
298 if (sbi->meta_inode) {
299 unsigned npages = META_MAPPING(sbi)->nrpages;
312 update_general_status(si->sbi);
315 si->sbi->sb->s_bdev, i++,
316 f2fs_readonly(si->sbi->sb) ? "RO": "RW",
317 is_set_ckpt_flags(si->sbi, CP_DISABLED_FLAG) ?
318 "Disabled": (f2fs_cp_error(si->sbi) ? "Error": "Good"));
327 SIT_I(si->sbi)->mounted_time);
328 if (test_opt(si->sbi, DISCARD))
502 f2fs_update_sit_info(si->sbi);
507 update_mem_info(si->sbi);
524 int f2fs_build_stats(struct f2fs_sb_info *sbi)
526 struct f2fs_super_block *raw_super = F2FS_RAW_SUPER(sbi);
530 si = f2fs_kzalloc(sbi, sizeof(struct f2fs_stat_info), GFP_KERNEL);
542 si->sbi = sbi;
543 sbi->stat_info = si;
545 atomic64_set(&sbi->total_hit_ext, 0);
546 atomic64_set(&sbi->read_hit_rbtree, 0);
547 atomic64_set(&sbi->read_hit_largest, 0);
548 atomic64_set(&sbi->read_hit_cached, 0);
550 atomic_set(&sbi->inline_xattr, 0);
551 atomic_set(&sbi->inline_inode, 0);
552 atomic_set(&sbi->inline_dir, 0);
553 atomic_set(&sbi->compr_inode, 0);
554 atomic64_set(&sbi->compr_blocks, 0);
555 atomic_set(&sbi->inplace_count, 0);
557 atomic_set(&sbi->meta_count[i], 0);
559 atomic_set(&sbi->vw_cnt, 0);
560 atomic_set(&sbi->max_aw_cnt, 0);
561 atomic_set(&sbi->max_vw_cnt, 0);
570 void f2fs_destroy_stats(struct f2fs_sb_info *sbi)
572 struct f2fs_stat_info *si = F2FS_STAT(sbi);