Lines Matching refs:cur
29 struct xfs_btree_cur *cur,
32 return M_IGEO(cur->bc_mp)->inobt_mnr[level != 0];
37 struct xfs_btree_cur *cur)
39 return xfs_inobt_init_cursor(cur->bc_ag.pag, cur->bc_tp,
40 cur->bc_ag.agbp, cur->bc_btnum);
45 struct xfs_btree_cur *cur,
49 struct xfs_buf *agbp = cur->bc_ag.agbp;
54 xfs_ialloc_log_agi(cur->bc_tp, agbp, XFS_AGI_ROOT | XFS_AGI_LEVEL);
59 struct xfs_btree_cur *cur,
63 struct xfs_buf *agbp = cur->bc_ag.agbp;
68 xfs_ialloc_log_agi(cur->bc_tp, agbp,
75 struct xfs_btree_cur *cur,
78 struct xfs_buf *agbp = cur->bc_ag.agbp;
81 if (!xfs_has_inobtcounts(cur->bc_mp))
84 if (cur->bc_btnum == XFS_BTNUM_FINO)
86 else if (cur->bc_btnum == XFS_BTNUM_INO)
88 xfs_ialloc_log_agi(cur->bc_tp, agbp, XFS_AGI_IBLOCKS);
93 struct xfs_btree_cur *cur,
104 args.tp = cur->bc_tp;
105 args.mp = cur->bc_mp;
106 args.pag = cur->bc_ag.pag;
126 xfs_inobt_mod_blockcount(cur, 1);
132 struct xfs_btree_cur *cur,
137 return __xfs_inobt_alloc_block(cur, start, new, stat, XFS_AG_RESV_NONE);
142 struct xfs_btree_cur *cur,
147 if (cur->bc_mp->m_finobt_nores)
148 return xfs_inobt_alloc_block(cur, start, new, stat);
149 return __xfs_inobt_alloc_block(cur, start, new, stat,
155 struct xfs_btree_cur *cur,
161 xfs_inobt_mod_blockcount(cur, -1);
162 fsbno = XFS_DADDR_TO_FSB(cur->bc_mp, xfs_buf_daddr(bp));
163 return xfs_free_extent_later(cur->bc_tp, fsbno, 1,
169 struct xfs_btree_cur *cur,
172 return __xfs_inobt_free_block(cur, bp, XFS_AG_RESV_NONE);
177 struct xfs_btree_cur *cur,
180 if (cur->bc_mp->m_finobt_nores)
181 return xfs_inobt_free_block(cur, bp);
182 return __xfs_inobt_free_block(cur, bp, XFS_AG_RESV_METADATA);
187 struct xfs_btree_cur *cur,
190 return M_IGEO(cur->bc_mp)->inobt_mxr[level != 0];
215 struct xfs_btree_cur *cur,
218 rec->inobt.ir_startino = cpu_to_be32(cur->bc_rec.i.ir_startino);
219 if (xfs_has_sparseinodes(cur->bc_mp)) {
221 cpu_to_be16(cur->bc_rec.i.ir_holemask);
222 rec->inobt.ir_u.sp.ir_count = cur->bc_rec.i.ir_count;
223 rec->inobt.ir_u.sp.ir_freecount = cur->bc_rec.i.ir_freecount;
227 cpu_to_be32(cur->bc_rec.i.ir_freecount);
229 rec->inobt.ir_free = cpu_to_be64(cur->bc_rec.i.ir_free);
237 struct xfs_btree_cur *cur,
240 struct xfs_agi *agi = cur->bc_ag.agbp->b_addr;
242 ASSERT(cur->bc_ag.pag->pag_agno == be32_to_cpu(agi->agi_seqno));
249 struct xfs_btree_cur *cur,
252 struct xfs_agi *agi = cur->bc_ag.agbp->b_addr;
254 ASSERT(cur->bc_ag.pag->pag_agno == be32_to_cpu(agi->agi_seqno));
260 struct xfs_btree_cur *cur,
264 cur->bc_rec.i.ir_startino;
269 struct xfs_btree_cur *cur,
370 struct xfs_btree_cur *cur,
380 struct xfs_btree_cur *cur,
390 struct xfs_btree_cur *cur,
455 struct xfs_btree_cur *cur;
457 cur = xfs_btree_alloc_cursor(mp, tp, btnum,
460 cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_ibt_2);
461 cur->bc_ops = &xfs_inobt_ops;
463 cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_fibt_2);
464 cur->bc_ops = &xfs_finobt_ops;
468 cur->bc_flags |= XFS_BTREE_CRC_BLOCKS;
470 cur->bc_ag.pag = xfs_perag_hold(pag);
471 return cur;
482 struct xfs_btree_cur *cur;
485 cur = xfs_inobt_init_common(pag, tp, btnum);
487 cur->bc_nlevels = be32_to_cpu(agi->agi_level);
489 cur->bc_nlevels = be32_to_cpu(agi->agi_free_level);
490 cur->bc_ag.agbp = agbp;
491 return cur;
501 struct xfs_btree_cur *cur;
503 cur = xfs_inobt_init_common(pag, NULL, btnum);
504 xfs_btree_stage_afakeroot(cur, afake);
505 return cur;
514 struct xfs_btree_cur *cur,
519 struct xbtree_afakeroot *afake = cur->bc_ag.afake;
522 ASSERT(cur->bc_flags & XFS_BTREE_STAGING);
524 if (cur->bc_btnum == XFS_BTNUM_INO) {
528 if (xfs_has_inobtcounts(cur->bc_mp)) {
533 xfs_btree_commit_afakeroot(cur, tp, agbp, &xfs_inobt_ops);
538 if (xfs_has_inobtcounts(cur->bc_mp)) {
543 xfs_btree_commit_afakeroot(cur, tp, agbp, &xfs_finobt_ops);
733 struct xfs_btree_cur *cur;
743 cur = xfs_inobt_init_cursor(pag, tp, *agi_bpp, which);
744 *curpp = cur;
756 struct xfs_btree_cur *cur = NULL;
759 error = xfs_inobt_cur(pag, tp, btnum, &cur, &agbp);
763 error = xfs_btree_count_blocks(cur, tree_blocks);
764 xfs_btree_del_cursor(cur, error);