Lines Matching refs:cur

27 	struct xfs_btree_cur	*cur)
29 return xfs_allocbt_init_cursor(cur->bc_mp, cur->bc_tp,
30 cur->bc_ag.agbp, cur->bc_ag.pag, cur->bc_btnum);
35 struct xfs_btree_cur *cur,
39 struct xfs_buf *agbp = cur->bc_ag.agbp;
41 int btnum = cur->bc_btnum;
47 cur->bc_ag.pag->pagf_levels[btnum] += inc;
49 xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_ROOTS | XFS_AGF_LEVELS);
54 struct xfs_btree_cur *cur,
63 error = xfs_alloc_get_freelist(cur->bc_ag.pag, cur->bc_tp,
64 cur->bc_ag.agbp, &bno, 1);
73 atomic64_inc(&cur->bc_mp->m_allocbt_blks);
74 xfs_extent_busy_reuse(cur->bc_mp, cur->bc_ag.pag, bno, 1, false);
84 struct xfs_btree_cur *cur,
87 struct xfs_buf *agbp = cur->bc_ag.agbp;
91 bno = xfs_daddr_to_agbno(cur->bc_mp, xfs_buf_daddr(bp));
92 error = xfs_alloc_put_freelist(cur->bc_ag.pag, cur->bc_tp, agbp, NULL,
97 atomic64_dec(&cur->bc_mp->m_allocbt_blks);
98 xfs_extent_busy_insert(cur->bc_tp, agbp->b_pag, bno, 1,
108 struct xfs_btree_cur *cur,
114 struct xfs_agf *agf = cur->bc_ag.agbp->b_addr;
119 ASSERT(cur->bc_btnum == XFS_BTNUM_CNT);
146 rrp = XFS_ALLOC_REC_ADDR(cur->bc_mp, block, numrecs);
159 pag = cur->bc_ag.agbp->b_pag;
161 xfs_alloc_log_agf(cur->bc_tp, cur->bc_ag.agbp, XFS_AGF_LONGEST);
166 struct xfs_btree_cur *cur,
169 return cur->bc_mp->m_alloc_mnr[level != 0];
174 struct xfs_btree_cur *cur,
177 return cur->bc_mp->m_alloc_mxr[level != 0];
213 struct xfs_btree_cur *cur,
216 rec->alloc.ar_startblock = cpu_to_be32(cur->bc_rec.a.ar_startblock);
217 rec->alloc.ar_blockcount = cpu_to_be32(cur->bc_rec.a.ar_blockcount);
222 struct xfs_btree_cur *cur,
225 struct xfs_agf *agf = cur->bc_ag.agbp->b_addr;
227 ASSERT(cur->bc_ag.pag->pag_agno == be32_to_cpu(agf->agf_seqno));
229 ptr->s = agf->agf_roots[cur->bc_btnum];
234 struct xfs_btree_cur *cur,
237 struct xfs_alloc_rec_incore *rec = &cur->bc_rec.a;
245 struct xfs_btree_cur *cur,
248 struct xfs_alloc_rec_incore *rec = &cur->bc_rec.a;
261 struct xfs_btree_cur *cur,
274 struct xfs_btree_cur *cur,
388 struct xfs_btree_cur *cur,
398 struct xfs_btree_cur *cur,
409 struct xfs_btree_cur *cur,
422 struct xfs_btree_cur *cur,
435 struct xfs_btree_cur *cur,
500 struct xfs_btree_cur *cur;
504 cur = xfs_btree_alloc_cursor(mp, tp, btnum, mp->m_alloc_maxlevels,
506 cur->bc_ag.abt.active = false;
509 cur->bc_ops = &xfs_cntbt_ops;
510 cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtc_2);
511 cur->bc_flags = XFS_BTREE_LASTREC_UPDATE;
513 cur->bc_ops = &xfs_bnobt_ops;
514 cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtb_2);
517 cur->bc_ag.pag = xfs_perag_hold(pag);
520 cur->bc_flags |= XFS_BTREE_CRC_BLOCKS;
522 return cur;
537 struct xfs_btree_cur *cur;
539 cur = xfs_allocbt_init_common(mp, tp, pag, btnum);
541 cur->bc_nlevels = be32_to_cpu(agf->agf_levels[XFS_BTNUM_CNT]);
543 cur->bc_nlevels = be32_to_cpu(agf->agf_levels[XFS_BTNUM_BNO]);
545 cur->bc_ag.agbp = agbp;
547 return cur;
558 struct xfs_btree_cur *cur;
560 cur = xfs_allocbt_init_common(mp, NULL, pag, btnum);
561 xfs_btree_stage_afakeroot(cur, afake);
562 return cur;
571 struct xfs_btree_cur *cur,
576 struct xbtree_afakeroot *afake = cur->bc_ag.afake;
578 ASSERT(cur->bc_flags & XFS_BTREE_STAGING);
580 agf->agf_roots[cur->bc_btnum] = cpu_to_be32(afake->af_root);
581 agf->agf_levels[cur->bc_btnum] = cpu_to_be32(afake->af_levels);
584 if (cur->bc_btnum == XFS_BTNUM_BNO) {
585 xfs_btree_commit_afakeroot(cur, tp, agbp, &xfs_bnobt_ops);
587 cur->bc_flags |= XFS_BTREE_LASTREC_UPDATE;
588 xfs_btree_commit_afakeroot(cur, tp, agbp, &xfs_cntbt_ops);