Home
last modified time | relevance | path

Searched refs:m_sb_lock (Results 1 - 25 of 27) sorted by relevance

12

/kernel/linux/linux-5.10/fs/xfs/
H A Dxfs_health.c101 spin_lock(&mp->m_sb_lock); in xfs_fs_mark_sick()
104 spin_unlock(&mp->m_sb_lock); in xfs_fs_mark_sick()
116 spin_lock(&mp->m_sb_lock); in xfs_fs_mark_healthy()
119 spin_unlock(&mp->m_sb_lock); in xfs_fs_mark_healthy()
129 spin_lock(&mp->m_sb_lock); in xfs_fs_measure_sickness()
132 spin_unlock(&mp->m_sb_lock); in xfs_fs_measure_sickness()
144 spin_lock(&mp->m_sb_lock); in xfs_rt_mark_sick()
147 spin_unlock(&mp->m_sb_lock); in xfs_rt_mark_sick()
159 spin_lock(&mp->m_sb_lock); in xfs_rt_mark_healthy()
162 spin_unlock(&mp->m_sb_lock); in xfs_rt_mark_healthy()
[all...]
H A Dxfs_fsops.c298 spin_lock(&mp->m_sb_lock); in xfs_fs_counts()
300 spin_unlock(&mp->m_sb_lock); in xfs_fs_counts()
346 * We do this under the m_sb_lock so that if we are near ENOSPC, we will in xfs_reserve_blocks()
351 spin_lock(&mp->m_sb_lock); in xfs_reserve_blocks()
367 spin_unlock(&mp->m_sb_lock); in xfs_reserve_blocks()
369 spin_lock(&mp->m_sb_lock); in xfs_reserve_blocks()
383 * m_sb_lock. Set the reserve size even if there's not enough free in xfs_reserve_blocks()
403 spin_unlock(&mp->m_sb_lock); in xfs_reserve_blocks()
407 spin_lock(&mp->m_sb_lock); in xfs_reserve_blocks()
415 spin_unlock(&mp->m_sb_lock); in xfs_reserve_blocks()
[all...]
H A Dxfs_mount.c551 /* It is OK to look at sb_qflags in the mount path without m_sb_lock. */ in xfs_mount_reset_sbqflags()
554 spin_lock(&mp->m_sb_lock); in xfs_mount_reset_sbqflags()
556 spin_unlock(&mp->m_sb_lock); in xfs_mount_reset_sbqflags()
1225 spin_lock(&mp->m_sb_lock); in xfs_mod_fdblocks()
1235 spin_unlock(&mp->m_sb_lock); in xfs_mod_fdblocks()
1264 spin_lock(&mp->m_sb_lock); in xfs_mod_fdblocks()
1272 spin_unlock(&mp->m_sb_lock); in xfs_mod_fdblocks()
1279 spin_unlock(&mp->m_sb_lock); in xfs_mod_fdblocks()
1291 spin_lock(&mp->m_sb_lock); in xfs_mod_frextents()
1297 spin_unlock(&mp->m_sb_lock); in xfs_mod_frextents()
[all...]
H A Dxfs_qm_syscalls.c39 spin_lock(&mp->m_sb_lock); in xfs_qm_log_quotaoff()
41 spin_unlock(&mp->m_sb_lock); in xfs_qm_log_quotaoff()
133 spin_lock(&mp->m_sb_lock); in xfs_qm_scall_quotaoff()
135 spin_unlock(&mp->m_sb_lock); in xfs_qm_scall_quotaoff()
398 spin_lock(&mp->m_sb_lock); in xfs_qm_scall_quotaon()
401 spin_unlock(&mp->m_sb_lock); in xfs_qm_scall_quotaon()
H A Dxfs_trans.c617 spin_lock(&mp->m_sb_lock); in xfs_trans_unreserve_and_mod_sb()
630 spin_unlock(&mp->m_sb_lock); in xfs_trans_unreserve_and_mod_sb()
H A Dxfs_mount.h145 * Callers must hold m_sb_lock to access these two fields.
151 * Callers must hold m_sb_lock to access this field.
163 spinlock_t ____cacheline_aligned m_sb_lock; /* sb counter lock */ member
H A Dxfs_ioctl.c1967 spin_lock(&mp->m_sb_lock); in xfs_ioc_getlabel()
1969 spin_unlock(&mp->m_sb_lock); in xfs_ioc_getlabel()
2005 spin_lock(&mp->m_sb_lock); in xfs_ioc_setlabel()
2008 spin_unlock(&mp->m_sb_lock); in xfs_ioc_setlabel()
H A Dxfs_qm.c802 spin_lock(&mp->m_sb_lock); in xfs_qm_qino_alloc()
820 spin_unlock(&mp->m_sb_lock); in xfs_qm_qino_alloc()
1462 * We actually don't have to acquire the m_sb_lock at all. in xfs_qm_mount_quotas()
1465 spin_lock(&mp->m_sb_lock); in xfs_qm_mount_quotas()
1468 spin_unlock(&mp->m_sb_lock); in xfs_qm_mount_quotas()
H A Dxfs_super.c809 spin_lock(&mp->m_sb_lock); in xfs_fs_statfs()
813 spin_unlock(&mp->m_sb_lock); in xfs_fs_statfs()
1860 spin_lock_init(&mp->m_sb_lock); in xfs_init_fs_context()
/kernel/linux/linux-6.6/fs/xfs/
H A Dxfs_health.c99 spin_lock(&mp->m_sb_lock); in xfs_fs_mark_sick()
102 spin_unlock(&mp->m_sb_lock); in xfs_fs_mark_sick()
114 spin_lock(&mp->m_sb_lock); in xfs_fs_mark_healthy()
117 spin_unlock(&mp->m_sb_lock); in xfs_fs_mark_healthy()
127 spin_lock(&mp->m_sb_lock); in xfs_fs_measure_sickness()
130 spin_unlock(&mp->m_sb_lock); in xfs_fs_measure_sickness()
142 spin_lock(&mp->m_sb_lock); in xfs_rt_mark_sick()
145 spin_unlock(&mp->m_sb_lock); in xfs_rt_mark_sick()
157 spin_lock(&mp->m_sb_lock); in xfs_rt_mark_healthy()
160 spin_unlock(&mp->m_sb_lock); in xfs_rt_mark_healthy()
[all...]
H A Dxfs_fsops.c405 * We do this under the m_sb_lock so that if we are near ENOSPC, we will in xfs_reserve_blocks()
410 spin_lock(&mp->m_sb_lock); in xfs_reserve_blocks()
426 spin_unlock(&mp->m_sb_lock); in xfs_reserve_blocks()
428 spin_lock(&mp->m_sb_lock); in xfs_reserve_blocks()
442 * m_sb_lock. Set the reserve size even if there's not enough free in xfs_reserve_blocks()
462 spin_unlock(&mp->m_sb_lock); in xfs_reserve_blocks()
466 spin_lock(&mp->m_sb_lock); in xfs_reserve_blocks()
474 spin_unlock(&mp->m_sb_lock); in xfs_reserve_blocks()
H A Dxfs_mount.c438 /* It is OK to look at sb_qflags in the mount path without m_sb_lock. */ in xfs_mount_reset_sbqflags()
441 spin_lock(&mp->m_sb_lock); in xfs_mount_reset_sbqflags()
443 spin_unlock(&mp->m_sb_lock); in xfs_mount_reset_sbqflags()
1166 spin_lock(&mp->m_sb_lock); in xfs_mod_freecounter()
1176 spin_unlock(&mp->m_sb_lock); in xfs_mod_freecounter()
1218 spin_lock(&mp->m_sb_lock); in xfs_mod_freecounter()
1226 spin_unlock(&mp->m_sb_lock); in xfs_mod_freecounter()
1233 spin_unlock(&mp->m_sb_lock); in xfs_mod_freecounter()
H A Dxfs_qm_syscalls.c44 spin_lock(&mp->m_sb_lock); in xfs_qm_scall_quotaoff()
46 spin_unlock(&mp->m_sb_lock); in xfs_qm_scall_quotaoff()
182 spin_lock(&mp->m_sb_lock); in xfs_qm_scall_quotaon()
185 spin_unlock(&mp->m_sb_lock); in xfs_qm_scall_quotaon()
H A Dxfs_trans.c518 spin_lock(&mp->m_sb_lock); in xfs_trans_apply_sb_deltas()
521 spin_unlock(&mp->m_sb_lock); in xfs_trans_apply_sb_deltas()
644 spin_lock(&mp->m_sb_lock); in xfs_trans_unreserve_and_mod_sb()
662 spin_unlock(&mp->m_sb_lock); in xfs_trans_unreserve_and_mod_sb()
H A Dxfs_qm.c803 spin_lock(&mp->m_sb_lock); in xfs_qm_qino_alloc()
821 spin_unlock(&mp->m_sb_lock); in xfs_qm_qino_alloc()
1493 * We actually don't have to acquire the m_sb_lock at all. in xfs_qm_mount_quotas()
1496 spin_lock(&mp->m_sb_lock); in xfs_qm_mount_quotas()
1499 spin_unlock(&mp->m_sb_lock); in xfs_qm_mount_quotas()
H A Dxfs_ioctl.c1759 spin_lock(&mp->m_sb_lock); in xfs_ioc_getlabel()
1761 spin_unlock(&mp->m_sb_lock); in xfs_ioc_getlabel()
1797 spin_lock(&mp->m_sb_lock); in xfs_ioc_setlabel()
1800 spin_unlock(&mp->m_sb_lock); in xfs_ioc_setlabel()
H A Dxfs_mount.h166 * Callers must hold m_sb_lock to access these two fields.
172 * Callers must hold m_sb_lock to access this field.
184 spinlock_t ____cacheline_aligned m_sb_lock; /* sb counter lock */ member
H A Dxfs_rtalloc.c1342 spin_lock(&mp->m_sb_lock); in xfs_rtalloc_reinit_frextents()
1344 spin_unlock(&mp->m_sb_lock); in xfs_rtalloc_reinit_frextents()
H A Dxfs_super.c854 spin_lock(&mp->m_sb_lock); in xfs_fs_statfs()
858 spin_unlock(&mp->m_sb_lock); in xfs_fs_statfs()
2001 spin_lock_init(&mp->m_sb_lock); in xfs_init_fs_context()
/kernel/linux/linux-6.6/fs/xfs/scrub/
H A Drepair.c690 spin_lock(&sc->mp->m_sb_lock); in xrep_force_quotacheck()
692 spin_unlock(&sc->mp->m_sb_lock); in xrep_force_quotacheck()
/kernel/linux/linux-5.10/fs/xfs/scrub/
H A Drepair.c911 spin_lock(&sc->mp->m_sb_lock); in xrep_force_quotacheck()
913 spin_unlock(&sc->mp->m_sb_lock); in xrep_force_quotacheck()
/kernel/linux/linux-5.10/fs/xfs/libxfs/
H A Dxfs_sb.c934 spin_lock(&mp->m_sb_lock); in xfs_initialize_perag_data()
938 spin_unlock(&mp->m_sb_lock); in xfs_initialize_perag_data()
H A Dxfs_attr_leaf.c630 spin_lock(&mp->m_sb_lock); in xfs_sbversion_add_attr2()
633 spin_unlock(&mp->m_sb_lock); in xfs_sbversion_add_attr2()
636 spin_unlock(&mp->m_sb_lock); in xfs_sbversion_add_attr2()
/kernel/linux/linux-6.6/fs/xfs/libxfs/
H A Dxfs_ag.c225 spin_lock(&mp->m_sb_lock); in xfs_initialize_perag_data()
229 spin_unlock(&mp->m_sb_lock); in xfs_initialize_perag_data()
H A Dxfs_attr_leaf.c671 spin_lock(&mp->m_sb_lock); in xfs_sbversion_add_attr2()
673 spin_unlock(&mp->m_sb_lock); in xfs_sbversion_add_attr2()

Completed in 36 milliseconds

12