Lines Matching refs:mp

250 struct vnode *__mnt_vnode_next_all(struct vnode **mvp, struct mount *mp);
251 struct vnode *__mnt_vnode_first_all(struct vnode **mvp, struct mount *mp);
252 void __mnt_vnode_markerfree_all(struct vnode **mvp, struct mount *mp);
254 #define MNT_VNODE_FOREACH_ALL(vp, mp, mvp) \ argument
255 for (vp = __mnt_vnode_first_all(&(mvp), (mp)); \
256 (vp) != NULL; vp = __mnt_vnode_next_all(&(mvp), (mp)))
258 #define MNT_VNODE_FOREACH_ALL_ABORT(mp, mvp) \ argument
260 MNT_ILOCK(mp); \
261 __mnt_vnode_markerfree_all(&(mvp), (mp)); \
263 mtx_assert(MNT_MTX(mp), MA_NOTOWNED); \
270 struct vnode *__mnt_vnode_next_lazy(struct vnode **mvp, struct mount *mp,
272 struct vnode *__mnt_vnode_first_lazy(struct vnode **mvp, struct mount *mp,
274 void __mnt_vnode_markerfree_lazy(struct vnode **mvp, struct mount *mp);
276 #define MNT_VNODE_FOREACH_LAZY(vp, mp, mvp, cb, cbarg) \ argument
277 for (vp = __mnt_vnode_first_lazy(&(mvp), (mp), (cb), (cbarg)); \
279 vp = __mnt_vnode_next_lazy(&(mvp), (mp), (cb), (cbarg)))
281 #define MNT_VNODE_FOREACH_LAZY_ABORT(mp, mvp) \ argument
282 __mnt_vnode_markerfree_lazy(&(mvp), (mp))
284 #define MNT_ILOCK(mp) mtx_lock(&(mp)->mnt_mtx) argument
285 #define MNT_ITRYLOCK(mp) mtx_trylock(&(mp)->mnt_mtx) argument
286 #define MNT_IUNLOCK(mp) mtx_unlock(&(mp)->mnt_mtx) argument
287 #define MNT_MTX(mp) (&(mp)->mnt_mtx) argument
289 #define MNT_REF(mp) do { \ argument
290 mtx_assert(MNT_MTX(mp), MA_OWNED); \
291 mp->mnt_ref++; \
293 #define MNT_REL(mp) do { \ argument
294 mtx_assert(MNT_MTX(mp), MA_OWNED); \
295 (mp)->mnt_ref--; \
296 if ((mp)->mnt_vfs_ops && (mp)->mnt_ref < 0) \
297 vfs_dump_mount_counters(mp); \
298 if ((mp)->mnt_ref == 0 && (mp)->mnt_vfs_ops) \
299 wakeup((mp)); \
487 MNT_SHARED_WRITES(struct mount *mp) in MNT_SHARED_WRITES() argument
490 return (mp != NULL && (mp->mnt_kern_flag & MNTK_SHARED_WRITES) != 0); in MNT_SHARED_WRITES()
494 MNT_EXTENDED_SHARED(struct mount *mp) in MNT_EXTENDED_SHARED() argument
497 return (mp != NULL && (mp->mnt_kern_flag & MNTK_EXTENDED_SHARED) != 0); in MNT_EXTENDED_SHARED()
756 typedef int vfs_unmount_t(struct mount *mp, int mntflags);
757 typedef int vfs_root_t(struct mount *mp, int flags, struct vnode **vpp);
758 typedef int vfs_quotactl_t(struct mount *mp, int cmds, uid_t uid, void *arg);
759 typedef int vfs_statfs_t(struct mount *mp, struct statfs *sbp);
760 typedef int vfs_sync_t(struct mount *mp, int waitfor);
761 typedef int vfs_vget_t(struct mount *mp, ino_t ino, int flags,
763 typedef int vfs_fhtovp_t(struct mount *mp, struct fid *fhp,
765 typedef int vfs_checkexp_t(struct mount *mp, struct sockaddr *nam,
770 typedef int vfs_extattrctl_t(struct mount *mp, int cmd,
773 typedef int vfs_mount_t(struct mount *mp);
774 typedef int vfs_sysctl_t(struct mount *mp, fsctlop_t op,
776 typedef void vfs_susp_clean_t(struct mount *mp);
777 typedef void vfs_notify_lowervp_t(struct mount *mp, struct vnode *lowervp);
778 typedef void vfs_purge_t(struct mount *mp);
1052 #define vfs_assert_mount_counters(mp) do { } while (0) argument
1053 #define vfs_dump_mount_counters(mp) do { } while (0) argument
1074 #define vfs_mount_pcpu(mp) zpcpu_get(mp->mnt_pcpu) argument
1075 #define vfs_mount_pcpu_remote(mp, cpu) zpcpu_get_cpu(mp->mnt_pcpu, cpu) argument
1077 #define vfs_op_thread_entered(mp) ({ \ argument
1079 struct mount_pcpu *_mpcpu = vfs_mount_pcpu(mp); \
1083 #define vfs_op_thread_enter_crit(mp, _mpcpu) ({ \ argument
1086 _mpcpu = vfs_mount_pcpu(mp); \
1090 if (__predict_false(mp->mnt_vfs_ops > 0)) { \
1091 vfs_op_thread_exit_crit(mp, _mpcpu); \
1097 #define vfs_op_thread_enter(mp, _mpcpu) ({ \ argument
1100 _retval = vfs_op_thread_enter_crit(mp, _mpcpu); \
1106 #define vfs_op_thread_exit_crit(mp, _mpcpu) do { \ argument
1107 MPASS(_mpcpu == vfs_mount_pcpu(mp)); \
1113 #define vfs_op_thread_exit(mp, _mpcpu) do { \ argument
1114 vfs_op_thread_exit_crit(mp, _mpcpu); \