[XFS] Fix UP build breakage due to undefined m_icsb_mutex.
SGI-PV: 952227
SGI-Modid: xfs-linux-melb:xfs-kern:27692a
Signed-off-by: David Chinner <dgc@sgi.com>
Signed-off-by: Lachlan McIlroy <lachlan@sgi.com>
Signed-off-by: Tim Shimmin <tes@sgi.com>
diff --git a/fs/xfs/xfs_mount.c b/fs/xfs/xfs_mount.c
index 37c612c..cbe7293 100644
--- a/fs/xfs/xfs_mount.c
+++ b/fs/xfs/xfs_mount.c
@@ -543,11 +543,11 @@
ASSERT(XFS_BUF_VALUSEMA(bp) <= 0);
}
- mutex_lock(&mp->m_icsb_mutex);
+ xfs_icsb_lock(mp);
xfs_icsb_balance_counter(mp, XFS_SBS_ICOUNT, 0, 0);
xfs_icsb_balance_counter(mp, XFS_SBS_IFREE, 0, 0);
xfs_icsb_balance_counter(mp, XFS_SBS_FDBLOCKS, 0, 0);
- mutex_unlock(&mp->m_icsb_mutex);
+ xfs_icsb_unlock(mp);
mp->m_sb_bp = bp;
xfs_buf_relse(bp);
@@ -1736,17 +1736,17 @@
memset(cntp, 0, sizeof(xfs_icsb_cnts_t));
break;
case CPU_ONLINE:
- mutex_lock(&mp->m_icsb_mutex);
+ xfs_icsb_lock(mp);
xfs_icsb_balance_counter(mp, XFS_SBS_ICOUNT, 0, 0);
xfs_icsb_balance_counter(mp, XFS_SBS_IFREE, 0, 0);
xfs_icsb_balance_counter(mp, XFS_SBS_FDBLOCKS, 0, 0);
- mutex_unlock(&mp->m_icsb_mutex);
+ xfs_icsb_unlock(mp);
break;
case CPU_DEAD:
/* Disable all the counters, then fold the dead cpu's
* count into the total on the global superblock and
* re-enable the counters. */
- mutex_lock(&mp->m_icsb_mutex);
+ xfs_icsb_lock(mp);
s = XFS_SB_LOCK(mp);
xfs_icsb_disable_counter(mp, XFS_SBS_ICOUNT);
xfs_icsb_disable_counter(mp, XFS_SBS_IFREE);
@@ -1765,7 +1765,7 @@
xfs_icsb_balance_counter(mp, XFS_SBS_FDBLOCKS,
XFS_ICSB_SB_LOCKED, 0);
XFS_SB_UNLOCK(mp, s);
- mutex_unlock(&mp->m_icsb_mutex);
+ xfs_icsb_unlock(mp);
break;
}
@@ -1813,6 +1813,7 @@
unregister_hotcpu_notifier(&mp->m_icsb_notifier);
free_percpu(mp->m_sb_cnts);
}
+ mutex_destroy(&mp->m_icsb_mutex);
}
STATIC_INLINE void
@@ -2156,7 +2157,7 @@
* the superblock lock. We still need to hold the superblock
* lock, however, when we modify the global structures.
*/
- mutex_lock(&mp->m_icsb_mutex);
+ xfs_icsb_lock(mp);
/*
* Now running atomically.
@@ -2165,7 +2166,7 @@
* Drop the lock and try again in the fast path....
*/
if (!(xfs_icsb_counter_disabled(mp, field))) {
- mutex_unlock(&mp->m_icsb_mutex);
+ xfs_icsb_unlock(mp);
goto again;
}
@@ -2192,7 +2193,7 @@
*/
if (ret != ENOSPC)
xfs_icsb_balance_counter(mp, field, 0, 0);
- mutex_unlock(&mp->m_icsb_mutex);
+ xfs_icsb_unlock(mp);
return ret;
balance_counter:
@@ -2205,7 +2206,7 @@
* do more balances than strictly necessary but it is not
* the common slowpath case.
*/
- mutex_lock(&mp->m_icsb_mutex);
+ xfs_icsb_lock(mp);
/*
* running atomically.
@@ -2216,7 +2217,7 @@
* another balance operation being required.
*/
xfs_icsb_balance_counter(mp, field, 0, delta);
- mutex_unlock(&mp->m_icsb_mutex);
+ xfs_icsb_unlock(mp);
goto again;
}
diff --git a/fs/xfs/xfs_mount.h b/fs/xfs/xfs_mount.h
index a2295df..416c89f 100644
--- a/fs/xfs/xfs_mount.h
+++ b/fs/xfs/xfs_mount.h
@@ -18,6 +18,7 @@
#ifndef __XFS_MOUNT_H__
#define __XFS_MOUNT_H__
+
typedef struct xfs_trans_reservations {
uint tr_write; /* extent alloc trans */
uint tr_itruncate; /* truncate trans */
@@ -564,6 +565,26 @@
}
/*
+ * Per-cpu superblock locking functions
+ */
+#ifdef HAVE_PERCPU_SB
+STATIC_INLINE void
+xfs_icsb_lock(xfs_mount_t *mp)
+{
+ mutex_lock(&mp->m_icsb_mutex);
+}
+
+STATIC_INLINE void
+xfs_icsb_unlock(xfs_mount_t *mp)
+{
+ mutex_unlock(&mp->m_icsb_mutex);
+}
+#else
+#define xfs_icsb_lock(mp)
+#define xfs_icsb_unlock(mp)
+#endif
+
+/*
* This structure is for use by the xfs_mod_incore_sb_batch() routine.
*/
typedef struct xfs_mod_sb {