blob: c282ad41f3d1860bba4301495a7303de40b57d06 [file] [log] [blame]
David Teiglandb3b94fa2006-01-16 16:50:04 +00001/*
2 * Copyright (C) Sistina Software, Inc. 1997-2003 All rights reserved.
Bob Petersonda6dd402007-12-11 18:49:21 -06003 * Copyright (C) 2004-2007 Red Hat, Inc. All rights reserved.
David Teiglandb3b94fa2006-01-16 16:50:04 +00004 *
5 * This copyrighted material is made available to anyone wishing to use,
6 * modify, copy, or redistribute it subject to the terms and conditions
Steven Whitehousee9fc2aa2006-09-01 11:05:15 -04007 * of the GNU General Public License version 2.
David Teiglandb3b94fa2006-01-16 16:50:04 +00008 */
9
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +010010#include <linux/bio.h>
David Teiglandb3b94fa2006-01-16 16:50:04 +000011#include <linux/sched.h>
12#include <linux/slab.h>
13#include <linux/spinlock.h>
14#include <linux/completion.h>
15#include <linux/buffer_head.h>
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +010016#include <linux/statfs.h>
17#include <linux/seq_file.h>
18#include <linux/mount.h>
19#include <linux/kthread.h>
20#include <linux/delay.h>
Steven Whitehouse5c676f62006-02-27 17:23:27 -050021#include <linux/gfs2_ondisk.h>
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +010022#include <linux/crc32.h>
23#include <linux/time.h>
David Teiglandb3b94fa2006-01-16 16:50:04 +000024
25#include "gfs2.h"
Steven Whitehouse5c676f62006-02-27 17:23:27 -050026#include "incore.h"
David Teiglandb3b94fa2006-01-16 16:50:04 +000027#include "bmap.h"
28#include "dir.h"
David Teiglandb3b94fa2006-01-16 16:50:04 +000029#include "glock.h"
30#include "glops.h"
31#include "inode.h"
32#include "log.h"
33#include "meta_io.h"
34#include "quota.h"
35#include "recovery.h"
36#include "rgrp.h"
37#include "super.h"
38#include "trans.h"
Steven Whitehouse5c676f62006-02-27 17:23:27 -050039#include "util.h"
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +010040#include "sys.h"
Steven Whitehouse307cf6e2009-08-26 18:51:04 +010041#include "xattr.h"
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +010042
43#define args_neq(a1, a2, x) ((a1)->ar_##x != (a2)->ar_##x)
44
45enum {
46 Opt_lockproto,
47 Opt_locktable,
48 Opt_hostdata,
49 Opt_spectator,
50 Opt_ignore_local_fs,
51 Opt_localflocks,
52 Opt_localcaching,
53 Opt_debug,
54 Opt_nodebug,
55 Opt_upgrade,
56 Opt_acl,
57 Opt_noacl,
58 Opt_quota_off,
59 Opt_quota_account,
60 Opt_quota_on,
61 Opt_quota,
62 Opt_noquota,
63 Opt_suiddir,
64 Opt_nosuiddir,
65 Opt_data_writeback,
66 Opt_data_ordered,
67 Opt_meta,
68 Opt_discard,
69 Opt_nodiscard,
70 Opt_commit,
Bob Petersond34843d2009-08-24 10:44:18 +010071 Opt_err_withdraw,
72 Opt_err_panic,
Benjamin Marzinski3d3c10f2009-10-20 02:39:44 -050073 Opt_statfs_quantum,
74 Opt_statfs_percent,
75 Opt_quota_quantum,
Christoph Hellwigf25934c2009-10-30 08:03:27 +010076 Opt_barrier,
77 Opt_nobarrier,
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +010078 Opt_error,
79};
80
81static const match_table_t tokens = {
82 {Opt_lockproto, "lockproto=%s"},
83 {Opt_locktable, "locktable=%s"},
84 {Opt_hostdata, "hostdata=%s"},
85 {Opt_spectator, "spectator"},
86 {Opt_ignore_local_fs, "ignore_local_fs"},
87 {Opt_localflocks, "localflocks"},
88 {Opt_localcaching, "localcaching"},
89 {Opt_debug, "debug"},
90 {Opt_nodebug, "nodebug"},
91 {Opt_upgrade, "upgrade"},
92 {Opt_acl, "acl"},
93 {Opt_noacl, "noacl"},
94 {Opt_quota_off, "quota=off"},
95 {Opt_quota_account, "quota=account"},
96 {Opt_quota_on, "quota=on"},
97 {Opt_quota, "quota"},
98 {Opt_noquota, "noquota"},
99 {Opt_suiddir, "suiddir"},
100 {Opt_nosuiddir, "nosuiddir"},
101 {Opt_data_writeback, "data=writeback"},
102 {Opt_data_ordered, "data=ordered"},
103 {Opt_meta, "meta"},
104 {Opt_discard, "discard"},
105 {Opt_nodiscard, "nodiscard"},
106 {Opt_commit, "commit=%d"},
Bob Petersond34843d2009-08-24 10:44:18 +0100107 {Opt_err_withdraw, "errors=withdraw"},
108 {Opt_err_panic, "errors=panic"},
Benjamin Marzinski3d3c10f2009-10-20 02:39:44 -0500109 {Opt_statfs_quantum, "statfs_quantum=%d"},
110 {Opt_statfs_percent, "statfs_percent=%d"},
111 {Opt_quota_quantum, "quota_quantum=%d"},
Christoph Hellwigf25934c2009-10-30 08:03:27 +0100112 {Opt_barrier, "barrier"},
113 {Opt_nobarrier, "nobarrier"},
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100114 {Opt_error, NULL}
115};
116
117/**
118 * gfs2_mount_args - Parse mount options
Steven Whitehousef55073f2009-09-28 10:30:49 +0100119 * @args: The structure into which the parsed options will be written
120 * @options: The options to parse
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100121 *
122 * Return: errno
123 */
124
Steven Whitehousef55073f2009-09-28 10:30:49 +0100125int gfs2_mount_args(struct gfs2_args *args, char *options)
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100126{
127 char *o;
128 int token;
129 substring_t tmp[MAX_OPT_ARGS];
130 int rv;
131
132 /* Split the options into tokens with the "," character and
133 process them */
134
135 while (1) {
136 o = strsep(&options, ",");
137 if (o == NULL)
138 break;
139 if (*o == '\0')
140 continue;
141
142 token = match_token(o, tokens, tmp);
143 switch (token) {
144 case Opt_lockproto:
145 match_strlcpy(args->ar_lockproto, &tmp[0],
146 GFS2_LOCKNAME_LEN);
147 break;
148 case Opt_locktable:
149 match_strlcpy(args->ar_locktable, &tmp[0],
150 GFS2_LOCKNAME_LEN);
151 break;
152 case Opt_hostdata:
153 match_strlcpy(args->ar_hostdata, &tmp[0],
154 GFS2_LOCKNAME_LEN);
155 break;
156 case Opt_spectator:
157 args->ar_spectator = 1;
158 break;
159 case Opt_ignore_local_fs:
160 args->ar_ignore_local_fs = 1;
161 break;
162 case Opt_localflocks:
163 args->ar_localflocks = 1;
164 break;
165 case Opt_localcaching:
166 args->ar_localcaching = 1;
167 break;
168 case Opt_debug:
Bob Petersond34843d2009-08-24 10:44:18 +0100169 if (args->ar_errors == GFS2_ERRORS_PANIC) {
Steven Whitehousef55073f2009-09-28 10:30:49 +0100170 printk(KERN_WARNING "GFS2: -o debug and -o errors=panic "
Bob Petersond34843d2009-08-24 10:44:18 +0100171 "are mutually exclusive.\n");
172 return -EINVAL;
173 }
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100174 args->ar_debug = 1;
175 break;
176 case Opt_nodebug:
177 args->ar_debug = 0;
178 break;
179 case Opt_upgrade:
180 args->ar_upgrade = 1;
181 break;
182 case Opt_acl:
183 args->ar_posix_acl = 1;
184 break;
185 case Opt_noacl:
186 args->ar_posix_acl = 0;
187 break;
188 case Opt_quota_off:
189 case Opt_noquota:
190 args->ar_quota = GFS2_QUOTA_OFF;
191 break;
192 case Opt_quota_account:
193 args->ar_quota = GFS2_QUOTA_ACCOUNT;
194 break;
195 case Opt_quota_on:
196 case Opt_quota:
197 args->ar_quota = GFS2_QUOTA_ON;
198 break;
199 case Opt_suiddir:
200 args->ar_suiddir = 1;
201 break;
202 case Opt_nosuiddir:
203 args->ar_suiddir = 0;
204 break;
205 case Opt_data_writeback:
206 args->ar_data = GFS2_DATA_WRITEBACK;
207 break;
208 case Opt_data_ordered:
209 args->ar_data = GFS2_DATA_ORDERED;
210 break;
211 case Opt_meta:
212 args->ar_meta = 1;
213 break;
214 case Opt_discard:
215 args->ar_discard = 1;
216 break;
217 case Opt_nodiscard:
218 args->ar_discard = 0;
219 break;
220 case Opt_commit:
221 rv = match_int(&tmp[0], &args->ar_commit);
222 if (rv || args->ar_commit <= 0) {
Steven Whitehousef55073f2009-09-28 10:30:49 +0100223 printk(KERN_WARNING "GFS2: commit mount option requires a positive numeric argument\n");
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100224 return rv ? rv : -EINVAL;
225 }
226 break;
Benjamin Marzinski3d3c10f2009-10-20 02:39:44 -0500227 case Opt_statfs_quantum:
228 rv = match_int(&tmp[0], &args->ar_statfs_quantum);
229 if (rv || args->ar_statfs_quantum < 0) {
230 printk(KERN_WARNING "GFS2: statfs_quantum mount option requires a non-negative numeric argument\n");
231 return rv ? rv : -EINVAL;
232 }
233 break;
234 case Opt_quota_quantum:
235 rv = match_int(&tmp[0], &args->ar_quota_quantum);
236 if (rv || args->ar_quota_quantum <= 0) {
237 printk(KERN_WARNING "GFS2: quota_quantum mount option requires a positive numeric argument\n");
238 return rv ? rv : -EINVAL;
239 }
240 break;
241 case Opt_statfs_percent:
242 rv = match_int(&tmp[0], &args->ar_statfs_percent);
243 if (rv || args->ar_statfs_percent < 0 ||
244 args->ar_statfs_percent > 100) {
245 printk(KERN_WARNING "statfs_percent mount option requires a numeric argument between 0 and 100\n");
246 return rv ? rv : -EINVAL;
247 }
248 break;
Bob Petersond34843d2009-08-24 10:44:18 +0100249 case Opt_err_withdraw:
250 args->ar_errors = GFS2_ERRORS_WITHDRAW;
251 break;
252 case Opt_err_panic:
253 if (args->ar_debug) {
Steven Whitehousef55073f2009-09-28 10:30:49 +0100254 printk(KERN_WARNING "GFS2: -o debug and -o errors=panic "
Bob Petersond34843d2009-08-24 10:44:18 +0100255 "are mutually exclusive.\n");
256 return -EINVAL;
257 }
258 args->ar_errors = GFS2_ERRORS_PANIC;
259 break;
Christoph Hellwigf25934c2009-10-30 08:03:27 +0100260 case Opt_barrier:
261 args->ar_nobarrier = 0;
262 break;
263 case Opt_nobarrier:
264 args->ar_nobarrier = 1;
265 break;
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100266 case Opt_error:
267 default:
Steven Whitehousef55073f2009-09-28 10:30:49 +0100268 printk(KERN_WARNING "GFS2: invalid mount option: %s\n", o);
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100269 return -EINVAL;
270 }
271 }
272
273 return 0;
274}
David Teiglandb3b94fa2006-01-16 16:50:04 +0000275
Steven Whitehousefefc03b2008-12-19 15:32:06 +0000276/**
277 * gfs2_jindex_free - Clear all the journal index information
278 * @sdp: The GFS2 superblock
279 *
280 */
281
282void gfs2_jindex_free(struct gfs2_sbd *sdp)
283{
284 struct list_head list, *head;
285 struct gfs2_jdesc *jd;
286 struct gfs2_journal_extent *jext;
287
288 spin_lock(&sdp->sd_jindex_spin);
289 list_add(&list, &sdp->sd_jindex_list);
290 list_del_init(&sdp->sd_jindex_list);
291 sdp->sd_journals = 0;
292 spin_unlock(&sdp->sd_jindex_spin);
293
294 while (!list_empty(&list)) {
295 jd = list_entry(list.next, struct gfs2_jdesc, jd_list);
296 head = &jd->extent_list;
297 while (!list_empty(head)) {
298 jext = list_entry(head->next,
299 struct gfs2_journal_extent,
300 extent_list);
301 list_del(&jext->extent_list);
302 kfree(jext);
303 }
304 list_del(&jd->jd_list);
305 iput(jd->jd_inode);
306 kfree(jd);
307 }
308}
309
David Teiglandb3b94fa2006-01-16 16:50:04 +0000310static struct gfs2_jdesc *jdesc_find_i(struct list_head *head, unsigned int jid)
311{
312 struct gfs2_jdesc *jd;
313 int found = 0;
314
315 list_for_each_entry(jd, head, jd_list) {
316 if (jd->jd_jid == jid) {
317 found = 1;
318 break;
319 }
320 }
321
322 if (!found)
323 jd = NULL;
324
325 return jd;
326}
327
328struct gfs2_jdesc *gfs2_jdesc_find(struct gfs2_sbd *sdp, unsigned int jid)
329{
330 struct gfs2_jdesc *jd;
331
332 spin_lock(&sdp->sd_jindex_spin);
333 jd = jdesc_find_i(&sdp->sd_jindex_list, jid);
334 spin_unlock(&sdp->sd_jindex_spin);
335
336 return jd;
337}
338
David Teiglandb3b94fa2006-01-16 16:50:04 +0000339int gfs2_jdesc_check(struct gfs2_jdesc *jd)
340{
Steven Whitehousefeaa7bb2006-06-14 15:32:57 -0400341 struct gfs2_inode *ip = GFS2_I(jd->jd_inode);
342 struct gfs2_sbd *sdp = GFS2_SB(jd->jd_inode);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000343 int ar;
344 int error;
345
Steven Whitehousec9e98882008-11-04 09:47:33 +0000346 if (ip->i_disksize < (8 << 20) || ip->i_disksize > (1 << 30) ||
347 (ip->i_disksize & (sdp->sd_sb.sb_bsize - 1))) {
David Teiglandb3b94fa2006-01-16 16:50:04 +0000348 gfs2_consist_inode(ip);
349 return -EIO;
350 }
Steven Whitehousec9e98882008-11-04 09:47:33 +0000351 jd->jd_blocks = ip->i_disksize >> sdp->sd_sb.sb_bsize_shift;
David Teiglandb3b94fa2006-01-16 16:50:04 +0000352
Steven Whitehousec9e98882008-11-04 09:47:33 +0000353 error = gfs2_write_alloc_required(ip, 0, ip->i_disksize, &ar);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000354 if (!error && ar) {
355 gfs2_consist_inode(ip);
356 error = -EIO;
357 }
358
359 return error;
360}
361
David Teiglandb3b94fa2006-01-16 16:50:04 +0000362/**
363 * gfs2_make_fs_rw - Turn a Read-Only FS into a Read-Write one
364 * @sdp: the filesystem
365 *
366 * Returns: errno
367 */
368
369int gfs2_make_fs_rw(struct gfs2_sbd *sdp)
370{
Steven Whitehousefeaa7bb2006-06-14 15:32:57 -0400371 struct gfs2_inode *ip = GFS2_I(sdp->sd_jdesc->jd_inode);
Steven Whitehouse5c676f62006-02-27 17:23:27 -0500372 struct gfs2_glock *j_gl = ip->i_gl;
David Teiglandb3b94fa2006-01-16 16:50:04 +0000373 struct gfs2_holder t_gh;
Al Viro55167622006-10-13 21:47:13 -0400374 struct gfs2_log_header_host head;
David Teiglandb3b94fa2006-01-16 16:50:04 +0000375 int error;
376
Steven Whitehouse1c0f4872007-01-22 12:10:39 -0500377 error = gfs2_glock_nq_init(sdp->sd_trans_gl, LM_ST_SHARED, 0, &t_gh);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000378 if (error)
379 return error;
380
Steven Whitehouse1a14d3a2006-11-20 10:37:45 -0500381 j_gl->gl_ops->go_inval(j_gl, DIO_METADATA);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000382
383 error = gfs2_find_jhead(sdp->sd_jdesc, &head);
384 if (error)
385 goto fail;
386
387 if (!(head.lh_flags & GFS2_LOG_HEAD_UNMOUNT)) {
388 gfs2_consist(sdp);
389 error = -EIO;
390 goto fail;
391 }
392
393 /* Initialize some head of the log stuff */
394 sdp->sd_log_sequence = head.lh_sequence + 1;
395 gfs2_log_pointers_init(sdp, head.lh_blkno);
396
David Teiglandb3b94fa2006-01-16 16:50:04 +0000397 error = gfs2_quota_init(sdp);
398 if (error)
Steven Whitehousea91ea692006-09-04 12:04:26 -0400399 goto fail;
David Teiglandb3b94fa2006-01-16 16:50:04 +0000400
401 set_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags);
402
403 gfs2_glock_dq_uninit(&t_gh);
404
405 return 0;
406
Steven Whitehousea91ea692006-09-04 12:04:26 -0400407fail:
David Teiglandb3b94fa2006-01-16 16:50:04 +0000408 t_gh.gh_flags |= GL_NOCACHE;
409 gfs2_glock_dq_uninit(&t_gh);
410
411 return error;
412}
413
Benjamin Marzinski1946f702009-06-25 15:09:51 -0500414void gfs2_statfs_change_in(struct gfs2_statfs_change_host *sc, const void *buf)
Steven Whitehousebb8d8a62007-06-01 14:11:58 +0100415{
416 const struct gfs2_statfs_change *str = buf;
417
418 sc->sc_total = be64_to_cpu(str->sc_total);
419 sc->sc_free = be64_to_cpu(str->sc_free);
420 sc->sc_dinodes = be64_to_cpu(str->sc_dinodes);
421}
422
423static void gfs2_statfs_change_out(const struct gfs2_statfs_change_host *sc, void *buf)
424{
425 struct gfs2_statfs_change *str = buf;
426
427 str->sc_total = cpu_to_be64(sc->sc_total);
428 str->sc_free = cpu_to_be64(sc->sc_free);
429 str->sc_dinodes = cpu_to_be64(sc->sc_dinodes);
430}
431
David Teiglandb3b94fa2006-01-16 16:50:04 +0000432int gfs2_statfs_init(struct gfs2_sbd *sdp)
433{
Steven Whitehousefeaa7bb2006-06-14 15:32:57 -0400434 struct gfs2_inode *m_ip = GFS2_I(sdp->sd_statfs_inode);
Al Virobd209cc2006-10-13 23:43:19 -0400435 struct gfs2_statfs_change_host *m_sc = &sdp->sd_statfs_master;
Steven Whitehousefeaa7bb2006-06-14 15:32:57 -0400436 struct gfs2_inode *l_ip = GFS2_I(sdp->sd_sc_inode);
Al Virobd209cc2006-10-13 23:43:19 -0400437 struct gfs2_statfs_change_host *l_sc = &sdp->sd_statfs_local;
David Teiglandb3b94fa2006-01-16 16:50:04 +0000438 struct buffer_head *m_bh, *l_bh;
439 struct gfs2_holder gh;
440 int error;
441
442 error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, GL_NOCACHE,
443 &gh);
444 if (error)
445 return error;
446
447 error = gfs2_meta_inode_buffer(m_ip, &m_bh);
448 if (error)
449 goto out;
450
451 if (sdp->sd_args.ar_spectator) {
452 spin_lock(&sdp->sd_statfs_spin);
453 gfs2_statfs_change_in(m_sc, m_bh->b_data +
454 sizeof(struct gfs2_dinode));
455 spin_unlock(&sdp->sd_statfs_spin);
456 } else {
457 error = gfs2_meta_inode_buffer(l_ip, &l_bh);
458 if (error)
459 goto out_m_bh;
460
461 spin_lock(&sdp->sd_statfs_spin);
462 gfs2_statfs_change_in(m_sc, m_bh->b_data +
463 sizeof(struct gfs2_dinode));
464 gfs2_statfs_change_in(l_sc, l_bh->b_data +
465 sizeof(struct gfs2_dinode));
466 spin_unlock(&sdp->sd_statfs_spin);
467
468 brelse(l_bh);
469 }
470
Steven Whitehousea91ea692006-09-04 12:04:26 -0400471out_m_bh:
David Teiglandb3b94fa2006-01-16 16:50:04 +0000472 brelse(m_bh);
Steven Whitehousea91ea692006-09-04 12:04:26 -0400473out:
David Teiglandb3b94fa2006-01-16 16:50:04 +0000474 gfs2_glock_dq_uninit(&gh);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000475 return 0;
476}
477
Steven Whitehousecd915492006-09-04 12:49:07 -0400478void gfs2_statfs_change(struct gfs2_sbd *sdp, s64 total, s64 free,
479 s64 dinodes)
David Teiglandb3b94fa2006-01-16 16:50:04 +0000480{
Steven Whitehousefeaa7bb2006-06-14 15:32:57 -0400481 struct gfs2_inode *l_ip = GFS2_I(sdp->sd_sc_inode);
Al Virobd209cc2006-10-13 23:43:19 -0400482 struct gfs2_statfs_change_host *l_sc = &sdp->sd_statfs_local;
Benjamin Marzinski3d3c10f2009-10-20 02:39:44 -0500483 struct gfs2_statfs_change_host *m_sc = &sdp->sd_statfs_master;
David Teiglandb3b94fa2006-01-16 16:50:04 +0000484 struct buffer_head *l_bh;
Benjamin Marzinskic14f5732009-10-26 13:29:47 -0500485 s64 x, y;
486 int need_sync = 0;
David Teiglandb3b94fa2006-01-16 16:50:04 +0000487 int error;
488
489 error = gfs2_meta_inode_buffer(l_ip, &l_bh);
490 if (error)
491 return;
492
Steven Whitehoused4e9c4c2006-01-18 11:19:28 +0000493 gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000494
495 spin_lock(&sdp->sd_statfs_spin);
496 l_sc->sc_total += total;
497 l_sc->sc_free += free;
498 l_sc->sc_dinodes += dinodes;
Steven Whitehouse907b9bc2006-09-25 09:26:04 -0400499 gfs2_statfs_change_out(l_sc, l_bh->b_data + sizeof(struct gfs2_dinode));
Benjamin Marzinskic14f5732009-10-26 13:29:47 -0500500 if (sdp->sd_args.ar_statfs_percent) {
501 x = 100 * l_sc->sc_free;
502 y = m_sc->sc_free * sdp->sd_args.ar_statfs_percent;
503 if (x >= y || x <= -y)
504 need_sync = 1;
505 }
David Teiglandb3b94fa2006-01-16 16:50:04 +0000506 spin_unlock(&sdp->sd_statfs_spin);
507
508 brelse(l_bh);
Benjamin Marzinskic14f5732009-10-26 13:29:47 -0500509 if (need_sync)
Benjamin Marzinski3d3c10f2009-10-20 02:39:44 -0500510 gfs2_wake_up_statfs(sdp);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000511}
512
Benjamin Marzinski1946f702009-06-25 15:09:51 -0500513void update_statfs(struct gfs2_sbd *sdp, struct buffer_head *m_bh,
514 struct buffer_head *l_bh)
515{
516 struct gfs2_inode *m_ip = GFS2_I(sdp->sd_statfs_inode);
517 struct gfs2_inode *l_ip = GFS2_I(sdp->sd_sc_inode);
518 struct gfs2_statfs_change_host *m_sc = &sdp->sd_statfs_master;
519 struct gfs2_statfs_change_host *l_sc = &sdp->sd_statfs_local;
520
521 gfs2_trans_add_bh(l_ip->i_gl, l_bh, 1);
522
523 spin_lock(&sdp->sd_statfs_spin);
524 m_sc->sc_total += l_sc->sc_total;
525 m_sc->sc_free += l_sc->sc_free;
526 m_sc->sc_dinodes += l_sc->sc_dinodes;
527 memset(l_sc, 0, sizeof(struct gfs2_statfs_change));
528 memset(l_bh->b_data + sizeof(struct gfs2_dinode),
529 0, sizeof(struct gfs2_statfs_change));
530 spin_unlock(&sdp->sd_statfs_spin);
531
532 gfs2_trans_add_bh(m_ip->i_gl, m_bh, 1);
533 gfs2_statfs_change_out(m_sc, m_bh->b_data + sizeof(struct gfs2_dinode));
534}
535
Steven Whitehouse8c42d632009-09-11 14:36:44 +0100536int gfs2_statfs_sync(struct super_block *sb, int type)
David Teiglandb3b94fa2006-01-16 16:50:04 +0000537{
Steven Whitehouse8c42d632009-09-11 14:36:44 +0100538 struct gfs2_sbd *sdp = sb->s_fs_info;
Steven Whitehousefeaa7bb2006-06-14 15:32:57 -0400539 struct gfs2_inode *m_ip = GFS2_I(sdp->sd_statfs_inode);
540 struct gfs2_inode *l_ip = GFS2_I(sdp->sd_sc_inode);
Al Virobd209cc2006-10-13 23:43:19 -0400541 struct gfs2_statfs_change_host *m_sc = &sdp->sd_statfs_master;
542 struct gfs2_statfs_change_host *l_sc = &sdp->sd_statfs_local;
David Teiglandb3b94fa2006-01-16 16:50:04 +0000543 struct gfs2_holder gh;
544 struct buffer_head *m_bh, *l_bh;
545 int error;
546
547 error = gfs2_glock_nq_init(m_ip->i_gl, LM_ST_EXCLUSIVE, GL_NOCACHE,
548 &gh);
549 if (error)
550 return error;
551
552 error = gfs2_meta_inode_buffer(m_ip, &m_bh);
553 if (error)
554 goto out;
555
556 spin_lock(&sdp->sd_statfs_spin);
557 gfs2_statfs_change_in(m_sc, m_bh->b_data +
Steven Whitehouse907b9bc2006-09-25 09:26:04 -0400558 sizeof(struct gfs2_dinode));
David Teiglandb3b94fa2006-01-16 16:50:04 +0000559 if (!l_sc->sc_total && !l_sc->sc_free && !l_sc->sc_dinodes) {
560 spin_unlock(&sdp->sd_statfs_spin);
561 goto out_bh;
562 }
563 spin_unlock(&sdp->sd_statfs_spin);
564
565 error = gfs2_meta_inode_buffer(l_ip, &l_bh);
566 if (error)
567 goto out_bh;
568
569 error = gfs2_trans_begin(sdp, 2 * RES_DINODE, 0);
570 if (error)
571 goto out_bh2;
572
Benjamin Marzinski1946f702009-06-25 15:09:51 -0500573 update_statfs(sdp, m_bh, l_bh);
Benjamin Marzinski3d3c10f2009-10-20 02:39:44 -0500574 sdp->sd_statfs_force_sync = 0;
David Teiglandb3b94fa2006-01-16 16:50:04 +0000575
576 gfs2_trans_end(sdp);
577
Steven Whitehousea91ea692006-09-04 12:04:26 -0400578out_bh2:
David Teiglandb3b94fa2006-01-16 16:50:04 +0000579 brelse(l_bh);
Steven Whitehousea91ea692006-09-04 12:04:26 -0400580out_bh:
David Teiglandb3b94fa2006-01-16 16:50:04 +0000581 brelse(m_bh);
Steven Whitehousea91ea692006-09-04 12:04:26 -0400582out:
David Teiglandb3b94fa2006-01-16 16:50:04 +0000583 gfs2_glock_dq_uninit(&gh);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000584 return error;
585}
586
David Teiglandb3b94fa2006-01-16 16:50:04 +0000587struct lfcc {
588 struct list_head list;
589 struct gfs2_holder gh;
590};
591
592/**
593 * gfs2_lock_fs_check_clean - Stop all writes to the FS and check that all
594 * journals are clean
595 * @sdp: the file system
596 * @state: the state to put the transaction lock into
597 * @t_gh: the hold on the transaction lock
598 *
599 * Returns: errno
600 */
601
Adrian Bunk08bc2db2006-04-28 10:59:12 -0400602static int gfs2_lock_fs_check_clean(struct gfs2_sbd *sdp,
603 struct gfs2_holder *t_gh)
David Teiglandb3b94fa2006-01-16 16:50:04 +0000604{
Steven Whitehouse5c676f62006-02-27 17:23:27 -0500605 struct gfs2_inode *ip;
David Teiglandb3b94fa2006-01-16 16:50:04 +0000606 struct gfs2_jdesc *jd;
607 struct lfcc *lfcc;
608 LIST_HEAD(list);
Al Viro55167622006-10-13 21:47:13 -0400609 struct gfs2_log_header_host lh;
David Teiglandb3b94fa2006-01-16 16:50:04 +0000610 int error;
611
David Teiglandb3b94fa2006-01-16 16:50:04 +0000612 list_for_each_entry(jd, &sdp->sd_jindex_list, jd_list) {
613 lfcc = kmalloc(sizeof(struct lfcc), GFP_KERNEL);
614 if (!lfcc) {
615 error = -ENOMEM;
616 goto out;
617 }
Steven Whitehousefeaa7bb2006-06-14 15:32:57 -0400618 ip = GFS2_I(jd->jd_inode);
619 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_SHARED, 0, &lfcc->gh);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000620 if (error) {
621 kfree(lfcc);
622 goto out;
623 }
624 list_add(&lfcc->list, &list);
625 }
626
627 error = gfs2_glock_nq_init(sdp->sd_trans_gl, LM_ST_DEFERRED,
Steven Whitehouse6802e342008-05-21 17:03:22 +0100628 GL_NOCACHE, t_gh);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000629
630 list_for_each_entry(jd, &sdp->sd_jindex_list, jd_list) {
631 error = gfs2_jdesc_check(jd);
632 if (error)
633 break;
634 error = gfs2_find_jhead(jd, &lh);
635 if (error)
636 break;
637 if (!(lh.lh_flags & GFS2_LOG_HEAD_UNMOUNT)) {
638 error = -EBUSY;
639 break;
640 }
641 }
642
643 if (error)
644 gfs2_glock_dq_uninit(t_gh);
645
Steven Whitehousea91ea692006-09-04 12:04:26 -0400646out:
David Teiglandb3b94fa2006-01-16 16:50:04 +0000647 while (!list_empty(&list)) {
648 lfcc = list_entry(list.next, struct lfcc, list);
649 list_del(&lfcc->list);
650 gfs2_glock_dq_uninit(&lfcc->gh);
651 kfree(lfcc);
652 }
David Teiglandb3b94fa2006-01-16 16:50:04 +0000653 return error;
654}
655
656/**
657 * gfs2_freeze_fs - freezes the file system
658 * @sdp: the file system
659 *
660 * This function flushes data and meta data for all machines by
661 * aquiring the transaction log exclusively. All journals are
662 * ensured to be in a clean state as well.
663 *
664 * Returns: errno
665 */
666
667int gfs2_freeze_fs(struct gfs2_sbd *sdp)
668{
669 int error = 0;
670
Steven Whitehousef55ab262006-02-21 12:51:39 +0000671 mutex_lock(&sdp->sd_freeze_lock);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000672
673 if (!sdp->sd_freeze_count++) {
674 error = gfs2_lock_fs_check_clean(sdp, &sdp->sd_freeze_gh);
675 if (error)
676 sdp->sd_freeze_count--;
677 }
678
Steven Whitehousef55ab262006-02-21 12:51:39 +0000679 mutex_unlock(&sdp->sd_freeze_lock);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000680
681 return error;
682}
683
684/**
685 * gfs2_unfreeze_fs - unfreezes the file system
686 * @sdp: the file system
687 *
688 * This function allows the file system to proceed by unlocking
689 * the exclusively held transaction lock. Other GFS2 nodes are
690 * now free to acquire the lock shared and go on with their lives.
691 *
692 */
693
694void gfs2_unfreeze_fs(struct gfs2_sbd *sdp)
695{
Steven Whitehousef55ab262006-02-21 12:51:39 +0000696 mutex_lock(&sdp->sd_freeze_lock);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000697
698 if (sdp->sd_freeze_count && !--sdp->sd_freeze_count)
699 gfs2_glock_dq_uninit(&sdp->sd_freeze_gh);
700
Steven Whitehousef55ab262006-02-21 12:51:39 +0000701 mutex_unlock(&sdp->sd_freeze_lock);
David Teiglandb3b94fa2006-01-16 16:50:04 +0000702}
703
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100704
705/**
706 * gfs2_write_inode - Make sure the inode is stable on the disk
707 * @inode: The inode
708 * @sync: synchronous write flag
709 *
710 * Returns: errno
711 */
712
713static int gfs2_write_inode(struct inode *inode, int sync)
714{
715 struct gfs2_inode *ip = GFS2_I(inode);
716 struct gfs2_sbd *sdp = GFS2_SB(inode);
717 struct gfs2_holder gh;
718 struct buffer_head *bh;
719 struct timespec atime;
720 struct gfs2_dinode *di;
721 int ret = 0;
722
723 /* Check this is a "normal" inode, etc */
724 if (!test_bit(GIF_USER, &ip->i_flags) ||
725 (current->flags & PF_MEMALLOC))
726 return 0;
727 ret = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh);
728 if (ret)
729 goto do_flush;
730 ret = gfs2_trans_begin(sdp, RES_DINODE, 0);
731 if (ret)
732 goto do_unlock;
733 ret = gfs2_meta_inode_buffer(ip, &bh);
734 if (ret == 0) {
735 di = (struct gfs2_dinode *)bh->b_data;
736 atime.tv_sec = be64_to_cpu(di->di_atime);
737 atime.tv_nsec = be32_to_cpu(di->di_atime_nsec);
738 if (timespec_compare(&inode->i_atime, &atime) > 0) {
739 gfs2_trans_add_bh(ip->i_gl, bh, 1);
740 gfs2_dinode_out(ip, bh->b_data);
741 }
742 brelse(bh);
743 }
744 gfs2_trans_end(sdp);
745do_unlock:
746 gfs2_glock_dq_uninit(&gh);
747do_flush:
748 if (sync != 0)
749 gfs2_log_flush(GFS2_SB(inode), ip->i_gl);
750 return ret;
751}
752
753/**
754 * gfs2_make_fs_ro - Turn a Read-Write FS into a Read-Only one
755 * @sdp: the filesystem
756 *
757 * Returns: errno
758 */
759
760static int gfs2_make_fs_ro(struct gfs2_sbd *sdp)
761{
762 struct gfs2_holder t_gh;
763 int error;
764
Benjamin Marzinskib94a1702009-07-23 18:52:34 -0500765 flush_workqueue(gfs2_delete_workqueue);
Steven Whitehouse8c42d632009-09-11 14:36:44 +0100766 gfs2_quota_sync(sdp->sd_vfs, 0);
767 gfs2_statfs_sync(sdp->sd_vfs, 0);
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100768
769 error = gfs2_glock_nq_init(sdp->sd_trans_gl, LM_ST_SHARED, GL_NOCACHE,
770 &t_gh);
771 if (error && !test_bit(SDF_SHUTDOWN, &sdp->sd_flags))
772 return error;
773
774 gfs2_meta_syncfs(sdp);
775 gfs2_log_shutdown(sdp);
776
777 clear_bit(SDF_JOURNAL_LIVE, &sdp->sd_flags);
778
779 if (t_gh.gh_gl)
780 gfs2_glock_dq_uninit(&t_gh);
781
782 gfs2_quota_cleanup(sdp);
783
784 return error;
785}
786
787static int gfs2_umount_recovery_wait(void *word)
788{
789 schedule();
790 return 0;
791}
792
793/**
794 * gfs2_put_super - Unmount the filesystem
795 * @sb: The VFS superblock
796 *
797 */
798
799static void gfs2_put_super(struct super_block *sb)
800{
801 struct gfs2_sbd *sdp = sb->s_fs_info;
802 int error;
803 struct gfs2_jdesc *jd;
804
805 /* Unfreeze the filesystem, if we need to */
806
807 mutex_lock(&sdp->sd_freeze_lock);
808 if (sdp->sd_freeze_count)
809 gfs2_glock_dq_uninit(&sdp->sd_freeze_gh);
810 mutex_unlock(&sdp->sd_freeze_lock);
811
812 /* No more recovery requests */
813 set_bit(SDF_NORECOVERY, &sdp->sd_flags);
814 smp_mb();
815
816 /* Wait on outstanding recovery */
817restart:
818 spin_lock(&sdp->sd_jindex_spin);
819 list_for_each_entry(jd, &sdp->sd_jindex_list, jd_list) {
820 if (!test_bit(JDF_RECOVERY, &jd->jd_flags))
821 continue;
822 spin_unlock(&sdp->sd_jindex_spin);
823 wait_on_bit(&jd->jd_flags, JDF_RECOVERY,
824 gfs2_umount_recovery_wait, TASK_UNINTERRUPTIBLE);
825 goto restart;
826 }
827 spin_unlock(&sdp->sd_jindex_spin);
828
829 kthread_stop(sdp->sd_quotad_process);
830 kthread_stop(sdp->sd_logd_process);
831
832 if (!(sb->s_flags & MS_RDONLY)) {
833 error = gfs2_make_fs_ro(sdp);
834 if (error)
835 gfs2_io_error(sdp);
836 }
837 /* At this point, we're through modifying the disk */
838
839 /* Release stuff */
840
841 iput(sdp->sd_jindex);
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100842 iput(sdp->sd_statfs_inode);
843 iput(sdp->sd_rindex);
844 iput(sdp->sd_quota_inode);
845
846 gfs2_glock_put(sdp->sd_rename_gl);
847 gfs2_glock_put(sdp->sd_trans_gl);
848
849 if (!sdp->sd_args.ar_spectator) {
850 gfs2_glock_dq_uninit(&sdp->sd_journal_gh);
851 gfs2_glock_dq_uninit(&sdp->sd_jinode_gh);
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100852 gfs2_glock_dq_uninit(&sdp->sd_sc_gh);
853 gfs2_glock_dq_uninit(&sdp->sd_qc_gh);
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100854 iput(sdp->sd_sc_inode);
855 iput(sdp->sd_qc_inode);
856 }
857
858 gfs2_glock_dq_uninit(&sdp->sd_live_gh);
859 gfs2_clear_rgrpd(sdp);
860 gfs2_jindex_free(sdp);
861 /* Take apart glock structures and buffer lists */
862 gfs2_gl_hash_clear(sdp);
863 /* Unmount the locking protocol */
864 gfs2_lm_unmount(sdp);
865
866 /* At this point, we're through participating in the lockspace */
867 gfs2_sys_fs_del(sdp);
868}
869
870/**
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100871 * gfs2_sync_fs - sync the filesystem
872 * @sb: the superblock
873 *
874 * Flushes the log to disk.
875 */
876
877static int gfs2_sync_fs(struct super_block *sb, int wait)
878{
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +0100879 if (wait && sb->s_fs_info)
880 gfs2_log_flush(sb->s_fs_info, NULL);
881 return 0;
882}
883
884/**
885 * gfs2_freeze - prevent further writes to the filesystem
886 * @sb: the VFS structure for the filesystem
887 *
888 */
889
890static int gfs2_freeze(struct super_block *sb)
891{
892 struct gfs2_sbd *sdp = sb->s_fs_info;
893 int error;
894
895 if (test_bit(SDF_SHUTDOWN, &sdp->sd_flags))
896 return -EINVAL;
897
898 for (;;) {
899 error = gfs2_freeze_fs(sdp);
900 if (!error)
901 break;
902
903 switch (error) {
904 case -EBUSY:
905 fs_err(sdp, "waiting for recovery before freeze\n");
906 break;
907
908 default:
909 fs_err(sdp, "error freezing FS: %d\n", error);
910 break;
911 }
912
913 fs_err(sdp, "retrying...\n");
914 msleep(1000);
915 }
916 return 0;
917}
918
919/**
920 * gfs2_unfreeze - reallow writes to the filesystem
921 * @sb: the VFS structure for the filesystem
922 *
923 */
924
925static int gfs2_unfreeze(struct super_block *sb)
926{
927 gfs2_unfreeze_fs(sb->s_fs_info);
928 return 0;
929}
930
931/**
932 * statfs_fill - fill in the sg for a given RG
933 * @rgd: the RG
934 * @sc: the sc structure
935 *
936 * Returns: 0 on success, -ESTALE if the LVB is invalid
937 */
938
939static int statfs_slow_fill(struct gfs2_rgrpd *rgd,
940 struct gfs2_statfs_change_host *sc)
941{
942 gfs2_rgrp_verify(rgd);
943 sc->sc_total += rgd->rd_data;
944 sc->sc_free += rgd->rd_free;
945 sc->sc_dinodes += rgd->rd_dinodes;
946 return 0;
947}
948
949/**
950 * gfs2_statfs_slow - Stat a filesystem using asynchronous locking
951 * @sdp: the filesystem
952 * @sc: the sc info that will be returned
953 *
954 * Any error (other than a signal) will cause this routine to fall back
955 * to the synchronous version.
956 *
957 * FIXME: This really shouldn't busy wait like this.
958 *
959 * Returns: errno
960 */
961
962static int gfs2_statfs_slow(struct gfs2_sbd *sdp, struct gfs2_statfs_change_host *sc)
963{
964 struct gfs2_holder ri_gh;
965 struct gfs2_rgrpd *rgd_next;
966 struct gfs2_holder *gha, *gh;
967 unsigned int slots = 64;
968 unsigned int x;
969 int done;
970 int error = 0, err;
971
972 memset(sc, 0, sizeof(struct gfs2_statfs_change_host));
973 gha = kcalloc(slots, sizeof(struct gfs2_holder), GFP_KERNEL);
974 if (!gha)
975 return -ENOMEM;
976
977 error = gfs2_rindex_hold(sdp, &ri_gh);
978 if (error)
979 goto out;
980
981 rgd_next = gfs2_rgrpd_get_first(sdp);
982
983 for (;;) {
984 done = 1;
985
986 for (x = 0; x < slots; x++) {
987 gh = gha + x;
988
989 if (gh->gh_gl && gfs2_glock_poll(gh)) {
990 err = gfs2_glock_wait(gh);
991 if (err) {
992 gfs2_holder_uninit(gh);
993 error = err;
994 } else {
995 if (!error)
996 error = statfs_slow_fill(
997 gh->gh_gl->gl_object, sc);
998 gfs2_glock_dq_uninit(gh);
999 }
1000 }
1001
1002 if (gh->gh_gl)
1003 done = 0;
1004 else if (rgd_next && !error) {
1005 error = gfs2_glock_nq_init(rgd_next->rd_gl,
1006 LM_ST_SHARED,
1007 GL_ASYNC,
1008 gh);
1009 rgd_next = gfs2_rgrpd_get_next(rgd_next);
1010 done = 0;
1011 }
1012
1013 if (signal_pending(current))
1014 error = -ERESTARTSYS;
1015 }
1016
1017 if (done)
1018 break;
1019
1020 yield();
1021 }
1022
1023 gfs2_glock_dq_uninit(&ri_gh);
1024
1025out:
1026 kfree(gha);
1027 return error;
1028}
1029
1030/**
1031 * gfs2_statfs_i - Do a statfs
1032 * @sdp: the filesystem
1033 * @sg: the sg structure
1034 *
1035 * Returns: errno
1036 */
1037
1038static int gfs2_statfs_i(struct gfs2_sbd *sdp, struct gfs2_statfs_change_host *sc)
1039{
1040 struct gfs2_statfs_change_host *m_sc = &sdp->sd_statfs_master;
1041 struct gfs2_statfs_change_host *l_sc = &sdp->sd_statfs_local;
1042
1043 spin_lock(&sdp->sd_statfs_spin);
1044
1045 *sc = *m_sc;
1046 sc->sc_total += l_sc->sc_total;
1047 sc->sc_free += l_sc->sc_free;
1048 sc->sc_dinodes += l_sc->sc_dinodes;
1049
1050 spin_unlock(&sdp->sd_statfs_spin);
1051
1052 if (sc->sc_free < 0)
1053 sc->sc_free = 0;
1054 if (sc->sc_free > sc->sc_total)
1055 sc->sc_free = sc->sc_total;
1056 if (sc->sc_dinodes < 0)
1057 sc->sc_dinodes = 0;
1058
1059 return 0;
1060}
1061
1062/**
1063 * gfs2_statfs - Gather and return stats about the filesystem
1064 * @sb: The superblock
1065 * @statfsbuf: The buffer
1066 *
1067 * Returns: 0 on success or error code
1068 */
1069
1070static int gfs2_statfs(struct dentry *dentry, struct kstatfs *buf)
1071{
1072 struct super_block *sb = dentry->d_inode->i_sb;
1073 struct gfs2_sbd *sdp = sb->s_fs_info;
1074 struct gfs2_statfs_change_host sc;
1075 int error;
1076
1077 if (gfs2_tune_get(sdp, gt_statfs_slow))
1078 error = gfs2_statfs_slow(sdp, &sc);
1079 else
1080 error = gfs2_statfs_i(sdp, &sc);
1081
1082 if (error)
1083 return error;
1084
1085 buf->f_type = GFS2_MAGIC;
1086 buf->f_bsize = sdp->sd_sb.sb_bsize;
1087 buf->f_blocks = sc.sc_total;
1088 buf->f_bfree = sc.sc_free;
1089 buf->f_bavail = sc.sc_free;
1090 buf->f_files = sc.sc_dinodes + sc.sc_free;
1091 buf->f_ffree = sc.sc_free;
1092 buf->f_namelen = GFS2_FNAMESIZE;
1093
1094 return 0;
1095}
1096
1097/**
1098 * gfs2_remount_fs - called when the FS is remounted
1099 * @sb: the filesystem
1100 * @flags: the remount flags
1101 * @data: extra data passed in (not used right now)
1102 *
1103 * Returns: errno
1104 */
1105
1106static int gfs2_remount_fs(struct super_block *sb, int *flags, char *data)
1107{
1108 struct gfs2_sbd *sdp = sb->s_fs_info;
1109 struct gfs2_args args = sdp->sd_args; /* Default to current settings */
1110 struct gfs2_tune *gt = &sdp->sd_tune;
1111 int error;
1112
1113 spin_lock(&gt->gt_spin);
1114 args.ar_commit = gt->gt_log_flush_secs;
Benjamin Marzinski3d3c10f2009-10-20 02:39:44 -05001115 args.ar_quota_quantum = gt->gt_quota_quantum;
1116 if (gt->gt_statfs_slow)
1117 args.ar_statfs_quantum = 0;
1118 else
1119 args.ar_statfs_quantum = gt->gt_statfs_quantum;
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +01001120 spin_unlock(&gt->gt_spin);
Steven Whitehousef55073f2009-09-28 10:30:49 +01001121 error = gfs2_mount_args(&args, data);
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +01001122 if (error)
1123 return error;
1124
1125 /* Not allowed to change locking details */
1126 if (strcmp(args.ar_lockproto, sdp->sd_args.ar_lockproto) ||
1127 strcmp(args.ar_locktable, sdp->sd_args.ar_locktable) ||
1128 strcmp(args.ar_hostdata, sdp->sd_args.ar_hostdata))
1129 return -EINVAL;
1130
1131 /* Some flags must not be changed */
1132 if (args_neq(&args, &sdp->sd_args, spectator) ||
1133 args_neq(&args, &sdp->sd_args, ignore_local_fs) ||
1134 args_neq(&args, &sdp->sd_args, localflocks) ||
1135 args_neq(&args, &sdp->sd_args, localcaching) ||
1136 args_neq(&args, &sdp->sd_args, meta))
1137 return -EINVAL;
1138
1139 if (sdp->sd_args.ar_spectator)
1140 *flags |= MS_RDONLY;
1141
1142 if ((sb->s_flags ^ *flags) & MS_RDONLY) {
1143 if (*flags & MS_RDONLY)
1144 error = gfs2_make_fs_ro(sdp);
1145 else
1146 error = gfs2_make_fs_rw(sdp);
1147 if (error)
1148 return error;
1149 }
1150
1151 sdp->sd_args = args;
1152 if (sdp->sd_args.ar_posix_acl)
1153 sb->s_flags |= MS_POSIXACL;
1154 else
1155 sb->s_flags &= ~MS_POSIXACL;
Christoph Hellwigf25934c2009-10-30 08:03:27 +01001156 if (sdp->sd_args.ar_nobarrier)
1157 set_bit(SDF_NOBARRIERS, &sdp->sd_flags);
1158 else
1159 clear_bit(SDF_NOBARRIERS, &sdp->sd_flags);
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +01001160 spin_lock(&gt->gt_spin);
1161 gt->gt_log_flush_secs = args.ar_commit;
Benjamin Marzinski3d3c10f2009-10-20 02:39:44 -05001162 gt->gt_quota_quantum = args.ar_quota_quantum;
1163 if (args.ar_statfs_quantum) {
1164 gt->gt_statfs_slow = 0;
1165 gt->gt_statfs_quantum = args.ar_statfs_quantum;
1166 }
1167 else {
1168 gt->gt_statfs_slow = 1;
1169 gt->gt_statfs_quantum = 30;
1170 }
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +01001171 spin_unlock(&gt->gt_spin);
1172
Steven Whitehouse8633ecf2009-07-31 11:07:29 +01001173 gfs2_online_uevent(sdp);
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +01001174 return 0;
1175}
1176
1177/**
1178 * gfs2_drop_inode - Drop an inode (test for remote unlink)
1179 * @inode: The inode to drop
1180 *
1181 * If we've received a callback on an iopen lock then its because a
1182 * remote node tried to deallocate the inode but failed due to this node
1183 * still having the inode open. Here we mark the link count zero
1184 * since we know that it must have reached zero if the GLF_DEMOTE flag
1185 * is set on the iopen glock. If we didn't do a disk read since the
1186 * remote node removed the final link then we might otherwise miss
1187 * this event. This check ensures that this node will deallocate the
1188 * inode's blocks, or alternatively pass the baton on to another
1189 * node for later deallocation.
1190 */
1191
1192static void gfs2_drop_inode(struct inode *inode)
1193{
1194 struct gfs2_inode *ip = GFS2_I(inode);
1195
1196 if (test_bit(GIF_USER, &ip->i_flags) && inode->i_nlink) {
1197 struct gfs2_glock *gl = ip->i_iopen_gh.gh_gl;
1198 if (gl && test_bit(GLF_DEMOTE, &gl->gl_flags))
1199 clear_nlink(inode);
1200 }
1201 generic_drop_inode(inode);
1202}
1203
1204/**
1205 * gfs2_clear_inode - Deallocate an inode when VFS is done with it
1206 * @inode: The VFS inode
1207 *
1208 */
1209
1210static void gfs2_clear_inode(struct inode *inode)
1211{
1212 struct gfs2_inode *ip = GFS2_I(inode);
1213
1214 /* This tells us its a "real" inode and not one which only
1215 * serves to contain an address space (see rgrp.c, meta_io.c)
1216 * which therefore doesn't have its own glocks.
1217 */
1218 if (test_bit(GIF_USER, &ip->i_flags)) {
1219 ip->i_gl->gl_object = NULL;
1220 gfs2_glock_put(ip->i_gl);
1221 ip->i_gl = NULL;
1222 if (ip->i_iopen_gh.gh_gl) {
1223 ip->i_iopen_gh.gh_gl->gl_object = NULL;
1224 gfs2_glock_dq_uninit(&ip->i_iopen_gh);
1225 }
1226 }
1227}
1228
1229static int is_ancestor(const struct dentry *d1, const struct dentry *d2)
1230{
1231 do {
1232 if (d1 == d2)
1233 return 1;
1234 d1 = d1->d_parent;
1235 } while (!IS_ROOT(d1));
1236 return 0;
1237}
1238
1239/**
1240 * gfs2_show_options - Show mount options for /proc/mounts
1241 * @s: seq_file structure
1242 * @mnt: vfsmount
1243 *
1244 * Returns: 0 on success or error code
1245 */
1246
1247static int gfs2_show_options(struct seq_file *s, struct vfsmount *mnt)
1248{
1249 struct gfs2_sbd *sdp = mnt->mnt_sb->s_fs_info;
1250 struct gfs2_args *args = &sdp->sd_args;
Benjamin Marzinski3d3c10f2009-10-20 02:39:44 -05001251 int val;
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +01001252
1253 if (is_ancestor(mnt->mnt_root, sdp->sd_master_dir))
1254 seq_printf(s, ",meta");
1255 if (args->ar_lockproto[0])
1256 seq_printf(s, ",lockproto=%s", args->ar_lockproto);
1257 if (args->ar_locktable[0])
1258 seq_printf(s, ",locktable=%s", args->ar_locktable);
1259 if (args->ar_hostdata[0])
1260 seq_printf(s, ",hostdata=%s", args->ar_hostdata);
1261 if (args->ar_spectator)
1262 seq_printf(s, ",spectator");
1263 if (args->ar_ignore_local_fs)
1264 seq_printf(s, ",ignore_local_fs");
1265 if (args->ar_localflocks)
1266 seq_printf(s, ",localflocks");
1267 if (args->ar_localcaching)
1268 seq_printf(s, ",localcaching");
1269 if (args->ar_debug)
1270 seq_printf(s, ",debug");
1271 if (args->ar_upgrade)
1272 seq_printf(s, ",upgrade");
1273 if (args->ar_posix_acl)
1274 seq_printf(s, ",acl");
1275 if (args->ar_quota != GFS2_QUOTA_DEFAULT) {
1276 char *state;
1277 switch (args->ar_quota) {
1278 case GFS2_QUOTA_OFF:
1279 state = "off";
1280 break;
1281 case GFS2_QUOTA_ACCOUNT:
1282 state = "account";
1283 break;
1284 case GFS2_QUOTA_ON:
1285 state = "on";
1286 break;
1287 default:
1288 state = "unknown";
1289 break;
1290 }
1291 seq_printf(s, ",quota=%s", state);
1292 }
1293 if (args->ar_suiddir)
1294 seq_printf(s, ",suiddir");
1295 if (args->ar_data != GFS2_DATA_DEFAULT) {
1296 char *state;
1297 switch (args->ar_data) {
1298 case GFS2_DATA_WRITEBACK:
1299 state = "writeback";
1300 break;
1301 case GFS2_DATA_ORDERED:
1302 state = "ordered";
1303 break;
1304 default:
1305 state = "unknown";
1306 break;
1307 }
1308 seq_printf(s, ",data=%s", state);
1309 }
1310 if (args->ar_discard)
1311 seq_printf(s, ",discard");
Benjamin Marzinski3d3c10f2009-10-20 02:39:44 -05001312 val = sdp->sd_tune.gt_log_flush_secs;
1313 if (val != 60)
1314 seq_printf(s, ",commit=%d", val);
1315 val = sdp->sd_tune.gt_statfs_quantum;
1316 if (val != 30)
1317 seq_printf(s, ",statfs_quantum=%d", val);
1318 val = sdp->sd_tune.gt_quota_quantum;
1319 if (val != 60)
1320 seq_printf(s, ",quota_quantum=%d", val);
1321 if (args->ar_statfs_percent)
1322 seq_printf(s, ",statfs_percent=%d", args->ar_statfs_percent);
Bob Petersond34843d2009-08-24 10:44:18 +01001323 if (args->ar_errors != GFS2_ERRORS_DEFAULT) {
1324 const char *state;
1325
1326 switch (args->ar_errors) {
1327 case GFS2_ERRORS_WITHDRAW:
1328 state = "withdraw";
1329 break;
1330 case GFS2_ERRORS_PANIC:
1331 state = "panic";
1332 break;
1333 default:
1334 state = "unknown";
1335 break;
1336 }
1337 seq_printf(s, ",errors=%s", state);
1338 }
Steven Whitehousecdcfde62009-10-30 10:48:53 +00001339 if (test_bit(SDF_NOBARRIERS, &sdp->sd_flags))
1340 seq_printf(s, ",nobarrier");
1341
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +01001342 return 0;
1343}
1344
1345/*
1346 * We have to (at the moment) hold the inodes main lock to cover
1347 * the gap between unlocking the shared lock on the iopen lock and
1348 * taking the exclusive lock. I'd rather do a shared -> exclusive
1349 * conversion on the iopen lock, but we can change that later. This
1350 * is safe, just less efficient.
1351 */
1352
1353static void gfs2_delete_inode(struct inode *inode)
1354{
1355 struct gfs2_sbd *sdp = inode->i_sb->s_fs_info;
1356 struct gfs2_inode *ip = GFS2_I(inode);
1357 struct gfs2_holder gh;
1358 int error;
1359
1360 if (!test_bit(GIF_USER, &ip->i_flags))
1361 goto out;
1362
1363 error = gfs2_glock_nq_init(ip->i_gl, LM_ST_EXCLUSIVE, 0, &gh);
1364 if (unlikely(error)) {
1365 gfs2_glock_dq_uninit(&ip->i_iopen_gh);
1366 goto out;
1367 }
1368
Steven Whitehouseacf7e242009-09-08 18:00:30 +01001369 error = gfs2_check_blk_type(sdp, ip->i_no_addr, GFS2_BLKST_UNLINKED);
1370 if (error)
1371 goto out_truncate;
1372
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +01001373 gfs2_glock_dq_wait(&ip->i_iopen_gh);
1374 gfs2_holder_reinit(LM_ST_EXCLUSIVE, LM_FLAG_TRY_1CB | GL_NOCACHE, &ip->i_iopen_gh);
1375 error = gfs2_glock_nq(&ip->i_iopen_gh);
1376 if (error)
1377 goto out_truncate;
1378
1379 if (S_ISDIR(inode->i_mode) &&
1380 (ip->i_diskflags & GFS2_DIF_EXHASH)) {
1381 error = gfs2_dir_exhash_dealloc(ip);
1382 if (error)
1383 goto out_unlock;
1384 }
1385
1386 if (ip->i_eattr) {
1387 error = gfs2_ea_dealloc(ip);
1388 if (error)
1389 goto out_unlock;
1390 }
1391
1392 if (!gfs2_is_stuffed(ip)) {
1393 error = gfs2_file_dealloc(ip);
1394 if (error)
1395 goto out_unlock;
1396 }
1397
1398 error = gfs2_dinode_dealloc(ip);
1399 if (error)
1400 goto out_unlock;
1401
1402out_truncate:
1403 error = gfs2_trans_begin(sdp, 0, sdp->sd_jdesc->jd_blocks);
1404 if (error)
1405 goto out_unlock;
1406 /* Needs to be done before glock release & also in a transaction */
1407 truncate_inode_pages(&inode->i_data, 0);
1408 gfs2_trans_end(sdp);
1409
1410out_unlock:
1411 if (test_bit(HIF_HOLDER, &ip->i_iopen_gh.gh_iflags))
1412 gfs2_glock_dq(&ip->i_iopen_gh);
1413 gfs2_holder_uninit(&ip->i_iopen_gh);
1414 gfs2_glock_dq_uninit(&gh);
1415 if (error && error != GLR_TRYFAILED && error != -EROFS)
1416 fs_warn(sdp, "gfs2_delete_inode: %d\n", error);
1417out:
1418 truncate_inode_pages(&inode->i_data, 0);
1419 clear_inode(inode);
1420}
1421
1422static struct inode *gfs2_alloc_inode(struct super_block *sb)
1423{
1424 struct gfs2_inode *ip;
1425
1426 ip = kmem_cache_alloc(gfs2_inode_cachep, GFP_KERNEL);
1427 if (ip) {
1428 ip->i_flags = 0;
1429 ip->i_gl = NULL;
1430 }
1431 return &ip->i_inode;
1432}
1433
1434static void gfs2_destroy_inode(struct inode *inode)
1435{
1436 kmem_cache_free(gfs2_inode_cachep, inode);
1437}
1438
1439const struct super_operations gfs2_super_ops = {
1440 .alloc_inode = gfs2_alloc_inode,
1441 .destroy_inode = gfs2_destroy_inode,
1442 .write_inode = gfs2_write_inode,
1443 .delete_inode = gfs2_delete_inode,
1444 .put_super = gfs2_put_super,
Steven Whitehouse9e6e0a12009-05-22 10:36:01 +01001445 .sync_fs = gfs2_sync_fs,
1446 .freeze_fs = gfs2_freeze,
1447 .unfreeze_fs = gfs2_unfreeze,
1448 .statfs = gfs2_statfs,
1449 .remount_fs = gfs2_remount_fs,
1450 .clear_inode = gfs2_clear_inode,
1451 .drop_inode = gfs2_drop_inode,
1452 .show_options = gfs2_show_options,
1453};
1454