blob: 1c6cfb59128d6b522139cd26274efe76ca054415 [file] [log] [blame]
Koji Sato29619802009-04-06 19:01:31 -07001/*
2 * cpfile.c - NILFS checkpoint file.
3 *
4 * Copyright (C) 2006-2008 Nippon Telegraph and Telephone Corporation.
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
10 *
11 * This program is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 * GNU General Public License for more details.
15 *
16 * You should have received a copy of the GNU General Public License
17 * along with this program; if not, write to the Free Software
18 * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
19 *
20 * Written by Koji Sato <koji@osrg.net>.
21 */
22
23#include <linux/kernel.h>
24#include <linux/fs.h>
25#include <linux/string.h>
26#include <linux/buffer_head.h>
27#include <linux/errno.h>
28#include <linux/nilfs2_fs.h>
29#include "mdt.h"
30#include "cpfile.h"
31
32
33static inline unsigned long
34nilfs_cpfile_checkpoints_per_block(const struct inode *cpfile)
35{
36 return NILFS_MDT(cpfile)->mi_entries_per_block;
37}
38
39/* block number from the beginning of the file */
40static unsigned long
41nilfs_cpfile_get_blkoff(const struct inode *cpfile, __u64 cno)
42{
Ryusuke Konishi1f5abe72009-04-06 19:01:55 -070043 __u64 tcno = cno + NILFS_MDT(cpfile)->mi_first_entry_offset - 1;
Koji Sato29619802009-04-06 19:01:31 -070044 do_div(tcno, nilfs_cpfile_checkpoints_per_block(cpfile));
45 return (unsigned long)tcno;
46}
47
48/* offset in block */
49static unsigned long
50nilfs_cpfile_get_offset(const struct inode *cpfile, __u64 cno)
51{
52 __u64 tcno = cno + NILFS_MDT(cpfile)->mi_first_entry_offset - 1;
53 return do_div(tcno, nilfs_cpfile_checkpoints_per_block(cpfile));
54}
55
56static unsigned long
57nilfs_cpfile_checkpoints_in_block(const struct inode *cpfile,
58 __u64 curr,
59 __u64 max)
60{
61 return min_t(__u64,
62 nilfs_cpfile_checkpoints_per_block(cpfile) -
63 nilfs_cpfile_get_offset(cpfile, curr),
64 max - curr);
65}
66
67static inline int nilfs_cpfile_is_in_first(const struct inode *cpfile,
68 __u64 cno)
69{
70 return nilfs_cpfile_get_blkoff(cpfile, cno) == 0;
71}
72
73static unsigned int
74nilfs_cpfile_block_add_valid_checkpoints(const struct inode *cpfile,
75 struct buffer_head *bh,
76 void *kaddr,
77 unsigned int n)
78{
79 struct nilfs_checkpoint *cp = kaddr + bh_offset(bh);
80 unsigned int count;
81
82 count = le32_to_cpu(cp->cp_checkpoints_count) + n;
83 cp->cp_checkpoints_count = cpu_to_le32(count);
84 return count;
85}
86
87static unsigned int
88nilfs_cpfile_block_sub_valid_checkpoints(const struct inode *cpfile,
89 struct buffer_head *bh,
90 void *kaddr,
91 unsigned int n)
92{
93 struct nilfs_checkpoint *cp = kaddr + bh_offset(bh);
94 unsigned int count;
95
Ryusuke Konishi1f5abe72009-04-06 19:01:55 -070096 WARN_ON(le32_to_cpu(cp->cp_checkpoints_count) < n);
Koji Sato29619802009-04-06 19:01:31 -070097 count = le32_to_cpu(cp->cp_checkpoints_count) - n;
98 cp->cp_checkpoints_count = cpu_to_le32(count);
99 return count;
100}
101
102static inline struct nilfs_cpfile_header *
103nilfs_cpfile_block_get_header(const struct inode *cpfile,
104 struct buffer_head *bh,
105 void *kaddr)
106{
107 return kaddr + bh_offset(bh);
108}
109
110static struct nilfs_checkpoint *
111nilfs_cpfile_block_get_checkpoint(const struct inode *cpfile, __u64 cno,
112 struct buffer_head *bh,
113 void *kaddr)
114{
115 return kaddr + bh_offset(bh) + nilfs_cpfile_get_offset(cpfile, cno) *
116 NILFS_MDT(cpfile)->mi_entry_size;
117}
118
119static void nilfs_cpfile_block_init(struct inode *cpfile,
120 struct buffer_head *bh,
121 void *kaddr)
122{
123 struct nilfs_checkpoint *cp = kaddr + bh_offset(bh);
124 size_t cpsz = NILFS_MDT(cpfile)->mi_entry_size;
125 int n = nilfs_cpfile_checkpoints_per_block(cpfile);
126
127 while (n-- > 0) {
128 nilfs_checkpoint_set_invalid(cp);
129 cp = (void *)cp + cpsz;
130 }
131}
132
133static inline int nilfs_cpfile_get_header_block(struct inode *cpfile,
134 struct buffer_head **bhp)
135{
136 return nilfs_mdt_get_block(cpfile, 0, 0, NULL, bhp);
137}
138
139static inline int nilfs_cpfile_get_checkpoint_block(struct inode *cpfile,
140 __u64 cno,
141 int create,
142 struct buffer_head **bhp)
143{
144 return nilfs_mdt_get_block(cpfile,
145 nilfs_cpfile_get_blkoff(cpfile, cno),
146 create, nilfs_cpfile_block_init, bhp);
147}
148
149static inline int nilfs_cpfile_delete_checkpoint_block(struct inode *cpfile,
150 __u64 cno)
151{
152 return nilfs_mdt_delete_block(cpfile,
153 nilfs_cpfile_get_blkoff(cpfile, cno));
154}
155
156/**
157 * nilfs_cpfile_get_checkpoint - get a checkpoint
158 * @cpfile: inode of checkpoint file
159 * @cno: checkpoint number
160 * @create: create flag
161 * @cpp: pointer to a checkpoint
162 * @bhp: pointer to a buffer head
163 *
164 * Description: nilfs_cpfile_get_checkpoint() acquires the checkpoint
165 * specified by @cno. A new checkpoint will be created if @cno is the current
166 * checkpoint number and @create is nonzero.
167 *
168 * Return Value: On success, 0 is returned, and the checkpoint and the
169 * buffer head of the buffer on which the checkpoint is located are stored in
170 * the place pointed by @cpp and @bhp, respectively. On error, one of the
171 * following negative error codes is returned.
172 *
173 * %-EIO - I/O error.
174 *
175 * %-ENOMEM - Insufficient amount of memory available.
176 *
177 * %-ENOENT - No such checkpoint.
Ryusuke Konishi1f5abe72009-04-06 19:01:55 -0700178 *
179 * %-EINVAL - invalid checkpoint.
Koji Sato29619802009-04-06 19:01:31 -0700180 */
181int nilfs_cpfile_get_checkpoint(struct inode *cpfile,
182 __u64 cno,
183 int create,
184 struct nilfs_checkpoint **cpp,
185 struct buffer_head **bhp)
186{
187 struct buffer_head *header_bh, *cp_bh;
188 struct nilfs_cpfile_header *header;
189 struct nilfs_checkpoint *cp;
190 void *kaddr;
191 int ret;
192
Ryusuke Konishi1f5abe72009-04-06 19:01:55 -0700193 if (unlikely(cno < 1 || cno > nilfs_mdt_cno(cpfile) ||
194 (cno < nilfs_mdt_cno(cpfile) && create)))
195 return -EINVAL;
Koji Sato29619802009-04-06 19:01:31 -0700196
197 down_write(&NILFS_MDT(cpfile)->mi_sem);
198
199 ret = nilfs_cpfile_get_header_block(cpfile, &header_bh);
200 if (ret < 0)
201 goto out_sem;
202 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, create, &cp_bh);
203 if (ret < 0)
204 goto out_header;
205 kaddr = kmap(cp_bh->b_page);
206 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr);
207 if (nilfs_checkpoint_invalid(cp)) {
208 if (!create) {
209 kunmap(cp_bh->b_page);
210 brelse(cp_bh);
211 ret = -ENOENT;
212 goto out_header;
213 }
214 /* a newly-created checkpoint */
215 nilfs_checkpoint_clear_invalid(cp);
216 if (!nilfs_cpfile_is_in_first(cpfile, cno))
217 nilfs_cpfile_block_add_valid_checkpoints(cpfile, cp_bh,
218 kaddr, 1);
219 nilfs_mdt_mark_buffer_dirty(cp_bh);
220
221 kaddr = kmap_atomic(header_bh->b_page, KM_USER0);
222 header = nilfs_cpfile_block_get_header(cpfile, header_bh,
223 kaddr);
224 le64_add_cpu(&header->ch_ncheckpoints, 1);
225 kunmap_atomic(kaddr, KM_USER0);
226 nilfs_mdt_mark_buffer_dirty(header_bh);
227 nilfs_mdt_mark_dirty(cpfile);
228 }
229
230 if (cpp != NULL)
231 *cpp = cp;
232 *bhp = cp_bh;
233
234 out_header:
235 brelse(header_bh);
236
237 out_sem:
238 up_write(&NILFS_MDT(cpfile)->mi_sem);
239 return ret;
240}
241
242/**
243 * nilfs_cpfile_put_checkpoint - put a checkpoint
244 * @cpfile: inode of checkpoint file
245 * @cno: checkpoint number
246 * @bh: buffer head
247 *
248 * Description: nilfs_cpfile_put_checkpoint() releases the checkpoint
249 * specified by @cno. @bh must be the buffer head which has been returned by
250 * a previous call to nilfs_cpfile_get_checkpoint() with @cno.
251 */
252void nilfs_cpfile_put_checkpoint(struct inode *cpfile, __u64 cno,
253 struct buffer_head *bh)
254{
255 kunmap(bh->b_page);
256 brelse(bh);
257}
258
259/**
260 * nilfs_cpfile_delete_checkpoints - delete checkpoints
261 * @cpfile: inode of checkpoint file
262 * @start: start checkpoint number
263 * @end: end checkpoint numer
264 *
265 * Description: nilfs_cpfile_delete_checkpoints() deletes the checkpoints in
266 * the period from @start to @end, excluding @end itself. The checkpoints
267 * which have been already deleted are ignored.
268 *
269 * Return Value: On success, 0 is returned. On error, one of the following
270 * negative error codes is returned.
271 *
272 * %-EIO - I/O error.
273 *
274 * %-ENOMEM - Insufficient amount of memory available.
275 *
276 * %-EINVAL - invalid checkpoints.
277 */
278int nilfs_cpfile_delete_checkpoints(struct inode *cpfile,
279 __u64 start,
280 __u64 end)
281{
282 struct buffer_head *header_bh, *cp_bh;
283 struct nilfs_cpfile_header *header;
284 struct nilfs_checkpoint *cp;
285 size_t cpsz = NILFS_MDT(cpfile)->mi_entry_size;
286 __u64 cno;
287 void *kaddr;
288 unsigned long tnicps;
289 int ret, ncps, nicps, count, i;
290
Ryusuke Konishi1f5abe72009-04-06 19:01:55 -0700291 if (unlikely(start == 0 || start > end)) {
292 printk(KERN_ERR "%s: invalid range of checkpoint numbers: "
293 "[%llu, %llu)\n", __func__,
294 (unsigned long long)start, (unsigned long long)end);
295 return -EINVAL;
Koji Sato29619802009-04-06 19:01:31 -0700296 }
297
Koji Sato29619802009-04-06 19:01:31 -0700298 down_write(&NILFS_MDT(cpfile)->mi_sem);
299
300 ret = nilfs_cpfile_get_header_block(cpfile, &header_bh);
301 if (ret < 0)
302 goto out_sem;
303 tnicps = 0;
304
305 for (cno = start; cno < end; cno += ncps) {
306 ncps = nilfs_cpfile_checkpoints_in_block(cpfile, cno, end);
307 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &cp_bh);
308 if (ret < 0) {
309 if (ret != -ENOENT)
Jiro SEKIBAd9a0a342009-07-04 23:00:53 +0900310 break;
Koji Sato29619802009-04-06 19:01:31 -0700311 /* skip hole */
312 ret = 0;
313 continue;
314 }
315
316 kaddr = kmap_atomic(cp_bh->b_page, KM_USER0);
317 cp = nilfs_cpfile_block_get_checkpoint(
318 cpfile, cno, cp_bh, kaddr);
319 nicps = 0;
320 for (i = 0; i < ncps; i++, cp = (void *)cp + cpsz) {
Ryusuke Konishi1f5abe72009-04-06 19:01:55 -0700321 WARN_ON(nilfs_checkpoint_snapshot(cp));
Koji Sato29619802009-04-06 19:01:31 -0700322 if (!nilfs_checkpoint_invalid(cp)) {
323 nilfs_checkpoint_set_invalid(cp);
324 nicps++;
325 }
326 }
327 if (nicps > 0) {
328 tnicps += nicps;
329 nilfs_mdt_mark_buffer_dirty(cp_bh);
330 nilfs_mdt_mark_dirty(cpfile);
331 if (!nilfs_cpfile_is_in_first(cpfile, cno) &&
332 (count = nilfs_cpfile_block_sub_valid_checkpoints(
333 cpfile, cp_bh, kaddr, nicps)) == 0) {
334 /* make hole */
335 kunmap_atomic(kaddr, KM_USER0);
336 brelse(cp_bh);
337 ret = nilfs_cpfile_delete_checkpoint_block(
338 cpfile, cno);
339 if (ret == 0)
340 continue;
341 printk(KERN_ERR "%s: cannot delete block\n",
342 __func__);
Jiro SEKIBAd9a0a342009-07-04 23:00:53 +0900343 break;
Koji Sato29619802009-04-06 19:01:31 -0700344 }
345 }
346
347 kunmap_atomic(kaddr, KM_USER0);
348 brelse(cp_bh);
349 }
350
351 if (tnicps > 0) {
352 kaddr = kmap_atomic(header_bh->b_page, KM_USER0);
353 header = nilfs_cpfile_block_get_header(cpfile, header_bh,
354 kaddr);
Koji Sato6c98cd42009-04-06 19:01:32 -0700355 le64_add_cpu(&header->ch_ncheckpoints, -(u64)tnicps);
Koji Sato29619802009-04-06 19:01:31 -0700356 nilfs_mdt_mark_buffer_dirty(header_bh);
357 nilfs_mdt_mark_dirty(cpfile);
358 kunmap_atomic(kaddr, KM_USER0);
359 }
Ryusuke Konishi62013ab2009-05-30 21:50:58 +0900360
Koji Sato29619802009-04-06 19:01:31 -0700361 brelse(header_bh);
362
363 out_sem:
364 up_write(&NILFS_MDT(cpfile)->mi_sem);
365 return ret;
366}
367
368static void nilfs_cpfile_checkpoint_to_cpinfo(struct inode *cpfile,
369 struct nilfs_checkpoint *cp,
370 struct nilfs_cpinfo *ci)
371{
372 ci->ci_flags = le32_to_cpu(cp->cp_flags);
373 ci->ci_cno = le64_to_cpu(cp->cp_cno);
374 ci->ci_create = le64_to_cpu(cp->cp_create);
375 ci->ci_nblk_inc = le64_to_cpu(cp->cp_nblk_inc);
376 ci->ci_inodes_count = le64_to_cpu(cp->cp_inodes_count);
377 ci->ci_blocks_count = le64_to_cpu(cp->cp_blocks_count);
378 ci->ci_next = le64_to_cpu(cp->cp_snapshot_list.ssl_next);
379}
380
Ryusuke Konishi76068c42009-04-06 19:01:50 -0700381static ssize_t nilfs_cpfile_do_get_cpinfo(struct inode *cpfile, __u64 *cnop,
Ryusuke Konishi003ff182009-05-12 03:58:47 +0900382 void *buf, unsigned cisz, size_t nci)
Koji Sato29619802009-04-06 19:01:31 -0700383{
384 struct nilfs_checkpoint *cp;
Ryusuke Konishi003ff182009-05-12 03:58:47 +0900385 struct nilfs_cpinfo *ci = buf;
Koji Sato29619802009-04-06 19:01:31 -0700386 struct buffer_head *bh;
387 size_t cpsz = NILFS_MDT(cpfile)->mi_entry_size;
Ryusuke Konishi76068c42009-04-06 19:01:50 -0700388 __u64 cur_cno = nilfs_mdt_cno(cpfile), cno = *cnop;
Koji Sato29619802009-04-06 19:01:31 -0700389 void *kaddr;
390 int n, ret;
391 int ncps, i;
392
Ryusuke Konishi1f5abe72009-04-06 19:01:55 -0700393 if (cno == 0)
394 return -ENOENT; /* checkpoint number 0 is invalid */
Koji Sato29619802009-04-06 19:01:31 -0700395 down_read(&NILFS_MDT(cpfile)->mi_sem);
396
397 for (n = 0; cno < cur_cno && n < nci; cno += ncps) {
398 ncps = nilfs_cpfile_checkpoints_in_block(cpfile, cno, cur_cno);
399 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &bh);
400 if (ret < 0) {
401 if (ret != -ENOENT)
402 goto out;
403 continue; /* skip hole */
404 }
405
406 kaddr = kmap_atomic(bh->b_page, KM_USER0);
407 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, bh, kaddr);
408 for (i = 0; i < ncps && n < nci; i++, cp = (void *)cp + cpsz) {
Ryusuke Konishi003ff182009-05-12 03:58:47 +0900409 if (!nilfs_checkpoint_invalid(cp)) {
410 nilfs_cpfile_checkpoint_to_cpinfo(cpfile, cp,
411 ci);
412 ci = (void *)ci + cisz;
413 n++;
414 }
Koji Sato29619802009-04-06 19:01:31 -0700415 }
416 kunmap_atomic(kaddr, KM_USER0);
417 brelse(bh);
418 }
419
420 ret = n;
Ryusuke Konishi003ff182009-05-12 03:58:47 +0900421 if (n > 0) {
422 ci = (void *)ci - cisz;
423 *cnop = ci->ci_cno + 1;
424 }
Koji Sato29619802009-04-06 19:01:31 -0700425
426 out:
427 up_read(&NILFS_MDT(cpfile)->mi_sem);
428 return ret;
429}
430
Ryusuke Konishib028fcf2009-04-06 19:01:47 -0700431static ssize_t nilfs_cpfile_do_get_ssinfo(struct inode *cpfile, __u64 *cnop,
Ryusuke Konishi003ff182009-05-12 03:58:47 +0900432 void *buf, unsigned cisz, size_t nci)
Koji Sato29619802009-04-06 19:01:31 -0700433{
434 struct buffer_head *bh;
435 struct nilfs_cpfile_header *header;
436 struct nilfs_checkpoint *cp;
Ryusuke Konishi003ff182009-05-12 03:58:47 +0900437 struct nilfs_cpinfo *ci = buf;
Ryusuke Konishib028fcf2009-04-06 19:01:47 -0700438 __u64 curr = *cnop, next;
Koji Sato29619802009-04-06 19:01:31 -0700439 unsigned long curr_blkoff, next_blkoff;
440 void *kaddr;
Ryusuke Konishi7fa10d22009-04-06 19:01:48 -0700441 int n = 0, ret;
Koji Sato29619802009-04-06 19:01:31 -0700442
443 down_read(&NILFS_MDT(cpfile)->mi_sem);
444
Ryusuke Konishib028fcf2009-04-06 19:01:47 -0700445 if (curr == 0) {
Koji Sato29619802009-04-06 19:01:31 -0700446 ret = nilfs_cpfile_get_header_block(cpfile, &bh);
447 if (ret < 0)
448 goto out;
449 kaddr = kmap_atomic(bh->b_page, KM_USER0);
450 header = nilfs_cpfile_block_get_header(cpfile, bh, kaddr);
451 curr = le64_to_cpu(header->ch_snapshot_list.ssl_next);
452 kunmap_atomic(kaddr, KM_USER0);
453 brelse(bh);
454 if (curr == 0) {
455 ret = 0;
456 goto out;
457 }
Ryusuke Konishib028fcf2009-04-06 19:01:47 -0700458 } else if (unlikely(curr == ~(__u64)0)) {
459 ret = 0;
460 goto out;
461 }
462
Koji Sato29619802009-04-06 19:01:31 -0700463 curr_blkoff = nilfs_cpfile_get_blkoff(cpfile, curr);
464 ret = nilfs_cpfile_get_checkpoint_block(cpfile, curr, 0, &bh);
Ryusuke Konishi7fa10d22009-04-06 19:01:48 -0700465 if (unlikely(ret < 0)) {
466 if (ret == -ENOENT)
467 ret = 0; /* No snapshots (started from a hole block) */
Koji Sato29619802009-04-06 19:01:31 -0700468 goto out;
Ryusuke Konishi7fa10d22009-04-06 19:01:48 -0700469 }
Koji Sato29619802009-04-06 19:01:31 -0700470 kaddr = kmap_atomic(bh->b_page, KM_USER0);
Ryusuke Konishi7fa10d22009-04-06 19:01:48 -0700471 while (n < nci) {
472 cp = nilfs_cpfile_block_get_checkpoint(cpfile, curr, bh, kaddr);
473 curr = ~(__u64)0; /* Terminator */
474 if (unlikely(nilfs_checkpoint_invalid(cp) ||
475 !nilfs_checkpoint_snapshot(cp)))
Koji Sato29619802009-04-06 19:01:31 -0700476 break;
Ryusuke Konishi003ff182009-05-12 03:58:47 +0900477 nilfs_cpfile_checkpoint_to_cpinfo(cpfile, cp, ci);
478 ci = (void *)ci + cisz;
479 n++;
Ryusuke Konishi7fa10d22009-04-06 19:01:48 -0700480 next = le64_to_cpu(cp->cp_snapshot_list.ssl_next);
481 if (next == 0)
482 break; /* reach end of the snapshot list */
483
Koji Sato29619802009-04-06 19:01:31 -0700484 next_blkoff = nilfs_cpfile_get_blkoff(cpfile, next);
485 if (curr_blkoff != next_blkoff) {
486 kunmap_atomic(kaddr, KM_USER0);
487 brelse(bh);
488 ret = nilfs_cpfile_get_checkpoint_block(cpfile, next,
489 0, &bh);
Ryusuke Konishi7fa10d22009-04-06 19:01:48 -0700490 if (unlikely(ret < 0)) {
491 WARN_ON(ret == -ENOENT);
Koji Sato29619802009-04-06 19:01:31 -0700492 goto out;
Ryusuke Konishi7fa10d22009-04-06 19:01:48 -0700493 }
Koji Sato29619802009-04-06 19:01:31 -0700494 kaddr = kmap_atomic(bh->b_page, KM_USER0);
495 }
496 curr = next;
497 curr_blkoff = next_blkoff;
498 }
499 kunmap_atomic(kaddr, KM_USER0);
500 brelse(bh);
Ryusuke Konishib028fcf2009-04-06 19:01:47 -0700501 *cnop = curr;
Koji Sato29619802009-04-06 19:01:31 -0700502 ret = n;
503
504 out:
505 up_read(&NILFS_MDT(cpfile)->mi_sem);
506 return ret;
507}
508
509/**
510 * nilfs_cpfile_get_cpinfo -
511 * @cpfile:
512 * @cno:
513 * @ci:
514 * @nci:
515 */
Ryusuke Konishib028fcf2009-04-06 19:01:47 -0700516
517ssize_t nilfs_cpfile_get_cpinfo(struct inode *cpfile, __u64 *cnop, int mode,
Ryusuke Konishi003ff182009-05-12 03:58:47 +0900518 void *buf, unsigned cisz, size_t nci)
Koji Sato29619802009-04-06 19:01:31 -0700519{
520 switch (mode) {
521 case NILFS_CHECKPOINT:
Ryusuke Konishi003ff182009-05-12 03:58:47 +0900522 return nilfs_cpfile_do_get_cpinfo(cpfile, cnop, buf, cisz, nci);
Koji Sato29619802009-04-06 19:01:31 -0700523 case NILFS_SNAPSHOT:
Ryusuke Konishi003ff182009-05-12 03:58:47 +0900524 return nilfs_cpfile_do_get_ssinfo(cpfile, cnop, buf, cisz, nci);
Koji Sato29619802009-04-06 19:01:31 -0700525 default:
526 return -EINVAL;
527 }
528}
529
530/**
531 * nilfs_cpfile_delete_checkpoint -
532 * @cpfile:
533 * @cno:
534 */
535int nilfs_cpfile_delete_checkpoint(struct inode *cpfile, __u64 cno)
536{
537 struct nilfs_cpinfo ci;
Ryusuke Konishi76068c42009-04-06 19:01:50 -0700538 __u64 tcno = cno;
Koji Sato29619802009-04-06 19:01:31 -0700539 ssize_t nci;
Koji Sato29619802009-04-06 19:01:31 -0700540
Ryusuke Konishi003ff182009-05-12 03:58:47 +0900541 nci = nilfs_cpfile_do_get_cpinfo(cpfile, &tcno, &ci, sizeof(ci), 1);
Koji Sato29619802009-04-06 19:01:31 -0700542 if (nci < 0)
543 return nci;
544 else if (nci == 0 || ci.ci_cno != cno)
545 return -ENOENT;
Ryusuke Konishi30c25be2009-05-30 19:08:09 +0900546 else if (nilfs_cpinfo_snapshot(&ci))
547 return -EBUSY;
Koji Sato29619802009-04-06 19:01:31 -0700548
549 return nilfs_cpfile_delete_checkpoints(cpfile, cno, cno + 1);
550}
551
552static struct nilfs_snapshot_list *
553nilfs_cpfile_block_get_snapshot_list(const struct inode *cpfile,
554 __u64 cno,
555 struct buffer_head *bh,
556 void *kaddr)
557{
558 struct nilfs_cpfile_header *header;
559 struct nilfs_checkpoint *cp;
560 struct nilfs_snapshot_list *list;
561
562 if (cno != 0) {
563 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, bh, kaddr);
564 list = &cp->cp_snapshot_list;
565 } else {
566 header = nilfs_cpfile_block_get_header(cpfile, bh, kaddr);
567 list = &header->ch_snapshot_list;
568 }
569 return list;
570}
571
572static int nilfs_cpfile_set_snapshot(struct inode *cpfile, __u64 cno)
573{
574 struct buffer_head *header_bh, *curr_bh, *prev_bh, *cp_bh;
575 struct nilfs_cpfile_header *header;
576 struct nilfs_checkpoint *cp;
577 struct nilfs_snapshot_list *list;
578 __u64 curr, prev;
579 unsigned long curr_blkoff, prev_blkoff;
580 void *kaddr;
581 int ret;
582
Ryusuke Konishi1f5abe72009-04-06 19:01:55 -0700583 if (cno == 0)
584 return -ENOENT; /* checkpoint number 0 is invalid */
Koji Sato29619802009-04-06 19:01:31 -0700585 down_write(&NILFS_MDT(cpfile)->mi_sem);
586
587 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &cp_bh);
588 if (ret < 0)
589 goto out_sem;
590 kaddr = kmap_atomic(cp_bh->b_page, KM_USER0);
591 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr);
592 if (nilfs_checkpoint_invalid(cp)) {
593 ret = -ENOENT;
594 kunmap_atomic(kaddr, KM_USER0);
595 goto out_cp;
596 }
597 if (nilfs_checkpoint_snapshot(cp)) {
598 ret = 0;
599 kunmap_atomic(kaddr, KM_USER0);
600 goto out_cp;
601 }
602 kunmap_atomic(kaddr, KM_USER0);
603
604 ret = nilfs_cpfile_get_header_block(cpfile, &header_bh);
605 if (ret < 0)
606 goto out_cp;
607 kaddr = kmap_atomic(header_bh->b_page, KM_USER0);
608 header = nilfs_cpfile_block_get_header(cpfile, header_bh, kaddr);
609 list = &header->ch_snapshot_list;
610 curr_bh = header_bh;
611 get_bh(curr_bh);
612 curr = 0;
613 curr_blkoff = 0;
614 prev = le64_to_cpu(list->ssl_prev);
615 while (prev > cno) {
616 prev_blkoff = nilfs_cpfile_get_blkoff(cpfile, prev);
617 curr = prev;
618 if (curr_blkoff != prev_blkoff) {
619 kunmap_atomic(kaddr, KM_USER0);
620 brelse(curr_bh);
621 ret = nilfs_cpfile_get_checkpoint_block(cpfile, curr,
622 0, &curr_bh);
623 if (ret < 0)
624 goto out_header;
625 kaddr = kmap_atomic(curr_bh->b_page, KM_USER0);
626 }
627 curr_blkoff = prev_blkoff;
628 cp = nilfs_cpfile_block_get_checkpoint(
629 cpfile, curr, curr_bh, kaddr);
630 list = &cp->cp_snapshot_list;
631 prev = le64_to_cpu(list->ssl_prev);
632 }
633 kunmap_atomic(kaddr, KM_USER0);
634
635 if (prev != 0) {
636 ret = nilfs_cpfile_get_checkpoint_block(cpfile, prev, 0,
637 &prev_bh);
638 if (ret < 0)
639 goto out_curr;
640 } else {
641 prev_bh = header_bh;
642 get_bh(prev_bh);
643 }
644
645 kaddr = kmap_atomic(curr_bh->b_page, KM_USER0);
646 list = nilfs_cpfile_block_get_snapshot_list(
647 cpfile, curr, curr_bh, kaddr);
648 list->ssl_prev = cpu_to_le64(cno);
649 kunmap_atomic(kaddr, KM_USER0);
650
651 kaddr = kmap_atomic(cp_bh->b_page, KM_USER0);
652 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr);
653 cp->cp_snapshot_list.ssl_next = cpu_to_le64(curr);
654 cp->cp_snapshot_list.ssl_prev = cpu_to_le64(prev);
655 nilfs_checkpoint_set_snapshot(cp);
656 kunmap_atomic(kaddr, KM_USER0);
657
658 kaddr = kmap_atomic(prev_bh->b_page, KM_USER0);
659 list = nilfs_cpfile_block_get_snapshot_list(
660 cpfile, prev, prev_bh, kaddr);
661 list->ssl_next = cpu_to_le64(cno);
662 kunmap_atomic(kaddr, KM_USER0);
663
664 kaddr = kmap_atomic(header_bh->b_page, KM_USER0);
665 header = nilfs_cpfile_block_get_header(cpfile, header_bh, kaddr);
666 le64_add_cpu(&header->ch_nsnapshots, 1);
667 kunmap_atomic(kaddr, KM_USER0);
668
669 nilfs_mdt_mark_buffer_dirty(prev_bh);
670 nilfs_mdt_mark_buffer_dirty(curr_bh);
671 nilfs_mdt_mark_buffer_dirty(cp_bh);
672 nilfs_mdt_mark_buffer_dirty(header_bh);
673 nilfs_mdt_mark_dirty(cpfile);
674
675 brelse(prev_bh);
676
677 out_curr:
678 brelse(curr_bh);
679
680 out_header:
681 brelse(header_bh);
682
683 out_cp:
684 brelse(cp_bh);
685
686 out_sem:
687 up_write(&NILFS_MDT(cpfile)->mi_sem);
688 return ret;
689}
690
691static int nilfs_cpfile_clear_snapshot(struct inode *cpfile, __u64 cno)
692{
693 struct buffer_head *header_bh, *next_bh, *prev_bh, *cp_bh;
694 struct nilfs_cpfile_header *header;
695 struct nilfs_checkpoint *cp;
696 struct nilfs_snapshot_list *list;
697 __u64 next, prev;
698 void *kaddr;
699 int ret;
700
Ryusuke Konishi1f5abe72009-04-06 19:01:55 -0700701 if (cno == 0)
702 return -ENOENT; /* checkpoint number 0 is invalid */
Koji Sato29619802009-04-06 19:01:31 -0700703 down_write(&NILFS_MDT(cpfile)->mi_sem);
704
705 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &cp_bh);
706 if (ret < 0)
707 goto out_sem;
708 kaddr = kmap_atomic(cp_bh->b_page, KM_USER0);
709 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr);
710 if (nilfs_checkpoint_invalid(cp)) {
711 ret = -ENOENT;
712 kunmap_atomic(kaddr, KM_USER0);
713 goto out_cp;
714 }
715 if (!nilfs_checkpoint_snapshot(cp)) {
716 ret = 0;
717 kunmap_atomic(kaddr, KM_USER0);
718 goto out_cp;
719 }
720
721 list = &cp->cp_snapshot_list;
722 next = le64_to_cpu(list->ssl_next);
723 prev = le64_to_cpu(list->ssl_prev);
724 kunmap_atomic(kaddr, KM_USER0);
725
726 ret = nilfs_cpfile_get_header_block(cpfile, &header_bh);
727 if (ret < 0)
728 goto out_cp;
729 if (next != 0) {
730 ret = nilfs_cpfile_get_checkpoint_block(cpfile, next, 0,
731 &next_bh);
732 if (ret < 0)
733 goto out_header;
734 } else {
735 next_bh = header_bh;
736 get_bh(next_bh);
737 }
738 if (prev != 0) {
739 ret = nilfs_cpfile_get_checkpoint_block(cpfile, prev, 0,
740 &prev_bh);
741 if (ret < 0)
742 goto out_next;
743 } else {
744 prev_bh = header_bh;
745 get_bh(prev_bh);
746 }
747
748 kaddr = kmap_atomic(next_bh->b_page, KM_USER0);
749 list = nilfs_cpfile_block_get_snapshot_list(
750 cpfile, next, next_bh, kaddr);
751 list->ssl_prev = cpu_to_le64(prev);
752 kunmap_atomic(kaddr, KM_USER0);
753
754 kaddr = kmap_atomic(prev_bh->b_page, KM_USER0);
755 list = nilfs_cpfile_block_get_snapshot_list(
756 cpfile, prev, prev_bh, kaddr);
757 list->ssl_next = cpu_to_le64(next);
758 kunmap_atomic(kaddr, KM_USER0);
759
760 kaddr = kmap_atomic(cp_bh->b_page, KM_USER0);
761 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, cp_bh, kaddr);
762 cp->cp_snapshot_list.ssl_next = cpu_to_le64(0);
763 cp->cp_snapshot_list.ssl_prev = cpu_to_le64(0);
764 nilfs_checkpoint_clear_snapshot(cp);
765 kunmap_atomic(kaddr, KM_USER0);
766
767 kaddr = kmap_atomic(header_bh->b_page, KM_USER0);
768 header = nilfs_cpfile_block_get_header(cpfile, header_bh, kaddr);
769 le64_add_cpu(&header->ch_nsnapshots, -1);
770 kunmap_atomic(kaddr, KM_USER0);
771
772 nilfs_mdt_mark_buffer_dirty(next_bh);
773 nilfs_mdt_mark_buffer_dirty(prev_bh);
774 nilfs_mdt_mark_buffer_dirty(cp_bh);
775 nilfs_mdt_mark_buffer_dirty(header_bh);
776 nilfs_mdt_mark_dirty(cpfile);
777
778 brelse(prev_bh);
779
780 out_next:
781 brelse(next_bh);
782
783 out_header:
784 brelse(header_bh);
785
786 out_cp:
787 brelse(cp_bh);
788
789 out_sem:
790 up_write(&NILFS_MDT(cpfile)->mi_sem);
791 return ret;
792}
793
794/**
795 * nilfs_cpfile_is_snapshot -
796 * @cpfile: inode of checkpoint file
797 * @cno: checkpoint number
798 *
799 * Description:
800 *
801 * Return Value: On success, 1 is returned if the checkpoint specified by
802 * @cno is a snapshot, or 0 if not. On error, one of the following negative
803 * error codes is returned.
804 *
805 * %-EIO - I/O error.
806 *
807 * %-ENOMEM - Insufficient amount of memory available.
808 *
809 * %-ENOENT - No such checkpoint.
810 */
811int nilfs_cpfile_is_snapshot(struct inode *cpfile, __u64 cno)
812{
813 struct buffer_head *bh;
814 struct nilfs_checkpoint *cp;
815 void *kaddr;
816 int ret;
817
Zhu Yanhai43be0ec2009-08-12 14:17:59 +0800818 /* CP number is invalid if it's zero or larger than the
819 largest exist one.*/
820 if (cno == 0 || cno >= nilfs_mdt_cno(cpfile))
821 return -ENOENT;
Koji Sato29619802009-04-06 19:01:31 -0700822 down_read(&NILFS_MDT(cpfile)->mi_sem);
823
824 ret = nilfs_cpfile_get_checkpoint_block(cpfile, cno, 0, &bh);
825 if (ret < 0)
826 goto out;
827 kaddr = kmap_atomic(bh->b_page, KM_USER0);
828 cp = nilfs_cpfile_block_get_checkpoint(cpfile, cno, bh, kaddr);
Zhu Yanhai43be0ec2009-08-12 14:17:59 +0800829 if (nilfs_checkpoint_invalid(cp))
830 ret = -ENOENT;
831 else
832 ret = nilfs_checkpoint_snapshot(cp);
Koji Sato29619802009-04-06 19:01:31 -0700833 kunmap_atomic(kaddr, KM_USER0);
834 brelse(bh);
835
836 out:
837 up_read(&NILFS_MDT(cpfile)->mi_sem);
838 return ret;
839}
840
841/**
842 * nilfs_cpfile_change_cpmode - change checkpoint mode
843 * @cpfile: inode of checkpoint file
844 * @cno: checkpoint number
845 * @status: mode of checkpoint
846 *
847 * Description: nilfs_change_cpmode() changes the mode of the checkpoint
848 * specified by @cno. The mode @mode is NILFS_CHECKPOINT or NILFS_SNAPSHOT.
849 *
850 * Return Value: On success, 0 is returned. On error, one of the following
851 * negative error codes is returned.
852 *
853 * %-EIO - I/O error.
854 *
855 * %-ENOMEM - Insufficient amount of memory available.
856 *
857 * %-ENOENT - No such checkpoint.
858 */
859int nilfs_cpfile_change_cpmode(struct inode *cpfile, __u64 cno, int mode)
860{
861 struct the_nilfs *nilfs;
862 int ret;
863
864 nilfs = NILFS_MDT(cpfile)->mi_nilfs;
865
866 switch (mode) {
867 case NILFS_CHECKPOINT:
868 /*
869 * Check for protecting existing snapshot mounts:
Ryusuke Konishiaa7dfb82009-06-08 01:39:33 +0900870 * ns_mount_mutex is used to make this operation atomic and
Koji Sato29619802009-04-06 19:01:31 -0700871 * exclusive with a new mount job. Though it doesn't cover
872 * umount, it's enough for the purpose.
873 */
Ryusuke Konishiaa7dfb82009-06-08 01:39:33 +0900874 mutex_lock(&nilfs->ns_mount_mutex);
Koji Sato29619802009-04-06 19:01:31 -0700875 if (nilfs_checkpoint_is_mounted(nilfs, cno, 1)) {
876 /* Current implementation does not have to protect
877 plain read-only mounts since they are exclusive
878 with a read/write mount and are protected from the
879 cleaner. */
880 ret = -EBUSY;
881 } else
882 ret = nilfs_cpfile_clear_snapshot(cpfile, cno);
Ryusuke Konishiaa7dfb82009-06-08 01:39:33 +0900883 mutex_unlock(&nilfs->ns_mount_mutex);
Koji Sato29619802009-04-06 19:01:31 -0700884 return ret;
885 case NILFS_SNAPSHOT:
886 return nilfs_cpfile_set_snapshot(cpfile, cno);
887 default:
888 return -EINVAL;
889 }
890}
891
892/**
893 * nilfs_cpfile_get_stat - get checkpoint statistics
894 * @cpfile: inode of checkpoint file
895 * @stat: pointer to a structure of checkpoint statistics
896 *
897 * Description: nilfs_cpfile_get_stat() returns information about checkpoints.
898 *
899 * Return Value: On success, 0 is returned, and checkpoints information is
900 * stored in the place pointed by @stat. On error, one of the following
901 * negative error codes is returned.
902 *
903 * %-EIO - I/O error.
904 *
905 * %-ENOMEM - Insufficient amount of memory available.
906 */
907int nilfs_cpfile_get_stat(struct inode *cpfile, struct nilfs_cpstat *cpstat)
908{
909 struct buffer_head *bh;
910 struct nilfs_cpfile_header *header;
911 void *kaddr;
912 int ret;
913
914 down_read(&NILFS_MDT(cpfile)->mi_sem);
915
916 ret = nilfs_cpfile_get_header_block(cpfile, &bh);
917 if (ret < 0)
918 goto out_sem;
919 kaddr = kmap_atomic(bh->b_page, KM_USER0);
920 header = nilfs_cpfile_block_get_header(cpfile, bh, kaddr);
921 cpstat->cs_cno = nilfs_mdt_cno(cpfile);
922 cpstat->cs_ncps = le64_to_cpu(header->ch_ncheckpoints);
923 cpstat->cs_nsss = le64_to_cpu(header->ch_nsnapshots);
924 kunmap_atomic(kaddr, KM_USER0);
925 brelse(bh);
926
927 out_sem:
928 up_read(&NILFS_MDT(cpfile)->mi_sem);
929 return ret;
930}