blob: 1403c88f2b053fd521598360007b090f805478c0 [file] [log] [blame]
Mark Fashehccd979b2005-12-15 14:31:24 -08001/* -*- mode: c; c-basic-offset: 8; -*-
2 * vim: noexpandtab sw=8 ts=8 sts=0:
3 *
4 * io.c
5 *
6 * Buffer cache handling
7 *
8 * Copyright (C) 2002, 2004 Oracle. All rights reserved.
9 *
10 * This program is free software; you can redistribute it and/or
11 * modify it under the terms of the GNU General Public
12 * License as published by the Free Software Foundation; either
13 * version 2 of the License, or (at your option) any later version.
14 *
15 * This program is distributed in the hope that it will be useful,
16 * but WITHOUT ANY WARRANTY; without even the implied warranty of
17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
18 * General Public License for more details.
19 *
20 * You should have received a copy of the GNU General Public
21 * License along with this program; if not, write to the
22 * Free Software Foundation, Inc., 59 Temple Place - Suite 330,
23 * Boston, MA 021110-1307, USA.
24 */
25
26#include <linux/fs.h>
27#include <linux/types.h>
Mark Fashehccd979b2005-12-15 14:31:24 -080028#include <linux/highmem.h>
29
30#include <cluster/masklog.h>
31
32#include "ocfs2.h"
33
34#include "alloc.h"
35#include "inode.h"
36#include "journal.h"
37#include "uptodate.h"
Mark Fashehccd979b2005-12-15 14:31:24 -080038#include "buffer_head_io.h"
Tao Ma15057e92011-02-24 16:09:38 +080039#include "ocfs2_trace.h"
Mark Fashehccd979b2005-12-15 14:31:24 -080040
Joel Becker970e4932008-11-13 14:49:19 -080041/*
42 * Bits on bh->b_state used by ocfs2.
43 *
Mark Fashehb86c86f2008-11-18 17:16:47 -080044 * These MUST be after the JBD2 bits. Hence, we use BH_JBDPrivateStart.
Joel Becker970e4932008-11-13 14:49:19 -080045 */
46enum ocfs2_state_bits {
Mark Fashehb86c86f2008-11-18 17:16:47 -080047 BH_NeedsValidate = BH_JBDPrivateStart,
Joel Becker970e4932008-11-13 14:49:19 -080048};
49
50/* Expand the magic b_state functions */
51BUFFER_FNS(NeedsValidate, needs_validate);
52
Mark Fashehccd979b2005-12-15 14:31:24 -080053int ocfs2_write_block(struct ocfs2_super *osb, struct buffer_head *bh,
Joel Becker8cb471e2009-02-10 20:00:41 -080054 struct ocfs2_caching_info *ci)
Mark Fashehccd979b2005-12-15 14:31:24 -080055{
56 int ret = 0;
57
Tao Ma15057e92011-02-24 16:09:38 +080058 trace_ocfs2_write_block((unsigned long long)bh->b_blocknr, ci);
Mark Fashehccd979b2005-12-15 14:31:24 -080059
60 BUG_ON(bh->b_blocknr < OCFS2_SUPER_BLOCK_BLKNO);
61 BUG_ON(buffer_jbd(bh));
62
63 /* No need to check for a soft readonly file system here. non
64 * journalled writes are only ever done on system files which
65 * can get modified during recovery even if read-only. */
66 if (ocfs2_is_hard_readonly(osb)) {
67 ret = -EROFS;
Tao Mac1e8d352011-03-07 16:43:21 +080068 mlog_errno(ret);
Mark Fashehccd979b2005-12-15 14:31:24 -080069 goto out;
70 }
71
Joel Becker8cb471e2009-02-10 20:00:41 -080072 ocfs2_metadata_cache_io_lock(ci);
Mark Fashehccd979b2005-12-15 14:31:24 -080073
74 lock_buffer(bh);
75 set_buffer_uptodate(bh);
76
77 /* remove from dirty list before I/O. */
78 clear_buffer_dirty(bh);
79
Joel Beckerda1e9092008-10-09 17:20:29 -070080 get_bh(bh); /* for end_buffer_write_sync() */
Mark Fashehccd979b2005-12-15 14:31:24 -080081 bh->b_end_io = end_buffer_write_sync;
Mike Christie2a222ca2016-06-05 14:31:43 -050082 submit_bh(REQ_OP_WRITE, 0, bh);
Mark Fashehccd979b2005-12-15 14:31:24 -080083
84 wait_on_buffer(bh);
85
86 if (buffer_uptodate(bh)) {
Joel Becker8cb471e2009-02-10 20:00:41 -080087 ocfs2_set_buffer_uptodate(ci, bh);
Mark Fashehccd979b2005-12-15 14:31:24 -080088 } else {
89 /* We don't need to remove the clustered uptodate
90 * information for this bh as it's not marked locally
91 * uptodate. */
92 ret = -EIO;
Tao Mac1e8d352011-03-07 16:43:21 +080093 mlog_errno(ret);
Mark Fashehccd979b2005-12-15 14:31:24 -080094 }
95
Joel Becker8cb471e2009-02-10 20:00:41 -080096 ocfs2_metadata_cache_io_unlock(ci);
Mark Fashehccd979b2005-12-15 14:31:24 -080097out:
Mark Fashehccd979b2005-12-15 14:31:24 -080098 return ret;
99}
100
Changwei Ge65cbd122018-11-02 15:48:19 -0700101/* Caller must provide a bhs[] with all NULL or non-NULL entries, so it
102 * will be easier to handle read failure.
103 */
Joel Beckerda1e9092008-10-09 17:20:29 -0700104int ocfs2_read_blocks_sync(struct ocfs2_super *osb, u64 block,
105 unsigned int nr, struct buffer_head *bhs[])
106{
107 int status = 0;
108 unsigned int i;
109 struct buffer_head *bh;
Changwei Ge65cbd122018-11-02 15:48:19 -0700110 int new_bh = 0;
Joel Beckerda1e9092008-10-09 17:20:29 -0700111
Tao Ma15057e92011-02-24 16:09:38 +0800112 trace_ocfs2_read_blocks_sync((unsigned long long)block, nr);
113
114 if (!nr)
Joel Beckerda1e9092008-10-09 17:20:29 -0700115 goto bail;
Joel Beckerda1e9092008-10-09 17:20:29 -0700116
Changwei Ge65cbd122018-11-02 15:48:19 -0700117 /* Don't put buffer head and re-assign it to NULL if it is allocated
118 * outside since the caller can't be aware of this alternation!
119 */
120 new_bh = (bhs[0] == NULL);
121
Joel Beckerda1e9092008-10-09 17:20:29 -0700122 for (i = 0 ; i < nr ; i++) {
123 if (bhs[i] == NULL) {
124 bhs[i] = sb_getblk(osb->sb, block++);
125 if (bhs[i] == NULL) {
Rui Xiang7391a292013-11-12 15:06:54 -0800126 status = -ENOMEM;
Joel Beckerda1e9092008-10-09 17:20:29 -0700127 mlog_errno(status);
Changwei Ge65cbd122018-11-02 15:48:19 -0700128 break;
Joel Beckerda1e9092008-10-09 17:20:29 -0700129 }
130 }
131 bh = bhs[i];
132
133 if (buffer_jbd(bh)) {
Tao Ma15057e92011-02-24 16:09:38 +0800134 trace_ocfs2_read_blocks_sync_jbd(
135 (unsigned long long)bh->b_blocknr);
Joel Beckerda1e9092008-10-09 17:20:29 -0700136 continue;
137 }
138
139 if (buffer_dirty(bh)) {
140 /* This should probably be a BUG, or
141 * at least return an error. */
142 mlog(ML_ERROR,
143 "trying to sync read a dirty "
144 "buffer! (blocknr = %llu), skipping\n",
145 (unsigned long long)bh->b_blocknr);
146 continue;
147 }
148
149 lock_buffer(bh);
150 if (buffer_jbd(bh)) {
Gang He7186ee02016-06-24 14:50:13 -0700151#ifdef CATCH_BH_JBD_RACES
Joel Beckerda1e9092008-10-09 17:20:29 -0700152 mlog(ML_ERROR,
153 "block %llu had the JBD bit set "
154 "while I was in lock_buffer!",
155 (unsigned long long)bh->b_blocknr);
156 BUG();
Gang He7186ee02016-06-24 14:50:13 -0700157#else
158 unlock_buffer(bh);
159 continue;
160#endif
Joel Beckerda1e9092008-10-09 17:20:29 -0700161 }
162
Joel Beckerda1e9092008-10-09 17:20:29 -0700163 get_bh(bh); /* for end_buffer_read_sync() */
164 bh->b_end_io = end_buffer_read_sync;
Mike Christie2a222ca2016-06-05 14:31:43 -0500165 submit_bh(REQ_OP_READ, 0, bh);
Joel Beckerda1e9092008-10-09 17:20:29 -0700166 }
167
Changwei Ge65cbd122018-11-02 15:48:19 -0700168read_failure:
Joel Beckerda1e9092008-10-09 17:20:29 -0700169 for (i = nr; i > 0; i--) {
170 bh = bhs[i - 1];
171
Changwei Ge65cbd122018-11-02 15:48:19 -0700172 if (unlikely(status)) {
173 if (new_bh && bh) {
174 /* If middle bh fails, let previous bh
175 * finish its read and then put it to
176 * aovoid bh leak
177 */
178 if (!buffer_jbd(bh))
179 wait_on_buffer(bh);
180 put_bh(bh);
181 bhs[i - 1] = NULL;
182 } else if (bh && buffer_uptodate(bh)) {
183 clear_buffer_uptodate(bh);
184 }
185 continue;
186 }
187
Mark Fashehd6b58f82008-11-21 14:06:55 -0800188 /* No need to wait on the buffer if it's managed by JBD. */
189 if (!buffer_jbd(bh))
190 wait_on_buffer(bh);
Joel Beckerda1e9092008-10-09 17:20:29 -0700191
Joel Beckerda1e9092008-10-09 17:20:29 -0700192 if (!buffer_uptodate(bh)) {
193 /* Status won't be cleared from here on out,
194 * so we can safely record this and loop back
195 * to cleanup the other buffers. */
196 status = -EIO;
Changwei Ge65cbd122018-11-02 15:48:19 -0700197 goto read_failure;
Joel Beckerda1e9092008-10-09 17:20:29 -0700198 }
199 }
200
201bail:
202 return status;
203}
204
Changwei Ge65cbd122018-11-02 15:48:19 -0700205/* Caller must provide a bhs[] with all NULL or non-NULL entries, so it
206 * will be easier to handle read failure.
207 */
Joel Becker8cb471e2009-02-10 20:00:41 -0800208int ocfs2_read_blocks(struct ocfs2_caching_info *ci, u64 block, int nr,
Joel Becker970e4932008-11-13 14:49:19 -0800209 struct buffer_head *bhs[], int flags,
210 int (*validate)(struct super_block *sb,
211 struct buffer_head *bh))
Mark Fashehccd979b2005-12-15 14:31:24 -0800212{
213 int status = 0;
Mark Fashehccd979b2005-12-15 14:31:24 -0800214 int i, ignore_cache = 0;
215 struct buffer_head *bh;
Joel Becker8cb471e2009-02-10 20:00:41 -0800216 struct super_block *sb = ocfs2_metadata_cache_get_super(ci);
Changwei Ge65cbd122018-11-02 15:48:19 -0700217 int new_bh = 0;
Mark Fashehccd979b2005-12-15 14:31:24 -0800218
Tao Ma15057e92011-02-24 16:09:38 +0800219 trace_ocfs2_read_blocks_begin(ci, (unsigned long long)block, nr, flags);
Mark Fashehccd979b2005-12-15 14:31:24 -0800220
Joel Becker8cb471e2009-02-10 20:00:41 -0800221 BUG_ON(!ci);
Joel Beckerd4a8c932008-10-09 17:20:34 -0700222 BUG_ON((flags & OCFS2_BH_READAHEAD) &&
223 (flags & OCFS2_BH_IGNORE_CACHE));
Mark Fashehaa958872006-04-21 13:49:02 -0700224
Joel Becker31d33072008-10-09 17:20:30 -0700225 if (bhs == NULL) {
Mark Fashehccd979b2005-12-15 14:31:24 -0800226 status = -EINVAL;
227 mlog_errno(status);
228 goto bail;
229 }
230
231 if (nr < 0) {
232 mlog(ML_ERROR, "asked to read %d blocks!\n", nr);
233 status = -EINVAL;
234 mlog_errno(status);
235 goto bail;
236 }
237
238 if (nr == 0) {
Mark Fashehccd979b2005-12-15 14:31:24 -0800239 status = 0;
240 goto bail;
241 }
242
Changwei Ge65cbd122018-11-02 15:48:19 -0700243 /* Don't put buffer head and re-assign it to NULL if it is allocated
244 * outside since the caller can't be aware of this alternation!
245 */
246 new_bh = (bhs[0] == NULL);
247
Joel Becker8cb471e2009-02-10 20:00:41 -0800248 ocfs2_metadata_cache_io_lock(ci);
Mark Fashehccd979b2005-12-15 14:31:24 -0800249 for (i = 0 ; i < nr ; i++) {
250 if (bhs[i] == NULL) {
Joel Becker8cb471e2009-02-10 20:00:41 -0800251 bhs[i] = sb_getblk(sb, block++);
Mark Fashehccd979b2005-12-15 14:31:24 -0800252 if (bhs[i] == NULL) {
Joel Becker8cb471e2009-02-10 20:00:41 -0800253 ocfs2_metadata_cache_io_unlock(ci);
Rui Xiang7391a292013-11-12 15:06:54 -0800254 status = -ENOMEM;
Mark Fashehccd979b2005-12-15 14:31:24 -0800255 mlog_errno(status);
Changwei Ge65cbd122018-11-02 15:48:19 -0700256 /* Don't forget to put previous bh! */
257 break;
Mark Fashehccd979b2005-12-15 14:31:24 -0800258 }
259 }
260 bh = bhs[i];
Joel Beckerd4a8c932008-10-09 17:20:34 -0700261 ignore_cache = (flags & OCFS2_BH_IGNORE_CACHE);
Mark Fashehccd979b2005-12-15 14:31:24 -0800262
Mark Fashehaa958872006-04-21 13:49:02 -0700263 /* There are three read-ahead cases here which we need to
264 * be concerned with. All three assume a buffer has
265 * previously been submitted with OCFS2_BH_READAHEAD
266 * and it hasn't yet completed I/O.
267 *
268 * 1) The current request is sync to disk. This rarely
269 * happens these days, and never when performance
270 * matters - the code can just wait on the buffer
271 * lock and re-submit.
272 *
273 * 2) The current request is cached, but not
274 * readahead. ocfs2_buffer_uptodate() will return
275 * false anyway, so we'll wind up waiting on the
276 * buffer lock to do I/O. We re-check the request
277 * with after getting the lock to avoid a re-submit.
278 *
279 * 3) The current request is readahead (and so must
280 * also be a caching one). We short circuit if the
281 * buffer is locked (under I/O) and if it's in the
282 * uptodate cache. The re-check from #2 catches the
283 * case that the previous read-ahead completes just
284 * before our is-it-in-flight check.
285 */
286
Joel Becker8cb471e2009-02-10 20:00:41 -0800287 if (!ignore_cache && !ocfs2_buffer_uptodate(ci, bh)) {
Tao Mad7014852011-02-24 16:22:20 +0800288 trace_ocfs2_read_blocks_from_disk(
Mark Fashehccd979b2005-12-15 14:31:24 -0800289 (unsigned long long)bh->b_blocknr,
Joel Becker8cb471e2009-02-10 20:00:41 -0800290 (unsigned long long)ocfs2_metadata_cache_owner(ci));
Joel Beckerd4a8c932008-10-09 17:20:34 -0700291 /* We're using ignore_cache here to say
292 * "go to disk" */
Mark Fashehccd979b2005-12-15 14:31:24 -0800293 ignore_cache = 1;
294 }
295
Tao Ma15057e92011-02-24 16:09:38 +0800296 trace_ocfs2_read_blocks_bh((unsigned long long)bh->b_blocknr,
297 ignore_cache, buffer_jbd(bh), buffer_dirty(bh));
298
Mark Fashehccd979b2005-12-15 14:31:24 -0800299 if (buffer_jbd(bh)) {
Mark Fashehccd979b2005-12-15 14:31:24 -0800300 continue;
301 }
302
Joel Beckerd4a8c932008-10-09 17:20:34 -0700303 if (ignore_cache) {
Mark Fashehccd979b2005-12-15 14:31:24 -0800304 if (buffer_dirty(bh)) {
305 /* This should probably be a BUG, or
306 * at least return an error. */
Mark Fashehccd979b2005-12-15 14:31:24 -0800307 continue;
308 }
309
Mark Fashehaa958872006-04-21 13:49:02 -0700310 /* A read-ahead request was made - if the
311 * buffer is already under read-ahead from a
312 * previously submitted request than we are
313 * done here. */
314 if ((flags & OCFS2_BH_READAHEAD)
Joel Becker8cb471e2009-02-10 20:00:41 -0800315 && ocfs2_buffer_read_ahead(ci, bh))
Mark Fashehaa958872006-04-21 13:49:02 -0700316 continue;
317
Mark Fashehccd979b2005-12-15 14:31:24 -0800318 lock_buffer(bh);
319 if (buffer_jbd(bh)) {
320#ifdef CATCH_BH_JBD_RACES
321 mlog(ML_ERROR, "block %llu had the JBD bit set "
322 "while I was in lock_buffer!",
323 (unsigned long long)bh->b_blocknr);
324 BUG();
325#else
326 unlock_buffer(bh);
327 continue;
328#endif
329 }
Mark Fashehaa958872006-04-21 13:49:02 -0700330
331 /* Re-check ocfs2_buffer_uptodate() as a
332 * previously read-ahead buffer may have
333 * completed I/O while we were waiting for the
334 * buffer lock. */
Joel Beckerd4a8c932008-10-09 17:20:34 -0700335 if (!(flags & OCFS2_BH_IGNORE_CACHE)
Mark Fashehaa958872006-04-21 13:49:02 -0700336 && !(flags & OCFS2_BH_READAHEAD)
Joel Becker8cb471e2009-02-10 20:00:41 -0800337 && ocfs2_buffer_uptodate(ci, bh)) {
Mark Fashehaa958872006-04-21 13:49:02 -0700338 unlock_buffer(bh);
339 continue;
340 }
341
Mark Fashehccd979b2005-12-15 14:31:24 -0800342 get_bh(bh); /* for end_buffer_read_sync() */
Joel Becker970e4932008-11-13 14:49:19 -0800343 if (validate)
344 set_buffer_needs_validate(bh);
Mark Fashehccd979b2005-12-15 14:31:24 -0800345 bh->b_end_io = end_buffer_read_sync;
Mike Christie2a222ca2016-06-05 14:31:43 -0500346 submit_bh(REQ_OP_READ, 0, bh);
Mark Fashehccd979b2005-12-15 14:31:24 -0800347 continue;
348 }
349 }
350
Changwei Ge65cbd122018-11-02 15:48:19 -0700351read_failure:
Mark Fashehccd979b2005-12-15 14:31:24 -0800352 for (i = (nr - 1); i >= 0; i--) {
353 bh = bhs[i];
354
Mark Fashehaa958872006-04-21 13:49:02 -0700355 if (!(flags & OCFS2_BH_READAHEAD)) {
Changwei Ge65cbd122018-11-02 15:48:19 -0700356 if (unlikely(status)) {
357 /* Clear the buffers on error including those
358 * ever succeeded in reading
359 */
360 if (new_bh && bh) {
361 /* If middle bh fails, let previous bh
362 * finish its read and then put it to
363 * aovoid bh leak
364 */
365 if (!buffer_jbd(bh))
366 wait_on_buffer(bh);
367 put_bh(bh);
368 bhs[i] = NULL;
369 } else if (bh && buffer_uptodate(bh)) {
370 clear_buffer_uptodate(bh);
371 }
Goldwyn Rodrigues34237682015-09-04 15:44:20 -0700372 continue;
373 }
Mark Fashehaa958872006-04-21 13:49:02 -0700374 /* We know this can't have changed as we hold the
Joel Becker8cb471e2009-02-10 20:00:41 -0800375 * owner sem. Avoid doing any work on the bh if the
Mark Fashehaa958872006-04-21 13:49:02 -0700376 * journal has it. */
377 if (!buffer_jbd(bh))
378 wait_on_buffer(bh);
Mark Fashehccd979b2005-12-15 14:31:24 -0800379
Mark Fashehaa958872006-04-21 13:49:02 -0700380 if (!buffer_uptodate(bh)) {
381 /* Status won't be cleared from here on out,
382 * so we can safely record this and loop back
383 * to cleanup the other buffers. Don't need to
384 * remove the clustered uptodate information
385 * for this bh as it's not marked locally
386 * uptodate. */
387 status = -EIO;
Junxiao Bi44ae7182018-09-20 12:22:51 -0700388 clear_buffer_needs_validate(bh);
Changwei Ge65cbd122018-11-02 15:48:19 -0700389 goto read_failure;
Mark Fashehaa958872006-04-21 13:49:02 -0700390 }
Joel Becker970e4932008-11-13 14:49:19 -0800391
392 if (buffer_needs_validate(bh)) {
393 /* We never set NeedsValidate if the
394 * buffer was held by the journal, so
395 * that better not have changed */
396 BUG_ON(buffer_jbd(bh));
397 clear_buffer_needs_validate(bh);
Joel Becker8cb471e2009-02-10 20:00:41 -0800398 status = validate(sb, bh);
Changwei Ge65cbd122018-11-02 15:48:19 -0700399 if (status)
400 goto read_failure;
Joel Becker970e4932008-11-13 14:49:19 -0800401 }
Mark Fashehccd979b2005-12-15 14:31:24 -0800402 }
403
Mark Fashehaa958872006-04-21 13:49:02 -0700404 /* Always set the buffer in the cache, even if it was
405 * a forced read, or read-ahead which hasn't yet
406 * completed. */
Joel Becker8cb471e2009-02-10 20:00:41 -0800407 ocfs2_set_buffer_uptodate(ci, bh);
Mark Fashehccd979b2005-12-15 14:31:24 -0800408 }
Joel Becker8cb471e2009-02-10 20:00:41 -0800409 ocfs2_metadata_cache_io_unlock(ci);
Mark Fashehccd979b2005-12-15 14:31:24 -0800410
Tao Ma15057e92011-02-24 16:09:38 +0800411 trace_ocfs2_read_blocks_end((unsigned long long)block, nr,
412 flags, ignore_cache);
Mark Fashehccd979b2005-12-15 14:31:24 -0800413
414bail:
415
Mark Fashehccd979b2005-12-15 14:31:24 -0800416 return status;
417}
Tao Mad6590722007-12-18 15:47:03 +0800418
419/* Check whether the blkno is the super block or one of the backups. */
420static void ocfs2_check_super_or_backup(struct super_block *sb,
421 sector_t blkno)
422{
423 int i;
424 u64 backup_blkno;
425
426 if (blkno == OCFS2_SUPER_BLOCK_BLKNO)
427 return;
428
429 for (i = 0; i < OCFS2_MAX_BACKUP_SUPERBLOCKS; i++) {
430 backup_blkno = ocfs2_backup_super_blkno(sb, i);
431 if (backup_blkno == blkno)
432 return;
433 }
434
435 BUG();
436}
437
438/*
439 * Write super block and backups doesn't need to collaborate with journal,
Joel Becker8cb471e2009-02-10 20:00:41 -0800440 * so we don't need to lock ip_io_mutex and ci doesn't need to bea passed
Tao Mad6590722007-12-18 15:47:03 +0800441 * into this function.
442 */
443int ocfs2_write_super_or_backup(struct ocfs2_super *osb,
444 struct buffer_head *bh)
445{
446 int ret = 0;
Joel Beckera42ab8e2010-03-31 18:25:44 -0700447 struct ocfs2_dinode *di = (struct ocfs2_dinode *)bh->b_data;
Tao Mad6590722007-12-18 15:47:03 +0800448
Tao Mad6590722007-12-18 15:47:03 +0800449 BUG_ON(buffer_jbd(bh));
450 ocfs2_check_super_or_backup(osb->sb, bh->b_blocknr);
451
452 if (ocfs2_is_hard_readonly(osb) || ocfs2_is_soft_readonly(osb)) {
453 ret = -EROFS;
Tao Mac1e8d352011-03-07 16:43:21 +0800454 mlog_errno(ret);
Tao Mad6590722007-12-18 15:47:03 +0800455 goto out;
456 }
457
458 lock_buffer(bh);
459 set_buffer_uptodate(bh);
460
461 /* remove from dirty list before I/O. */
462 clear_buffer_dirty(bh);
463
464 get_bh(bh); /* for end_buffer_write_sync() */
465 bh->b_end_io = end_buffer_write_sync;
Joel Beckera42ab8e2010-03-31 18:25:44 -0700466 ocfs2_compute_meta_ecc(osb->sb, bh->b_data, &di->i_check);
Mike Christie2a222ca2016-06-05 14:31:43 -0500467 submit_bh(REQ_OP_WRITE, 0, bh);
Tao Mad6590722007-12-18 15:47:03 +0800468
469 wait_on_buffer(bh);
470
471 if (!buffer_uptodate(bh)) {
472 ret = -EIO;
Tao Mac1e8d352011-03-07 16:43:21 +0800473 mlog_errno(ret);
Tao Mad6590722007-12-18 15:47:03 +0800474 }
475
476out:
Tao Mad6590722007-12-18 15:47:03 +0800477 return ret;
478}