Ryusuke Konishi | 5442680 | 2009-04-06 19:01:29 -0700 | [diff] [blame] | 1 | /* |
| 2 | * alloc.c - NILFS dat/inode allocator |
| 3 | * |
| 4 | * Copyright (C) 2006-2008 Nippon Telegraph and Telephone Corporation. |
| 5 | * |
| 6 | * This program is free software; you can redistribute it and/or modify |
| 7 | * it under the terms of the GNU General Public License as published by |
| 8 | * the Free Software Foundation; either version 2 of the License, or |
| 9 | * (at your option) any later version. |
| 10 | * |
| 11 | * This program is distributed in the hope that it will be useful, |
| 12 | * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| 13 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| 14 | * GNU General Public License for more details. |
| 15 | * |
| 16 | * You should have received a copy of the GNU General Public License |
| 17 | * along with this program; if not, write to the Free Software |
| 18 | * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA |
| 19 | * |
| 20 | * Original code was written by Koji Sato <koji@osrg.net>. |
| 21 | * Two allocators were unified by Ryusuke Konishi <ryusuke@osrg.net>, |
| 22 | * Amagai Yoshiji <amagai@osrg.net>. |
| 23 | */ |
| 24 | |
| 25 | #include <linux/types.h> |
| 26 | #include <linux/buffer_head.h> |
| 27 | #include <linux/fs.h> |
| 28 | #include <linux/bitops.h> |
Tejun Heo | 5a0e3ad | 2010-03-24 17:04:11 +0900 | [diff] [blame] | 29 | #include <linux/slab.h> |
Ryusuke Konishi | 5442680 | 2009-04-06 19:01:29 -0700 | [diff] [blame] | 30 | #include "mdt.h" |
| 31 | #include "alloc.h" |
| 32 | |
| 33 | |
| 34 | static inline unsigned long |
| 35 | nilfs_palloc_groups_per_desc_block(const struct inode *inode) |
| 36 | { |
| 37 | return (1UL << inode->i_blkbits) / |
| 38 | sizeof(struct nilfs_palloc_group_desc); |
| 39 | } |
| 40 | |
| 41 | static inline unsigned long |
| 42 | nilfs_palloc_groups_count(const struct inode *inode) |
| 43 | { |
| 44 | return 1UL << (BITS_PER_LONG - (inode->i_blkbits + 3 /* log2(8) */)); |
| 45 | } |
| 46 | |
| 47 | int nilfs_palloc_init_blockgroup(struct inode *inode, unsigned entry_size) |
| 48 | { |
| 49 | struct nilfs_mdt_info *mi = NILFS_MDT(inode); |
| 50 | |
| 51 | mi->mi_bgl = kmalloc(sizeof(*mi->mi_bgl), GFP_NOFS); |
| 52 | if (!mi->mi_bgl) |
| 53 | return -ENOMEM; |
| 54 | |
| 55 | bgl_lock_init(mi->mi_bgl); |
| 56 | |
| 57 | nilfs_mdt_set_entry_size(inode, entry_size, 0); |
| 58 | |
| 59 | mi->mi_blocks_per_group = |
| 60 | DIV_ROUND_UP(nilfs_palloc_entries_per_group(inode), |
| 61 | mi->mi_entries_per_block) + 1; |
| 62 | /* Number of blocks in a group including entry blocks and |
| 63 | a bitmap block */ |
| 64 | mi->mi_blocks_per_desc_block = |
| 65 | nilfs_palloc_groups_per_desc_block(inode) * |
| 66 | mi->mi_blocks_per_group + 1; |
| 67 | /* Number of blocks per descriptor including the |
| 68 | descriptor block */ |
| 69 | return 0; |
| 70 | } |
| 71 | |
| 72 | static unsigned long nilfs_palloc_group(const struct inode *inode, __u64 nr, |
| 73 | unsigned long *offset) |
| 74 | { |
| 75 | __u64 group = nr; |
| 76 | |
| 77 | *offset = do_div(group, nilfs_palloc_entries_per_group(inode)); |
| 78 | return group; |
| 79 | } |
| 80 | |
| 81 | static unsigned long |
| 82 | nilfs_palloc_desc_blkoff(const struct inode *inode, unsigned long group) |
| 83 | { |
| 84 | unsigned long desc_block = |
| 85 | group / nilfs_palloc_groups_per_desc_block(inode); |
| 86 | return desc_block * NILFS_MDT(inode)->mi_blocks_per_desc_block; |
| 87 | } |
| 88 | |
| 89 | static unsigned long |
| 90 | nilfs_palloc_bitmap_blkoff(const struct inode *inode, unsigned long group) |
| 91 | { |
| 92 | unsigned long desc_offset = |
| 93 | group % nilfs_palloc_groups_per_desc_block(inode); |
| 94 | return nilfs_palloc_desc_blkoff(inode, group) + 1 + |
| 95 | desc_offset * NILFS_MDT(inode)->mi_blocks_per_group; |
| 96 | } |
| 97 | |
| 98 | static unsigned long |
| 99 | nilfs_palloc_group_desc_nfrees(struct inode *inode, unsigned long group, |
| 100 | const struct nilfs_palloc_group_desc *desc) |
| 101 | { |
| 102 | unsigned long nfree; |
| 103 | |
| 104 | spin_lock(nilfs_mdt_bgl_lock(inode, group)); |
| 105 | nfree = le32_to_cpu(desc->pg_nfrees); |
| 106 | spin_unlock(nilfs_mdt_bgl_lock(inode, group)); |
| 107 | return nfree; |
| 108 | } |
| 109 | |
| 110 | static void |
| 111 | nilfs_palloc_group_desc_add_entries(struct inode *inode, |
| 112 | unsigned long group, |
| 113 | struct nilfs_palloc_group_desc *desc, |
| 114 | u32 n) |
| 115 | { |
| 116 | spin_lock(nilfs_mdt_bgl_lock(inode, group)); |
| 117 | le32_add_cpu(&desc->pg_nfrees, n); |
| 118 | spin_unlock(nilfs_mdt_bgl_lock(inode, group)); |
| 119 | } |
| 120 | |
| 121 | static unsigned long |
| 122 | nilfs_palloc_entry_blkoff(const struct inode *inode, __u64 nr) |
| 123 | { |
| 124 | unsigned long group, group_offset; |
| 125 | |
| 126 | group = nilfs_palloc_group(inode, nr, &group_offset); |
| 127 | |
| 128 | return nilfs_palloc_bitmap_blkoff(inode, group) + 1 + |
| 129 | group_offset / NILFS_MDT(inode)->mi_entries_per_block; |
| 130 | } |
| 131 | |
| 132 | static void nilfs_palloc_desc_block_init(struct inode *inode, |
| 133 | struct buffer_head *bh, void *kaddr) |
| 134 | { |
| 135 | struct nilfs_palloc_group_desc *desc = kaddr + bh_offset(bh); |
| 136 | unsigned long n = nilfs_palloc_groups_per_desc_block(inode); |
| 137 | __le32 nfrees; |
| 138 | |
| 139 | nfrees = cpu_to_le32(nilfs_palloc_entries_per_group(inode)); |
| 140 | while (n-- > 0) { |
| 141 | desc->pg_nfrees = nfrees; |
| 142 | desc++; |
| 143 | } |
| 144 | } |
| 145 | |
Ryusuke Konishi | 70622a2 | 2009-11-14 18:40:27 +0900 | [diff] [blame] | 146 | static int nilfs_palloc_get_block(struct inode *inode, unsigned long blkoff, |
| 147 | int create, |
| 148 | void (*init_block)(struct inode *, |
| 149 | struct buffer_head *, |
| 150 | void *), |
| 151 | struct buffer_head **bhp, |
| 152 | struct nilfs_bh_assoc *prev, |
| 153 | spinlock_t *lock) |
| 154 | { |
| 155 | int ret; |
| 156 | |
| 157 | spin_lock(lock); |
| 158 | if (prev->bh && blkoff == prev->blkoff) { |
| 159 | get_bh(prev->bh); |
| 160 | *bhp = prev->bh; |
| 161 | spin_unlock(lock); |
| 162 | return 0; |
| 163 | } |
| 164 | spin_unlock(lock); |
| 165 | |
| 166 | ret = nilfs_mdt_get_block(inode, blkoff, create, init_block, bhp); |
| 167 | if (!ret) { |
| 168 | spin_lock(lock); |
| 169 | /* |
| 170 | * The following code must be safe for change of the |
| 171 | * cache contents during the get block call. |
| 172 | */ |
| 173 | brelse(prev->bh); |
| 174 | get_bh(*bhp); |
| 175 | prev->bh = *bhp; |
| 176 | prev->blkoff = blkoff; |
| 177 | spin_unlock(lock); |
| 178 | } |
| 179 | return ret; |
| 180 | } |
| 181 | |
Ryusuke Konishi | 5442680 | 2009-04-06 19:01:29 -0700 | [diff] [blame] | 182 | static int nilfs_palloc_get_desc_block(struct inode *inode, |
| 183 | unsigned long group, |
| 184 | int create, struct buffer_head **bhp) |
| 185 | { |
Ryusuke Konishi | 70622a2 | 2009-11-14 18:40:27 +0900 | [diff] [blame] | 186 | struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; |
| 187 | |
| 188 | return nilfs_palloc_get_block(inode, |
| 189 | nilfs_palloc_desc_blkoff(inode, group), |
| 190 | create, nilfs_palloc_desc_block_init, |
| 191 | bhp, &cache->prev_desc, &cache->lock); |
Ryusuke Konishi | 5442680 | 2009-04-06 19:01:29 -0700 | [diff] [blame] | 192 | } |
| 193 | |
| 194 | static int nilfs_palloc_get_bitmap_block(struct inode *inode, |
| 195 | unsigned long group, |
| 196 | int create, struct buffer_head **bhp) |
| 197 | { |
Ryusuke Konishi | 70622a2 | 2009-11-14 18:40:27 +0900 | [diff] [blame] | 198 | struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; |
| 199 | |
| 200 | return nilfs_palloc_get_block(inode, |
| 201 | nilfs_palloc_bitmap_blkoff(inode, group), |
| 202 | create, NULL, bhp, |
| 203 | &cache->prev_bitmap, &cache->lock); |
Ryusuke Konishi | 5442680 | 2009-04-06 19:01:29 -0700 | [diff] [blame] | 204 | } |
| 205 | |
| 206 | int nilfs_palloc_get_entry_block(struct inode *inode, __u64 nr, |
| 207 | int create, struct buffer_head **bhp) |
| 208 | { |
Ryusuke Konishi | 70622a2 | 2009-11-14 18:40:27 +0900 | [diff] [blame] | 209 | struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; |
| 210 | |
| 211 | return nilfs_palloc_get_block(inode, |
| 212 | nilfs_palloc_entry_blkoff(inode, nr), |
| 213 | create, NULL, bhp, |
| 214 | &cache->prev_entry, &cache->lock); |
Ryusuke Konishi | 5442680 | 2009-04-06 19:01:29 -0700 | [diff] [blame] | 215 | } |
| 216 | |
| 217 | static struct nilfs_palloc_group_desc * |
| 218 | nilfs_palloc_block_get_group_desc(const struct inode *inode, |
| 219 | unsigned long group, |
| 220 | const struct buffer_head *bh, void *kaddr) |
| 221 | { |
| 222 | return (struct nilfs_palloc_group_desc *)(kaddr + bh_offset(bh)) + |
| 223 | group % nilfs_palloc_groups_per_desc_block(inode); |
| 224 | } |
| 225 | |
Ryusuke Konishi | 5442680 | 2009-04-06 19:01:29 -0700 | [diff] [blame] | 226 | void *nilfs_palloc_block_get_entry(const struct inode *inode, __u64 nr, |
| 227 | const struct buffer_head *bh, void *kaddr) |
| 228 | { |
| 229 | unsigned long entry_offset, group_offset; |
| 230 | |
| 231 | nilfs_palloc_group(inode, nr, &group_offset); |
| 232 | entry_offset = group_offset % NILFS_MDT(inode)->mi_entries_per_block; |
| 233 | |
| 234 | return kaddr + bh_offset(bh) + |
| 235 | entry_offset * NILFS_MDT(inode)->mi_entry_size; |
| 236 | } |
| 237 | |
| 238 | static int nilfs_palloc_find_available_slot(struct inode *inode, |
| 239 | unsigned long group, |
| 240 | unsigned long target, |
| 241 | unsigned char *bitmap, |
| 242 | int bsize) /* size in bits */ |
| 243 | { |
| 244 | int curr, pos, end, i; |
| 245 | |
| 246 | if (target > 0) { |
| 247 | end = (target + BITS_PER_LONG - 1) & ~(BITS_PER_LONG - 1); |
| 248 | if (end > bsize) |
| 249 | end = bsize; |
| 250 | pos = nilfs_find_next_zero_bit(bitmap, end, target); |
| 251 | if (pos < end && |
| 252 | !nilfs_set_bit_atomic( |
| 253 | nilfs_mdt_bgl_lock(inode, group), pos, bitmap)) |
| 254 | return pos; |
| 255 | } else |
| 256 | end = 0; |
| 257 | |
| 258 | for (i = 0, curr = end; |
| 259 | i < bsize; |
| 260 | i += BITS_PER_LONG, curr += BITS_PER_LONG) { |
| 261 | /* wrap around */ |
| 262 | if (curr >= bsize) |
| 263 | curr = 0; |
| 264 | while (*((unsigned long *)bitmap + curr / BITS_PER_LONG) |
| 265 | != ~0UL) { |
| 266 | end = curr + BITS_PER_LONG; |
| 267 | if (end > bsize) |
| 268 | end = bsize; |
| 269 | pos = nilfs_find_next_zero_bit(bitmap, end, curr); |
| 270 | if ((pos < end) && |
| 271 | !nilfs_set_bit_atomic( |
| 272 | nilfs_mdt_bgl_lock(inode, group), pos, |
| 273 | bitmap)) |
| 274 | return pos; |
| 275 | } |
| 276 | } |
| 277 | return -ENOSPC; |
| 278 | } |
| 279 | |
| 280 | static unsigned long |
| 281 | nilfs_palloc_rest_groups_in_desc_block(const struct inode *inode, |
| 282 | unsigned long curr, unsigned long max) |
| 283 | { |
| 284 | return min_t(unsigned long, |
| 285 | nilfs_palloc_groups_per_desc_block(inode) - |
| 286 | curr % nilfs_palloc_groups_per_desc_block(inode), |
| 287 | max - curr + 1); |
| 288 | } |
| 289 | |
| 290 | int nilfs_palloc_prepare_alloc_entry(struct inode *inode, |
| 291 | struct nilfs_palloc_req *req) |
| 292 | { |
| 293 | struct buffer_head *desc_bh, *bitmap_bh; |
| 294 | struct nilfs_palloc_group_desc *desc; |
| 295 | unsigned char *bitmap; |
| 296 | void *desc_kaddr, *bitmap_kaddr; |
| 297 | unsigned long group, maxgroup, ngroups; |
| 298 | unsigned long group_offset, maxgroup_offset; |
| 299 | unsigned long n, entries_per_group, groups_per_desc_block; |
| 300 | unsigned long i, j; |
| 301 | int pos, ret; |
| 302 | |
| 303 | ngroups = nilfs_palloc_groups_count(inode); |
| 304 | maxgroup = ngroups - 1; |
| 305 | group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset); |
| 306 | entries_per_group = nilfs_palloc_entries_per_group(inode); |
| 307 | groups_per_desc_block = nilfs_palloc_groups_per_desc_block(inode); |
| 308 | |
| 309 | for (i = 0; i < ngroups; i += n) { |
| 310 | if (group >= ngroups) { |
| 311 | /* wrap around */ |
| 312 | group = 0; |
| 313 | maxgroup = nilfs_palloc_group(inode, req->pr_entry_nr, |
| 314 | &maxgroup_offset) - 1; |
| 315 | } |
| 316 | ret = nilfs_palloc_get_desc_block(inode, group, 1, &desc_bh); |
| 317 | if (ret < 0) |
| 318 | return ret; |
| 319 | desc_kaddr = kmap(desc_bh->b_page); |
| 320 | desc = nilfs_palloc_block_get_group_desc( |
| 321 | inode, group, desc_bh, desc_kaddr); |
| 322 | n = nilfs_palloc_rest_groups_in_desc_block(inode, group, |
| 323 | maxgroup); |
| 324 | for (j = 0; j < n; j++, desc++, group++) { |
| 325 | if (nilfs_palloc_group_desc_nfrees(inode, group, desc) |
| 326 | > 0) { |
| 327 | ret = nilfs_palloc_get_bitmap_block( |
| 328 | inode, group, 1, &bitmap_bh); |
| 329 | if (ret < 0) |
| 330 | goto out_desc; |
| 331 | bitmap_kaddr = kmap(bitmap_bh->b_page); |
Ryusuke Konishi | 141bbdb | 2009-11-14 13:48:06 +0900 | [diff] [blame] | 332 | bitmap = bitmap_kaddr + bh_offset(bitmap_bh); |
Ryusuke Konishi | 5442680 | 2009-04-06 19:01:29 -0700 | [diff] [blame] | 333 | pos = nilfs_palloc_find_available_slot( |
| 334 | inode, group, group_offset, bitmap, |
| 335 | entries_per_group); |
| 336 | if (pos >= 0) { |
| 337 | /* found a free entry */ |
| 338 | nilfs_palloc_group_desc_add_entries( |
| 339 | inode, group, desc, -1); |
| 340 | req->pr_entry_nr = |
| 341 | entries_per_group * group + pos; |
| 342 | kunmap(desc_bh->b_page); |
| 343 | kunmap(bitmap_bh->b_page); |
| 344 | |
| 345 | req->pr_desc_bh = desc_bh; |
| 346 | req->pr_bitmap_bh = bitmap_bh; |
| 347 | return 0; |
| 348 | } |
| 349 | kunmap(bitmap_bh->b_page); |
| 350 | brelse(bitmap_bh); |
| 351 | } |
| 352 | |
| 353 | group_offset = 0; |
| 354 | } |
| 355 | |
| 356 | kunmap(desc_bh->b_page); |
| 357 | brelse(desc_bh); |
| 358 | } |
| 359 | |
| 360 | /* no entries left */ |
| 361 | return -ENOSPC; |
| 362 | |
| 363 | out_desc: |
| 364 | kunmap(desc_bh->b_page); |
| 365 | brelse(desc_bh); |
| 366 | return ret; |
| 367 | } |
| 368 | |
| 369 | void nilfs_palloc_commit_alloc_entry(struct inode *inode, |
| 370 | struct nilfs_palloc_req *req) |
| 371 | { |
| 372 | nilfs_mdt_mark_buffer_dirty(req->pr_bitmap_bh); |
| 373 | nilfs_mdt_mark_buffer_dirty(req->pr_desc_bh); |
| 374 | nilfs_mdt_mark_dirty(inode); |
| 375 | |
| 376 | brelse(req->pr_bitmap_bh); |
| 377 | brelse(req->pr_desc_bh); |
| 378 | } |
| 379 | |
| 380 | void nilfs_palloc_commit_free_entry(struct inode *inode, |
| 381 | struct nilfs_palloc_req *req) |
| 382 | { |
| 383 | struct nilfs_palloc_group_desc *desc; |
| 384 | unsigned long group, group_offset; |
| 385 | unsigned char *bitmap; |
| 386 | void *desc_kaddr, *bitmap_kaddr; |
| 387 | |
| 388 | group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset); |
| 389 | desc_kaddr = kmap(req->pr_desc_bh->b_page); |
| 390 | desc = nilfs_palloc_block_get_group_desc(inode, group, |
| 391 | req->pr_desc_bh, desc_kaddr); |
| 392 | bitmap_kaddr = kmap(req->pr_bitmap_bh->b_page); |
Ryusuke Konishi | 141bbdb | 2009-11-14 13:48:06 +0900 | [diff] [blame] | 393 | bitmap = bitmap_kaddr + bh_offset(req->pr_bitmap_bh); |
Ryusuke Konishi | 5442680 | 2009-04-06 19:01:29 -0700 | [diff] [blame] | 394 | |
| 395 | if (!nilfs_clear_bit_atomic(nilfs_mdt_bgl_lock(inode, group), |
| 396 | group_offset, bitmap)) |
| 397 | printk(KERN_WARNING "%s: entry number %llu already freed\n", |
| 398 | __func__, (unsigned long long)req->pr_entry_nr); |
| 399 | |
| 400 | nilfs_palloc_group_desc_add_entries(inode, group, desc, 1); |
| 401 | |
| 402 | kunmap(req->pr_bitmap_bh->b_page); |
| 403 | kunmap(req->pr_desc_bh->b_page); |
| 404 | |
| 405 | nilfs_mdt_mark_buffer_dirty(req->pr_desc_bh); |
| 406 | nilfs_mdt_mark_buffer_dirty(req->pr_bitmap_bh); |
| 407 | nilfs_mdt_mark_dirty(inode); |
| 408 | |
| 409 | brelse(req->pr_bitmap_bh); |
| 410 | brelse(req->pr_desc_bh); |
| 411 | } |
| 412 | |
| 413 | void nilfs_palloc_abort_alloc_entry(struct inode *inode, |
| 414 | struct nilfs_palloc_req *req) |
| 415 | { |
| 416 | struct nilfs_palloc_group_desc *desc; |
| 417 | void *desc_kaddr, *bitmap_kaddr; |
| 418 | unsigned char *bitmap; |
| 419 | unsigned long group, group_offset; |
| 420 | |
| 421 | group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset); |
| 422 | desc_kaddr = kmap(req->pr_desc_bh->b_page); |
| 423 | desc = nilfs_palloc_block_get_group_desc(inode, group, |
| 424 | req->pr_desc_bh, desc_kaddr); |
| 425 | bitmap_kaddr = kmap(req->pr_bitmap_bh->b_page); |
Ryusuke Konishi | 141bbdb | 2009-11-14 13:48:06 +0900 | [diff] [blame] | 426 | bitmap = bitmap_kaddr + bh_offset(req->pr_bitmap_bh); |
Ryusuke Konishi | 5442680 | 2009-04-06 19:01:29 -0700 | [diff] [blame] | 427 | if (!nilfs_clear_bit_atomic(nilfs_mdt_bgl_lock(inode, group), |
| 428 | group_offset, bitmap)) |
| 429 | printk(KERN_WARNING "%s: entry numer %llu already freed\n", |
| 430 | __func__, (unsigned long long)req->pr_entry_nr); |
| 431 | |
| 432 | nilfs_palloc_group_desc_add_entries(inode, group, desc, 1); |
| 433 | |
| 434 | kunmap(req->pr_bitmap_bh->b_page); |
| 435 | kunmap(req->pr_desc_bh->b_page); |
| 436 | |
| 437 | brelse(req->pr_bitmap_bh); |
| 438 | brelse(req->pr_desc_bh); |
| 439 | |
| 440 | req->pr_entry_nr = 0; |
| 441 | req->pr_bitmap_bh = NULL; |
| 442 | req->pr_desc_bh = NULL; |
| 443 | } |
| 444 | |
| 445 | int nilfs_palloc_prepare_free_entry(struct inode *inode, |
| 446 | struct nilfs_palloc_req *req) |
| 447 | { |
| 448 | struct buffer_head *desc_bh, *bitmap_bh; |
| 449 | unsigned long group, group_offset; |
| 450 | int ret; |
| 451 | |
| 452 | group = nilfs_palloc_group(inode, req->pr_entry_nr, &group_offset); |
| 453 | ret = nilfs_palloc_get_desc_block(inode, group, 1, &desc_bh); |
| 454 | if (ret < 0) |
| 455 | return ret; |
| 456 | ret = nilfs_palloc_get_bitmap_block(inode, group, 1, &bitmap_bh); |
| 457 | if (ret < 0) { |
| 458 | brelse(desc_bh); |
| 459 | return ret; |
| 460 | } |
| 461 | |
| 462 | req->pr_desc_bh = desc_bh; |
| 463 | req->pr_bitmap_bh = bitmap_bh; |
| 464 | return 0; |
| 465 | } |
| 466 | |
| 467 | void nilfs_palloc_abort_free_entry(struct inode *inode, |
| 468 | struct nilfs_palloc_req *req) |
| 469 | { |
| 470 | brelse(req->pr_bitmap_bh); |
| 471 | brelse(req->pr_desc_bh); |
| 472 | |
| 473 | req->pr_entry_nr = 0; |
| 474 | req->pr_bitmap_bh = NULL; |
| 475 | req->pr_desc_bh = NULL; |
| 476 | } |
| 477 | |
| 478 | static int |
| 479 | nilfs_palloc_group_is_in(struct inode *inode, unsigned long group, __u64 nr) |
| 480 | { |
| 481 | __u64 first, last; |
| 482 | |
| 483 | first = group * nilfs_palloc_entries_per_group(inode); |
| 484 | last = first + nilfs_palloc_entries_per_group(inode) - 1; |
| 485 | return (nr >= first) && (nr <= last); |
| 486 | } |
| 487 | |
| 488 | int nilfs_palloc_freev(struct inode *inode, __u64 *entry_nrs, size_t nitems) |
| 489 | { |
| 490 | struct buffer_head *desc_bh, *bitmap_bh; |
| 491 | struct nilfs_palloc_group_desc *desc; |
| 492 | unsigned char *bitmap; |
| 493 | void *desc_kaddr, *bitmap_kaddr; |
| 494 | unsigned long group, group_offset; |
| 495 | int i, j, n, ret; |
| 496 | |
| 497 | for (i = 0; i < nitems; i += n) { |
| 498 | group = nilfs_palloc_group(inode, entry_nrs[i], &group_offset); |
| 499 | ret = nilfs_palloc_get_desc_block(inode, group, 0, &desc_bh); |
| 500 | if (ret < 0) |
| 501 | return ret; |
| 502 | ret = nilfs_palloc_get_bitmap_block(inode, group, 0, |
| 503 | &bitmap_bh); |
| 504 | if (ret < 0) { |
| 505 | brelse(desc_bh); |
| 506 | return ret; |
| 507 | } |
| 508 | desc_kaddr = kmap(desc_bh->b_page); |
| 509 | desc = nilfs_palloc_block_get_group_desc( |
| 510 | inode, group, desc_bh, desc_kaddr); |
| 511 | bitmap_kaddr = kmap(bitmap_bh->b_page); |
Ryusuke Konishi | 141bbdb | 2009-11-14 13:48:06 +0900 | [diff] [blame] | 512 | bitmap = bitmap_kaddr + bh_offset(bitmap_bh); |
Ryusuke Konishi | 5442680 | 2009-04-06 19:01:29 -0700 | [diff] [blame] | 513 | for (j = i, n = 0; |
| 514 | (j < nitems) && nilfs_palloc_group_is_in(inode, group, |
| 515 | entry_nrs[j]); |
| 516 | j++, n++) { |
| 517 | nilfs_palloc_group(inode, entry_nrs[j], &group_offset); |
| 518 | if (!nilfs_clear_bit_atomic( |
| 519 | nilfs_mdt_bgl_lock(inode, group), |
| 520 | group_offset, bitmap)) { |
| 521 | printk(KERN_WARNING |
| 522 | "%s: entry number %llu already freed\n", |
| 523 | __func__, |
| 524 | (unsigned long long)entry_nrs[j]); |
| 525 | } |
| 526 | } |
| 527 | nilfs_palloc_group_desc_add_entries(inode, group, desc, n); |
| 528 | |
| 529 | kunmap(bitmap_bh->b_page); |
| 530 | kunmap(desc_bh->b_page); |
| 531 | |
| 532 | nilfs_mdt_mark_buffer_dirty(desc_bh); |
| 533 | nilfs_mdt_mark_buffer_dirty(bitmap_bh); |
| 534 | nilfs_mdt_mark_dirty(inode); |
| 535 | |
| 536 | brelse(bitmap_bh); |
| 537 | brelse(desc_bh); |
| 538 | } |
| 539 | return 0; |
| 540 | } |
Ryusuke Konishi | db38d5a | 2009-11-14 15:54:27 +0900 | [diff] [blame] | 541 | |
| 542 | void nilfs_palloc_setup_cache(struct inode *inode, |
| 543 | struct nilfs_palloc_cache *cache) |
| 544 | { |
| 545 | NILFS_MDT(inode)->mi_palloc_cache = cache; |
| 546 | spin_lock_init(&cache->lock); |
| 547 | } |
| 548 | |
| 549 | void nilfs_palloc_clear_cache(struct inode *inode) |
| 550 | { |
| 551 | struct nilfs_palloc_cache *cache = NILFS_MDT(inode)->mi_palloc_cache; |
| 552 | |
| 553 | spin_lock(&cache->lock); |
| 554 | brelse(cache->prev_desc.bh); |
| 555 | brelse(cache->prev_bitmap.bh); |
| 556 | brelse(cache->prev_entry.bh); |
| 557 | cache->prev_desc.bh = NULL; |
| 558 | cache->prev_bitmap.bh = NULL; |
| 559 | cache->prev_entry.bh = NULL; |
| 560 | spin_unlock(&cache->lock); |
| 561 | } |
| 562 | |
| 563 | void nilfs_palloc_destroy_cache(struct inode *inode) |
| 564 | { |
| 565 | nilfs_palloc_clear_cache(inode); |
| 566 | NILFS_MDT(inode)->mi_palloc_cache = NULL; |
| 567 | } |