blob: c0ae274c0a22560ff9850aa04daf2e3f465e8f17 [file] [log] [blame]
Linus Torvalds1da177e2005-04-16 15:20:36 -07001/*
2 * linux/fs/hfsplus/bitmap.c
3 *
4 * Copyright (C) 2001
5 * Brad Boyer (flar@allandria.com)
6 * (C) 2003 Ardis Technologies <roman@ardistech.com>
7 *
8 * Handling of allocation file
9 */
10
11#include <linux/pagemap.h>
12
13#include "hfsplus_fs.h"
14#include "hfsplus_raw.h"
15
Kirill A. Shutemov09cbfea2016-04-01 15:29:47 +030016#define PAGE_CACHE_BITS (PAGE_SIZE * 8)
Linus Torvalds1da177e2005-04-16 15:20:36 -070017
Anton Salikhmetov2753cc22010-12-16 18:08:38 +020018int hfsplus_block_allocate(struct super_block *sb, u32 size,
19 u32 offset, u32 *max)
Linus Torvalds1da177e2005-04-16 15:20:36 -070020{
Christoph Hellwigdd73a012010-10-01 05:42:59 +020021 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
Linus Torvalds1da177e2005-04-16 15:20:36 -070022 struct page *page;
23 struct address_space *mapping;
24 __be32 *pptr, *curr, *end;
25 u32 mask, start, len, n;
26 __be32 val;
27 int i;
28
29 len = *max;
30 if (!len)
31 return size;
32
Joe Perchesc2b3e1f2013-04-30 15:27:54 -070033 hfs_dbg(BITMAP, "block_allocate: %u,%u,%u\n", size, offset, len);
Christoph Hellwigdd73a012010-10-01 05:42:59 +020034 mutex_lock(&sbi->alloc_mutex);
35 mapping = sbi->alloc_file->i_mapping;
Pekka Enberg090d2b12006-06-23 02:05:08 -070036 page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS, NULL);
Eric Sesterhenn649f1ee2008-10-15 22:04:10 -070037 if (IS_ERR(page)) {
38 start = size;
39 goto out;
40 }
Linus Torvalds1da177e2005-04-16 15:20:36 -070041 pptr = kmap(page);
42 curr = pptr + (offset & (PAGE_CACHE_BITS - 1)) / 32;
43 i = offset % 32;
44 offset &= ~(PAGE_CACHE_BITS - 1);
45 if ((size ^ offset) / PAGE_CACHE_BITS)
46 end = pptr + PAGE_CACHE_BITS / 32;
47 else
48 end = pptr + ((size + 31) & (PAGE_CACHE_BITS - 1)) / 32;
49
50 /* scan the first partial u32 for zero bits */
51 val = *curr;
52 if (~val) {
53 n = be32_to_cpu(val);
54 mask = (1U << 31) >> i;
55 for (; i < 32; mask >>= 1, i++) {
56 if (!(n & mask))
57 goto found;
58 }
59 }
60 curr++;
61
62 /* scan complete u32s for the first zero bit */
63 while (1) {
64 while (curr < end) {
65 val = *curr;
66 if (~val) {
67 n = be32_to_cpu(val);
68 mask = 1 << 31;
69 for (i = 0; i < 32; mask >>= 1, i++) {
70 if (!(n & mask))
71 goto found;
72 }
73 }
74 curr++;
75 }
76 kunmap(page);
77 offset += PAGE_CACHE_BITS;
78 if (offset >= size)
79 break;
Pekka Enberg090d2b12006-06-23 02:05:08 -070080 page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS,
81 NULL);
Eric Sesterhenn649f1ee2008-10-15 22:04:10 -070082 if (IS_ERR(page)) {
83 start = size;
84 goto out;
85 }
Linus Torvalds1da177e2005-04-16 15:20:36 -070086 curr = pptr = kmap(page);
87 if ((size ^ offset) / PAGE_CACHE_BITS)
88 end = pptr + PAGE_CACHE_BITS / 32;
89 else
90 end = pptr + ((size + 31) & (PAGE_CACHE_BITS - 1)) / 32;
91 }
Joe Perchesc2b3e1f2013-04-30 15:27:54 -070092 hfs_dbg(BITMAP, "bitmap full\n");
Linus Torvalds1da177e2005-04-16 15:20:36 -070093 start = size;
94 goto out;
95
96found:
97 start = offset + (curr - pptr) * 32 + i;
98 if (start >= size) {
Joe Perchesc2b3e1f2013-04-30 15:27:54 -070099 hfs_dbg(BITMAP, "bitmap full\n");
Linus Torvalds1da177e2005-04-16 15:20:36 -0700100 goto out;
101 }
102 /* do any partial u32 at the start */
103 len = min(size - start, len);
104 while (1) {
105 n |= mask;
106 if (++i >= 32)
107 break;
108 mask >>= 1;
109 if (!--len || n & mask)
110 goto done;
111 }
112 if (!--len)
113 goto done;
114 *curr++ = cpu_to_be32(n);
115 /* do full u32s */
116 while (1) {
117 while (curr < end) {
118 n = be32_to_cpu(*curr);
119 if (len < 32)
120 goto last;
121 if (n) {
122 len = 32;
123 goto last;
124 }
125 *curr++ = cpu_to_be32(0xffffffff);
126 len -= 32;
127 }
128 set_page_dirty(page);
129 kunmap(page);
130 offset += PAGE_CACHE_BITS;
Pekka Enberg090d2b12006-06-23 02:05:08 -0700131 page = read_mapping_page(mapping, offset / PAGE_CACHE_BITS,
132 NULL);
Eric Sesterhenn649f1ee2008-10-15 22:04:10 -0700133 if (IS_ERR(page)) {
134 start = size;
135 goto out;
136 }
Linus Torvalds1da177e2005-04-16 15:20:36 -0700137 pptr = kmap(page);
138 curr = pptr;
139 end = pptr + PAGE_CACHE_BITS / 32;
140 }
141last:
142 /* do any partial u32 at end */
143 mask = 1U << 31;
144 for (i = 0; i < len; i++) {
145 if (n & mask)
146 break;
147 n |= mask;
148 mask >>= 1;
149 }
150done:
151 *curr = cpu_to_be32(n);
152 set_page_dirty(page);
153 kunmap(page);
154 *max = offset + (curr - pptr) * 32 + i - start;
Christoph Hellwigdd73a012010-10-01 05:42:59 +0200155 sbi->free_blocks -= *max;
Artem Bityutskiy9e6c5822012-07-12 17:26:31 +0300156 hfsplus_mark_mdb_dirty(sb);
Joe Perchesc2b3e1f2013-04-30 15:27:54 -0700157 hfs_dbg(BITMAP, "-> %u,%u\n", start, *max);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700158out:
Christoph Hellwigdd73a012010-10-01 05:42:59 +0200159 mutex_unlock(&sbi->alloc_mutex);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700160 return start;
161}
162
163int hfsplus_block_free(struct super_block *sb, u32 offset, u32 count)
164{
Christoph Hellwigdd73a012010-10-01 05:42:59 +0200165 struct hfsplus_sb_info *sbi = HFSPLUS_SB(sb);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700166 struct page *page;
167 struct address_space *mapping;
168 __be32 *pptr, *curr, *end;
169 u32 mask, len, pnr;
170 int i;
171
172 /* is there any actual work to be done? */
173 if (!count)
174 return 0;
175
Joe Perchesc2b3e1f2013-04-30 15:27:54 -0700176 hfs_dbg(BITMAP, "block_free: %u,%u\n", offset, count);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700177 /* are all of the bits in range? */
Christoph Hellwigdd73a012010-10-01 05:42:59 +0200178 if ((offset + count) > sbi->total_blocks)
Alan Cox5daa6692012-12-20 15:05:24 -0800179 return -ENOENT;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700180
Christoph Hellwigdd73a012010-10-01 05:42:59 +0200181 mutex_lock(&sbi->alloc_mutex);
182 mapping = sbi->alloc_file->i_mapping;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700183 pnr = offset / PAGE_CACHE_BITS;
Pekka Enberg090d2b12006-06-23 02:05:08 -0700184 page = read_mapping_page(mapping, pnr, NULL);
Alan Cox5daa6692012-12-20 15:05:24 -0800185 if (IS_ERR(page))
186 goto kaboom;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700187 pptr = kmap(page);
188 curr = pptr + (offset & (PAGE_CACHE_BITS - 1)) / 32;
189 end = pptr + PAGE_CACHE_BITS / 32;
190 len = count;
191
192 /* do any partial u32 at the start */
193 i = offset % 32;
194 if (i) {
195 int j = 32 - i;
196 mask = 0xffffffffU << j;
197 if (j > count) {
198 mask |= 0xffffffffU >> (i + count);
199 *curr++ &= cpu_to_be32(mask);
200 goto out;
201 }
202 *curr++ &= cpu_to_be32(mask);
203 count -= j;
204 }
205
206 /* do full u32s */
207 while (1) {
208 while (curr < end) {
209 if (count < 32)
210 goto done;
211 *curr++ = 0;
212 count -= 32;
213 }
214 if (!count)
215 break;
216 set_page_dirty(page);
217 kunmap(page);
Pekka Enberg090d2b12006-06-23 02:05:08 -0700218 page = read_mapping_page(mapping, ++pnr, NULL);
Alan Cox5daa6692012-12-20 15:05:24 -0800219 if (IS_ERR(page))
220 goto kaboom;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700221 pptr = kmap(page);
222 curr = pptr;
223 end = pptr + PAGE_CACHE_BITS / 32;
224 }
225done:
226 /* do any partial u32 at end */
227 if (count) {
228 mask = 0xffffffffU >> count;
229 *curr &= cpu_to_be32(mask);
230 }
231out:
232 set_page_dirty(page);
233 kunmap(page);
Christoph Hellwigdd73a012010-10-01 05:42:59 +0200234 sbi->free_blocks += len;
Artem Bityutskiy9e6c5822012-07-12 17:26:31 +0300235 hfsplus_mark_mdb_dirty(sb);
Christoph Hellwigdd73a012010-10-01 05:42:59 +0200236 mutex_unlock(&sbi->alloc_mutex);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700237
238 return 0;
Alan Cox5daa6692012-12-20 15:05:24 -0800239
240kaboom:
Vyacheslav Dubeyko865f38a2013-04-30 15:27:58 -0700241 pr_crit("unable to mark blocks free: error %ld\n", PTR_ERR(page));
Alan Cox5daa6692012-12-20 15:05:24 -0800242 mutex_unlock(&sbi->alloc_mutex);
243
244 return -EIO;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700245}