blob: 2f48d716d6a77286168793c378a5e635a76284a8 [file] [log] [blame]
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001/* drivers/android/pmem.c
2 *
3 * Copyright (C) 2007 Google, Inc.
Vipul Gandhif752bf62012-01-09 15:34:04 -08004 * Copyright (c) 2009-2012, Code Aurora Forum. All rights reserved.
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07005 *
6 * This software is licensed under the terms of the GNU General Public
7 * License version 2, as published by the Free Software Foundation, and
8 * may be copied, distributed, and modified under those terms.
9 *
10 * This program is distributed in the hope that it will be useful,
11 * but WITHOUT ANY WARRANTY; without even the implied warranty of
12 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
13 * GNU General Public License for more details.
14 *
15 */
16
17#include <linux/miscdevice.h>
18#include <linux/platform_device.h>
19#include <linux/fs.h>
20#include <linux/file.h>
Laura Abbott511edaf2011-12-14 13:34:53 -080021#include <linux/fmem.h>
Rebecca Schultza4ff0e82008-07-24 11:22:53 -070022#include <linux/mm.h>
23#include <linux/list.h>
Rebecca Schultza4ff0e82008-07-24 11:22:53 -070024#include <linux/debugfs.h>
25#include <linux/android_pmem.h>
26#include <linux/mempolicy.h>
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070027#include <linux/kobject.h>
Naveen Ramaraj189f1882011-08-16 17:39:22 -070028#include <linux/pm_runtime.h>
29#include <linux/memory_alloc.h>
30#include <linux/vmalloc.h>
31#include <linux/io.h>
32#include <linux/mm_types.h>
Rebecca Schultza4ff0e82008-07-24 11:22:53 -070033#include <asm/io.h>
34#include <asm/uaccess.h>
35#include <asm/cacheflush.h>
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070036#include <asm/sizes.h>
Naveen Ramaraj189f1882011-08-16 17:39:22 -070037#include <asm/mach/map.h>
38#include <asm/page.h>
Rebecca Schultza4ff0e82008-07-24 11:22:53 -070039
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070040#define PMEM_MAX_DEVICES (10)
41
42#define PMEM_MAX_ORDER (128)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -070043#define PMEM_MIN_ALLOC PAGE_SIZE
44
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070045#define PMEM_INITIAL_NUM_BITMAP_ALLOCATIONS (64)
46
47#define PMEM_32BIT_WORD_ORDER (5)
48#define PMEM_BITS_PER_WORD_MASK (BITS_PER_LONG - 1)
49
50#ifdef CONFIG_ANDROID_PMEM_DEBUG
Rebecca Schultza4ff0e82008-07-24 11:22:53 -070051#define PMEM_DEBUG 1
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -070052#else
53#define PMEM_DEBUG 0
54#endif
55
56#define SYSTEM_ALLOC_RETRY 10
Rebecca Schultza4ff0e82008-07-24 11:22:53 -070057
58/* indicates that a refernce to this file has been taken via get_pmem_file,
59 * the file should not be released until put_pmem_file is called */
60#define PMEM_FLAGS_BUSY 0x1
61/* indicates that this is a suballocation of a larger master range */
62#define PMEM_FLAGS_CONNECTED 0x1 << 1
63/* indicates this is a master and not a sub allocation and that it is mmaped */
64#define PMEM_FLAGS_MASTERMAP 0x1 << 2
65/* submap and unsubmap flags indicate:
66 * 00: subregion has never been mmaped
67 * 10: subregion has been mmaped, reference to the mm was taken
68 * 11: subretion has ben released, refernece to the mm still held
69 * 01: subretion has been released, reference to the mm has been released
70 */
71#define PMEM_FLAGS_SUBMAP 0x1 << 3
72#define PMEM_FLAGS_UNSUBMAP 0x1 << 4
73
Rebecca Schultza4ff0e82008-07-24 11:22:53 -070074struct pmem_data {
75 /* in alloc mode: an index into the bitmap
76 * in no_alloc mode: the size of the allocation */
77 int index;
78 /* see flags above for descriptions */
79 unsigned int flags;
80 /* protects this data field, if the mm_mmap sem will be held at the
81 * same time as this sem, the mm sem must be taken first (as this is
82 * the order for vma_open and vma_close ops */
83 struct rw_semaphore sem;
84 /* info about the mmaping process */
85 struct vm_area_struct *vma;
86 /* task struct of the mapping process */
87 struct task_struct *task;
88 /* process id of teh mapping process */
89 pid_t pid;
90 /* file descriptor of the master */
91 int master_fd;
92 /* file struct of the master */
93 struct file *master_file;
94 /* a list of currently available regions if this is a suballocation */
95 struct list_head region_list;
96 /* a linked list of data so we can access them for debugging */
97 struct list_head list;
98#if PMEM_DEBUG
99 int ref;
100#endif
101};
102
103struct pmem_bits {
104 unsigned allocated:1; /* 1 if allocated, 0 if free */
105 unsigned order:7; /* size of the region in pmem space */
106};
107
108struct pmem_region_node {
109 struct pmem_region region;
110 struct list_head list;
111};
112
113#define PMEM_DEBUG_MSGS 0
114#if PMEM_DEBUG_MSGS
115#define DLOG(fmt,args...) \
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700116 do { pr_debug("[%s:%s:%d] "fmt, __FILE__, __func__, __LINE__, \
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700117 ##args); } \
118 while (0)
119#else
120#define DLOG(x...) do {} while (0)
121#endif
122
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700123enum pmem_align {
124 PMEM_ALIGN_4K,
125 PMEM_ALIGN_1M,
126};
127
128#define PMEM_NAME_SIZE 16
129
130struct alloc_list {
131 void *addr; /* physical addr of allocation */
132 void *aaddr; /* aligned physical addr */
133 unsigned int size; /* total size of allocation */
134 unsigned char __iomem *vaddr; /* Virtual addr */
135 struct list_head allocs;
136};
137
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700138struct pmem_info {
139 struct miscdevice dev;
140 /* physical start address of the remaped pmem space */
141 unsigned long base;
142 /* vitual start address of the remaped pmem space */
143 unsigned char __iomem *vbase;
144 /* total size of the pmem space */
145 unsigned long size;
146 /* number of entries in the pmem space */
147 unsigned long num_entries;
148 /* pfn of the garbage page in memory */
149 unsigned long garbage_pfn;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700150 /* which memory type (i.e. SMI, EBI1) this PMEM device is backed by */
151 unsigned memory_type;
152
153 char name[PMEM_NAME_SIZE];
154
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700155 /* index of the garbage page in the pmem space */
156 int garbage_index;
Naveen Ramaraj189f1882011-08-16 17:39:22 -0700157 /* reserved virtual address range */
158 struct vm_struct *area;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700159
160 enum pmem_allocator_type allocator_type;
161
162 int (*allocate)(const int,
163 const unsigned long,
164 const unsigned int);
165 int (*free)(int, int);
166 int (*free_space)(int, struct pmem_freespace *);
167 unsigned long (*len)(int, struct pmem_data *);
168 unsigned long (*start_addr)(int, struct pmem_data *);
169
170 /* actual size of memory element, e.g.: (4 << 10) is 4K */
171 unsigned int quantum;
172
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700173 /* indicates maps of this region should be cached, if a mix of
174 * cached and uncached is desired, set this and open the device with
175 * O_SYNC to get an uncached region */
176 unsigned cached;
177 unsigned buffered;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700178 union {
179 struct {
180 /* in all_or_nothing allocator mode the first mapper
181 * gets the whole space and sets this flag */
182 unsigned allocated;
183 } all_or_nothing;
184
185 struct {
186 /* the buddy allocator bitmap for the region
187 * indicating which entries are allocated and which
188 * are free.
189 */
190
191 struct pmem_bits *buddy_bitmap;
192 } buddy_bestfit;
193
194 struct {
195 unsigned int bitmap_free; /* # of zero bits/quanta */
196 uint32_t *bitmap;
197 int32_t bitmap_allocs;
198 struct {
199 short bit;
200 unsigned short quanta;
201 } *bitm_alloc;
202 } bitmap;
203
204 struct {
205 unsigned long used; /* Bytes currently allocated */
206 struct list_head alist; /* List of allocations */
207 } system_mem;
208 } allocator;
209
210 int id;
211 struct kobject kobj;
212
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700213 /* for debugging, creates a list of pmem file structs, the
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700214 * data_list_mutex should be taken before pmem_data->sem if both are
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700215 * needed */
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700216 struct mutex data_list_mutex;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700217 struct list_head data_list;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700218 /* arena_mutex protects the global allocation arena
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700219 *
220 * IF YOU TAKE BOTH LOCKS TAKE THEM IN THIS ORDER:
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700221 * down(pmem_data->sem) => mutex_lock(arena_mutex)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700222 */
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700223 struct mutex arena_mutex;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700224
225 long (*ioctl)(struct file *, unsigned int, unsigned long);
226 int (*release)(struct inode *, struct file *);
Laura Abbott1e36a022011-06-22 17:08:13 -0700227 /* reference count of allocations */
228 atomic_t allocation_cnt;
229 /*
230 * request function for a region when the allocation count goes
231 * from 0 -> 1
232 */
Laura Abbott72ae4bf2011-12-14 14:01:43 -0800233 int (*mem_request)(void *);
Laura Abbott1e36a022011-06-22 17:08:13 -0700234 /*
235 * release function for a region when the allocation count goes
236 * from 1 -> 0
237 */
Laura Abbott72ae4bf2011-12-14 14:01:43 -0800238 int (*mem_release)(void *);
Laura Abbott1e36a022011-06-22 17:08:13 -0700239 /*
240 * private data for the request/release callback
241 */
242 void *region_data;
243 /*
244 * map and unmap as needed
245 */
246 int map_on_demand;
Laura Abbott511edaf2011-12-14 13:34:53 -0800247 /*
248 * memory will be reused through fmem
249 */
250 int reusable;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700251};
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700252#define to_pmem_info_id(a) (container_of(a, struct pmem_info, kobj)->id)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700253
Laura Abbott1e36a022011-06-22 17:08:13 -0700254static void ioremap_pmem(int id);
255static void pmem_put_region(int id);
256static int pmem_get_region(int id);
257
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700258static struct pmem_info pmem[PMEM_MAX_DEVICES];
259static int id_count;
260
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700261#define PMEM_SYSFS_DIR_NAME "pmem_regions" /* under /sys/kernel/ */
262static struct kset *pmem_kset;
263
264#define PMEM_IS_FREE_BUDDY(id, index) \
265 (!(pmem[id].allocator.buddy_bestfit.buddy_bitmap[index].allocated))
266#define PMEM_BUDDY_ORDER(id, index) \
267 (pmem[id].allocator.buddy_bestfit.buddy_bitmap[index].order)
268#define PMEM_BUDDY_INDEX(id, index) \
269 (index ^ (1 << PMEM_BUDDY_ORDER(id, index)))
270#define PMEM_BUDDY_NEXT_INDEX(id, index) \
271 (index + (1 << PMEM_BUDDY_ORDER(id, index)))
272#define PMEM_OFFSET(index) (index * pmem[id].quantum)
273#define PMEM_START_ADDR(id, index) \
274 (PMEM_OFFSET(index) + pmem[id].base)
275#define PMEM_BUDDY_LEN(id, index) \
276 ((1 << PMEM_BUDDY_ORDER(id, index)) * pmem[id].quantum)
277#define PMEM_END_ADDR(id, index) \
278 (PMEM_START_ADDR(id, index) + PMEM_LEN(id, index))
279#define PMEM_START_VADDR(id, index) \
280 (PMEM_OFFSET(id, index) + pmem[id].vbase)
281#define PMEM_END_VADDR(id, index) \
282 (PMEM_START_VADDR(id, index) + PMEM_LEN(id, index))
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700283#define PMEM_REVOKED(data) (data->flags & PMEM_FLAGS_REVOKED)
284#define PMEM_IS_PAGE_ALIGNED(addr) (!((addr) & (~PAGE_MASK)))
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700285#define PMEM_IS_SUBMAP(data) \
286 ((data->flags & PMEM_FLAGS_SUBMAP) && \
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700287 (!(data->flags & PMEM_FLAGS_UNSUBMAP)))
288
289static int pmem_release(struct inode *, struct file *);
290static int pmem_mmap(struct file *, struct vm_area_struct *);
291static int pmem_open(struct inode *, struct file *);
292static long pmem_ioctl(struct file *, unsigned int, unsigned long);
293
294struct file_operations pmem_fops = {
295 .release = pmem_release,
296 .mmap = pmem_mmap,
297 .open = pmem_open,
298 .unlocked_ioctl = pmem_ioctl,
299};
300
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700301#define PMEM_ATTR(_name, _mode, _show, _store) { \
302 .attr = {.name = __stringify(_name), .mode = _mode }, \
303 .show = _show, \
304 .store = _store, \
305}
306
307struct pmem_attr {
308 struct attribute attr;
309 ssize_t(*show) (const int id, char * const);
310 ssize_t(*store) (const int id, const char * const, const size_t count);
311};
312#define to_pmem_attr(a) container_of(a, struct pmem_attr, attr)
313
314#define RW_PMEM_ATTR(name) \
315static struct pmem_attr pmem_attr_## name = \
316 PMEM_ATTR(name, S_IRUGO | S_IWUSR, show_pmem_## name, store_pmem_## name)
317
318#define RO_PMEM_ATTR(name) \
319static struct pmem_attr pmem_attr_## name = \
320 PMEM_ATTR(name, S_IRUGO, show_pmem_## name, NULL)
321
322#define WO_PMEM_ATTR(name) \
323static struct pmem_attr pmem_attr_## name = \
324 PMEM_ATTR(name, S_IWUSR, NULL, store_pmem_## name)
325
326static ssize_t show_pmem(struct kobject *kobj,
327 struct attribute *attr,
328 char *buf)
329{
330 struct pmem_attr *a = to_pmem_attr(attr);
331 return a->show ? a->show(to_pmem_info_id(kobj), buf) : -EIO;
332}
333
334static ssize_t store_pmem(struct kobject *kobj, struct attribute *attr,
335 const char *buf, size_t count)
336{
337 struct pmem_attr *a = to_pmem_attr(attr);
338 return a->store ? a->store(to_pmem_info_id(kobj), buf, count) : -EIO;
339}
340
341static struct sysfs_ops pmem_ops = {
342 .show = show_pmem,
343 .store = store_pmem,
344};
345
346static ssize_t show_pmem_base(int id, char *buf)
347{
348 return scnprintf(buf, PAGE_SIZE, "%lu(%#lx)\n",
349 pmem[id].base, pmem[id].base);
350}
351RO_PMEM_ATTR(base);
352
353static ssize_t show_pmem_size(int id, char *buf)
354{
355 return scnprintf(buf, PAGE_SIZE, "%lu(%#lx)\n",
356 pmem[id].size, pmem[id].size);
357}
358RO_PMEM_ATTR(size);
359
360static ssize_t show_pmem_allocator_type(int id, char *buf)
361{
362 switch (pmem[id].allocator_type) {
363 case PMEM_ALLOCATORTYPE_ALLORNOTHING:
364 return scnprintf(buf, PAGE_SIZE, "%s\n", "All or Nothing");
365 case PMEM_ALLOCATORTYPE_BUDDYBESTFIT:
366 return scnprintf(buf, PAGE_SIZE, "%s\n", "Buddy Bestfit");
367 case PMEM_ALLOCATORTYPE_BITMAP:
368 return scnprintf(buf, PAGE_SIZE, "%s\n", "Bitmap");
369 case PMEM_ALLOCATORTYPE_SYSTEM:
370 return scnprintf(buf, PAGE_SIZE, "%s\n", "System heap");
371 default:
372 return scnprintf(buf, PAGE_SIZE,
373 "??? Invalid allocator type (%d) for this region! "
374 "Something isn't right.\n",
375 pmem[id].allocator_type);
376 }
377}
378RO_PMEM_ATTR(allocator_type);
379
380static ssize_t show_pmem_mapped_regions(int id, char *buf)
381{
382 struct list_head *elt;
383 int ret;
384
385 ret = scnprintf(buf, PAGE_SIZE,
386 "pid #: mapped regions (offset, len) (offset,len)...\n");
387
388 mutex_lock(&pmem[id].data_list_mutex);
389 list_for_each(elt, &pmem[id].data_list) {
390 struct pmem_data *data =
391 list_entry(elt, struct pmem_data, list);
392 struct list_head *elt2;
393
394 down_read(&data->sem);
395 ret += scnprintf(buf + ret, PAGE_SIZE - ret, "pid %u:",
396 data->pid);
397 list_for_each(elt2, &data->region_list) {
398 struct pmem_region_node *region_node = list_entry(elt2,
399 struct pmem_region_node,
400 list);
401 ret += scnprintf(buf + ret, PAGE_SIZE - ret,
402 "(%lx,%lx) ",
403 region_node->region.offset,
404 region_node->region.len);
405 }
406 up_read(&data->sem);
407 ret += scnprintf(buf + ret, PAGE_SIZE - ret, "\n");
408 }
409 mutex_unlock(&pmem[id].data_list_mutex);
410 return ret;
411}
412RO_PMEM_ATTR(mapped_regions);
413
414#define PMEM_COMMON_SYSFS_ATTRS \
415 &pmem_attr_base.attr, \
416 &pmem_attr_size.attr, \
417 &pmem_attr_allocator_type.attr, \
418 &pmem_attr_mapped_regions.attr
419
420
421static ssize_t show_pmem_allocated(int id, char *buf)
422{
423 ssize_t ret;
424
425 mutex_lock(&pmem[id].arena_mutex);
426 ret = scnprintf(buf, PAGE_SIZE, "%s\n",
427 pmem[id].allocator.all_or_nothing.allocated ?
428 "is allocated" : "is NOT allocated");
429 mutex_unlock(&pmem[id].arena_mutex);
430 return ret;
431}
432RO_PMEM_ATTR(allocated);
433
434static struct attribute *pmem_allornothing_attrs[] = {
435 PMEM_COMMON_SYSFS_ATTRS,
436
437 &pmem_attr_allocated.attr,
438
439 NULL
440};
441
442static struct kobj_type pmem_allornothing_ktype = {
443 .sysfs_ops = &pmem_ops,
444 .default_attrs = pmem_allornothing_attrs,
445};
446
447static ssize_t show_pmem_total_entries(int id, char *buf)
448{
449 return scnprintf(buf, PAGE_SIZE, "%lu\n", pmem[id].num_entries);
450}
451RO_PMEM_ATTR(total_entries);
452
453static ssize_t show_pmem_quantum_size(int id, char *buf)
454{
455 return scnprintf(buf, PAGE_SIZE, "%u (%#x)\n",
456 pmem[id].quantum, pmem[id].quantum);
457}
458RO_PMEM_ATTR(quantum_size);
459
460static ssize_t show_pmem_buddy_bitmap_dump(int id, char *buf)
461{
462 int ret, i;
463
464 mutex_lock(&pmem[id].data_list_mutex);
465 ret = scnprintf(buf, PAGE_SIZE, "index\torder\tlength\tallocated\n");
466
467 for (i = 0; i < pmem[id].num_entries && (PAGE_SIZE - ret);
468 i = PMEM_BUDDY_NEXT_INDEX(id, i))
469 ret += scnprintf(buf + ret, PAGE_SIZE - ret, "%d\t%d\t%d\t%d\n",
470 i, PMEM_BUDDY_ORDER(id, i),
471 PMEM_BUDDY_LEN(id, i),
472 !PMEM_IS_FREE_BUDDY(id, i));
473
474 mutex_unlock(&pmem[id].data_list_mutex);
475 return ret;
476}
477RO_PMEM_ATTR(buddy_bitmap_dump);
478
479#define PMEM_BITMAP_BUDDY_BESTFIT_COMMON_SYSFS_ATTRS \
480 &pmem_attr_quantum_size.attr, \
481 &pmem_attr_total_entries.attr
482
483static struct attribute *pmem_buddy_bestfit_attrs[] = {
484 PMEM_COMMON_SYSFS_ATTRS,
485
486 PMEM_BITMAP_BUDDY_BESTFIT_COMMON_SYSFS_ATTRS,
487
488 &pmem_attr_buddy_bitmap_dump.attr,
489
490 NULL
491};
492
493static struct kobj_type pmem_buddy_bestfit_ktype = {
494 .sysfs_ops = &pmem_ops,
495 .default_attrs = pmem_buddy_bestfit_attrs,
496};
497
498static ssize_t show_pmem_free_quanta(int id, char *buf)
499{
500 ssize_t ret;
501
502 mutex_lock(&pmem[id].arena_mutex);
503 ret = scnprintf(buf, PAGE_SIZE, "%u\n",
504 pmem[id].allocator.bitmap.bitmap_free);
505 mutex_unlock(&pmem[id].arena_mutex);
506 return ret;
507}
508RO_PMEM_ATTR(free_quanta);
509
510static ssize_t show_pmem_bits_allocated(int id, char *buf)
511{
512 ssize_t ret;
513 unsigned int i;
514
515 mutex_lock(&pmem[id].arena_mutex);
516
517 ret = scnprintf(buf, PAGE_SIZE,
518 "id: %d\nbitnum\tindex\tquanta allocated\n", id);
519
520 for (i = 0; i < pmem[id].allocator.bitmap.bitmap_allocs; i++)
521 if (pmem[id].allocator.bitmap.bitm_alloc[i].bit != -1)
522 ret += scnprintf(buf + ret, PAGE_SIZE - ret,
523 "%u\t%u\t%u\n",
524 i,
525 pmem[id].allocator.bitmap.bitm_alloc[i].bit,
526 pmem[id].allocator.bitmap.bitm_alloc[i].quanta
527 );
528
529 mutex_unlock(&pmem[id].arena_mutex);
530 return ret;
531}
532RO_PMEM_ATTR(bits_allocated);
533
534static struct attribute *pmem_bitmap_attrs[] = {
535 PMEM_COMMON_SYSFS_ATTRS,
536
537 PMEM_BITMAP_BUDDY_BESTFIT_COMMON_SYSFS_ATTRS,
538
539 &pmem_attr_free_quanta.attr,
540 &pmem_attr_bits_allocated.attr,
541
542 NULL
543};
544
545static struct attribute *pmem_system_attrs[] = {
546 PMEM_COMMON_SYSFS_ATTRS,
547
548 NULL
549};
550
551static struct kobj_type pmem_bitmap_ktype = {
552 .sysfs_ops = &pmem_ops,
553 .default_attrs = pmem_bitmap_attrs,
554};
555
556static struct kobj_type pmem_system_ktype = {
557 .sysfs_ops = &pmem_ops,
558 .default_attrs = pmem_system_attrs,
559};
560
Laura Abbott1e36a022011-06-22 17:08:13 -0700561static int pmem_allocate_from_id(const int id, const unsigned long size,
562 const unsigned int align)
563{
564 int ret;
565 ret = pmem_get_region(id);
566
567 if (ret)
568 return -1;
569
570 ret = pmem[id].allocate(id, size, align);
571
572 if (ret < 0)
573 pmem_put_region(id);
574
575 return ret;
576}
577
578static int pmem_free_from_id(const int id, const int index)
579{
580 pmem_put_region(id);
581 return pmem[id].free(id, index);
582}
583
584static int pmem_get_region(int id)
585{
586 /* Must be called with arena mutex locked */
587 atomic_inc(&pmem[id].allocation_cnt);
588 if (!pmem[id].vbase) {
589 DLOG("PMEMDEBUG: mapping for %s", pmem[id].name);
Laura Abbott72ae4bf2011-12-14 14:01:43 -0800590 if (pmem[id].mem_request) {
591 int ret = pmem[id].mem_request(pmem[id].region_data);
592 if (ret) {
593 atomic_dec(&pmem[id].allocation_cnt);
594 return 1;
595 }
596 }
Laura Abbott1e36a022011-06-22 17:08:13 -0700597 ioremap_pmem(id);
598 }
599
600 if (pmem[id].vbase) {
601 return 0;
602 } else {
603 if (pmem[id].mem_release)
604 pmem[id].mem_release(pmem[id].region_data);
605 atomic_dec(&pmem[id].allocation_cnt);
606 return 1;
607 }
608}
609
610static void pmem_put_region(int id)
611{
612 /* Must be called with arena mutex locked */
613 if (atomic_dec_and_test(&pmem[id].allocation_cnt)) {
614 DLOG("PMEMDEBUG: unmapping for %s", pmem[id].name);
615 BUG_ON(!pmem[id].vbase);
616 if (pmem[id].map_on_demand) {
Naveen Ramaraj189f1882011-08-16 17:39:22 -0700617 /* unmap_kernel_range() flushes the caches
618 * and removes the page table entries
619 */
620 unmap_kernel_range((unsigned long)pmem[id].vbase,
621 pmem[id].size);
Laura Abbott1e36a022011-06-22 17:08:13 -0700622 pmem[id].vbase = NULL;
Laura Abbott72ae4bf2011-12-14 14:01:43 -0800623 if (pmem[id].mem_release) {
624 int ret = pmem[id].mem_release(
625 pmem[id].region_data);
626 WARN(ret, "mem_release failed");
627 }
Laura Abbott1e36a022011-06-22 17:08:13 -0700628
629 }
630 }
631}
632
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700633static int get_id(struct file *file)
634{
635 return MINOR(file->f_dentry->d_inode->i_rdev);
636}
637
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700638static char *get_name(struct file *file)
639{
640 int id = get_id(file);
641 return pmem[id].name;
642}
643
644static int is_pmem_file(struct file *file)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700645{
646 int id;
647
648 if (unlikely(!file || !file->f_dentry || !file->f_dentry->d_inode))
649 return 0;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700650
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700651 id = get_id(file);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700652 return (unlikely(id >= PMEM_MAX_DEVICES ||
653 file->f_dentry->d_inode->i_rdev !=
654 MKDEV(MISC_MAJOR, pmem[id].dev.minor))) ? 0 : 1;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700655}
656
657static int has_allocation(struct file *file)
658{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700659 /* must be called with at least read lock held on
660 * ((struct pmem_data *)(file->private_data))->sem which
661 * means that file is guaranteed not to be NULL upon entry!!
662 * check is_pmem_file first if not accessed via pmem_file_ops */
663 struct pmem_data *pdata = file->private_data;
664 return pdata && pdata->index != -1;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700665}
666
667static int is_master_owner(struct file *file)
668{
669 struct file *master_file;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700670 struct pmem_data *data = file->private_data;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700671 int put_needed, ret = 0;
672
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700673 if (!has_allocation(file))
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700674 return 0;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700675 if (PMEM_FLAGS_MASTERMAP & data->flags)
676 return 1;
677 master_file = fget_light(data->master_fd, &put_needed);
678 if (master_file && data->master_file == master_file)
679 ret = 1;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700680 if (master_file)
681 fput_light(master_file, put_needed);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700682 return ret;
683}
684
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700685static int pmem_free_all_or_nothing(int id, int index)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700686{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700687 /* caller should hold the lock on arena_mutex! */
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700688 DLOG("index %d\n", index);
689
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700690 pmem[id].allocator.all_or_nothing.allocated = 0;
691 return 0;
692}
693
694static int pmem_free_space_all_or_nothing(int id,
695 struct pmem_freespace *fs)
696{
697 /* caller should hold the lock on arena_mutex! */
698 fs->total = (unsigned long)
699 pmem[id].allocator.all_or_nothing.allocated == 0 ?
700 pmem[id].size : 0;
701
702 fs->largest = fs->total;
703 return 0;
704}
705
706
707static int pmem_free_buddy_bestfit(int id, int index)
708{
709 /* caller should hold the lock on arena_mutex! */
710 int curr = index;
711 DLOG("index %d\n", index);
712
713
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700714 /* clean up the bitmap, merging any buddies */
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700715 pmem[id].allocator.buddy_bestfit.buddy_bitmap[curr].allocated = 0;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700716 /* find a slots buddy Buddy# = Slot# ^ (1 << order)
717 * if the buddy is also free merge them
718 * repeat until the buddy is not free or end of the bitmap is reached
719 */
720 do {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700721 int buddy = PMEM_BUDDY_INDEX(id, curr);
722 if (buddy < pmem[id].num_entries &&
723 PMEM_IS_FREE_BUDDY(id, buddy) &&
724 PMEM_BUDDY_ORDER(id, buddy) ==
725 PMEM_BUDDY_ORDER(id, curr)) {
726 PMEM_BUDDY_ORDER(id, buddy)++;
727 PMEM_BUDDY_ORDER(id, curr)++;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700728 curr = min(buddy, curr);
729 } else {
730 break;
731 }
732 } while (curr < pmem[id].num_entries);
733
734 return 0;
735}
736
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700737
738static int pmem_free_space_buddy_bestfit(int id,
739 struct pmem_freespace *fs)
740{
741 /* caller should hold the lock on arena_mutex! */
742 int curr;
743 unsigned long size;
744 fs->total = 0;
745 fs->largest = 0;
746
747 for (curr = 0; curr < pmem[id].num_entries;
748 curr = PMEM_BUDDY_NEXT_INDEX(id, curr)) {
749 if (PMEM_IS_FREE_BUDDY(id, curr)) {
750 size = PMEM_BUDDY_LEN(id, curr);
751 if (size > fs->largest)
752 fs->largest = size;
753 fs->total += size;
754 }
755 }
756 return 0;
757}
758
759
760static inline uint32_t start_mask(int bit_start)
761{
762 return (uint32_t)(~0) << (bit_start & PMEM_BITS_PER_WORD_MASK);
763}
764
765static inline uint32_t end_mask(int bit_end)
766{
767 return (uint32_t)(~0) >>
768 ((BITS_PER_LONG - bit_end) & PMEM_BITS_PER_WORD_MASK);
769}
770
771static inline int compute_total_words(int bit_end, int word_index)
772{
773 return ((bit_end + BITS_PER_LONG - 1) >>
774 PMEM_32BIT_WORD_ORDER) - word_index;
775}
776
777static void bitmap_bits_clear_all(uint32_t *bitp, int bit_start, int bit_end)
778{
779 int word_index = bit_start >> PMEM_32BIT_WORD_ORDER, total_words;
780
781 total_words = compute_total_words(bit_end, word_index);
782 if (total_words > 0) {
783 if (total_words == 1) {
784 bitp[word_index] &=
785 ~(start_mask(bit_start) & end_mask(bit_end));
786 } else {
787 bitp[word_index++] &= ~start_mask(bit_start);
788 if (total_words > 2) {
789 int total_bytes;
790
791 total_words -= 2;
792 total_bytes = total_words << 2;
793
794 memset(&bitp[word_index], 0, total_bytes);
795 word_index += total_words;
796 }
797 bitp[word_index] &= ~end_mask(bit_end);
798 }
799 }
800}
801
802static int pmem_free_bitmap(int id, int bitnum)
803{
804 /* caller should hold the lock on arena_mutex! */
805 int i;
806 char currtask_name[FIELD_SIZEOF(struct task_struct, comm) + 1];
807
808 DLOG("bitnum %d\n", bitnum);
809
810 for (i = 0; i < pmem[id].allocator.bitmap.bitmap_allocs; i++) {
811 const int curr_bit =
812 pmem[id].allocator.bitmap.bitm_alloc[i].bit;
813
814 if (curr_bit == bitnum) {
815 const int curr_quanta =
816 pmem[id].allocator.bitmap.bitm_alloc[i].quanta;
817
818 bitmap_bits_clear_all(pmem[id].allocator.bitmap.bitmap,
819 curr_bit, curr_bit + curr_quanta);
820 pmem[id].allocator.bitmap.bitmap_free += curr_quanta;
821 pmem[id].allocator.bitmap.bitm_alloc[i].bit = -1;
822 pmem[id].allocator.bitmap.bitm_alloc[i].quanta = 0;
823 return 0;
824 }
825 }
826 printk(KERN_ALERT "pmem: %s: Attempt to free unallocated index %d, id"
827 " %d, pid %d(%s)\n", __func__, bitnum, id, current->pid,
828 get_task_comm(currtask_name, current));
829
830 return -1;
831}
832
833static int pmem_free_system(int id, int index)
834{
835 /* caller should hold the lock on arena_mutex! */
836 struct alloc_list *item;
837
838 DLOG("index %d\n", index);
839 if (index != 0)
840 item = (struct alloc_list *)index;
841 else
842 return 0;
843
844 if (item->vaddr != NULL) {
845 iounmap(item->vaddr);
846 kfree(__va(item->addr));
847 list_del(&item->allocs);
848 kfree(item);
849 }
850
851 return 0;
852}
853
854static int pmem_free_space_bitmap(int id, struct pmem_freespace *fs)
855{
856 int i, j;
857 int max_allocs = pmem[id].allocator.bitmap.bitmap_allocs;
858 int alloc_start = 0;
859 int next_alloc;
860 unsigned long size = 0;
861
862 fs->total = 0;
863 fs->largest = 0;
864
865 for (i = 0; i < max_allocs; i++) {
866
867 int alloc_quanta = 0;
868 int alloc_idx = 0;
869 next_alloc = pmem[id].num_entries;
870
871 /* Look for the lowest bit where next allocation starts */
872 for (j = 0; j < max_allocs; j++) {
873 const int curr_alloc = pmem[id].allocator.
874 bitmap.bitm_alloc[j].bit;
875 if (curr_alloc != -1) {
876 if (alloc_start == curr_alloc)
877 alloc_idx = j;
878 if (alloc_start >= curr_alloc)
879 continue;
880 if (curr_alloc < next_alloc)
881 next_alloc = curr_alloc;
882 }
883 }
884 alloc_quanta = pmem[id].allocator.bitmap.
885 bitm_alloc[alloc_idx].quanta;
886 size = (next_alloc - (alloc_start + alloc_quanta)) *
887 pmem[id].quantum;
888
889 if (size > fs->largest)
890 fs->largest = size;
891 fs->total += size;
892
893 if (next_alloc == pmem[id].num_entries)
894 break;
895 else
896 alloc_start = next_alloc;
897 }
898
899 return 0;
900}
901
902static int pmem_free_space_system(int id, struct pmem_freespace *fs)
903{
904 fs->total = pmem[id].size;
905 fs->largest = pmem[id].size;
906
907 return 0;
908}
909
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700910static void pmem_revoke(struct file *file, struct pmem_data *data);
911
912static int pmem_release(struct inode *inode, struct file *file)
913{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700914 struct pmem_data *data = file->private_data;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700915 struct pmem_region_node *region_node;
916 struct list_head *elt, *elt2;
917 int id = get_id(file), ret = 0;
918
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700919#if PMEM_DEBUG_MSGS
920 char currtask_name[FIELD_SIZEOF(struct task_struct, comm) + 1];
921#endif
922 DLOG("releasing memory pid %u(%s) file %p(%ld) dev %s(id: %d)\n",
923 current->pid, get_task_comm(currtask_name, current),
924 file, file_count(file), get_name(file), id);
925 mutex_lock(&pmem[id].data_list_mutex);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700926 /* if this file is a master, revoke all the memory in the connected
927 * files */
928 if (PMEM_FLAGS_MASTERMAP & data->flags) {
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700929 list_for_each(elt, &pmem[id].data_list) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700930 struct pmem_data *sub_data =
931 list_entry(elt, struct pmem_data, list);
932 int is_master;
933
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700934 down_read(&sub_data->sem);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700935 is_master = (PMEM_IS_SUBMAP(sub_data) &&
936 file == sub_data->master_file);
937 up_read(&sub_data->sem);
938
939 if (is_master)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700940 pmem_revoke(file, sub_data);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700941 }
942 }
943 list_del(&data->list);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700944 mutex_unlock(&pmem[id].data_list_mutex);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700945
946 down_write(&data->sem);
947
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700948 /* if it is not a connected file and it has an allocation, free it */
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700949 if (!(PMEM_FLAGS_CONNECTED & data->flags) && has_allocation(file)) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700950 mutex_lock(&pmem[id].arena_mutex);
Laura Abbott1e36a022011-06-22 17:08:13 -0700951 ret = pmem_free_from_id(id, data->index);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700952 mutex_unlock(&pmem[id].arena_mutex);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700953 }
954
955 /* if this file is a submap (mapped, connected file), downref the
956 * task struct */
957 if (PMEM_FLAGS_SUBMAP & data->flags)
958 if (data->task) {
959 put_task_struct(data->task);
960 data->task = NULL;
961 }
962
963 file->private_data = NULL;
964
965 list_for_each_safe(elt, elt2, &data->region_list) {
966 region_node = list_entry(elt, struct pmem_region_node, list);
967 list_del(elt);
968 kfree(region_node);
969 }
970 BUG_ON(!list_empty(&data->region_list));
971
972 up_write(&data->sem);
973 kfree(data);
974 if (pmem[id].release)
975 ret = pmem[id].release(inode, file);
976
977 return ret;
978}
979
980static int pmem_open(struct inode *inode, struct file *file)
981{
982 struct pmem_data *data;
983 int id = get_id(file);
984 int ret = 0;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700985#if PMEM_DEBUG_MSGS
986 char currtask_name[FIELD_SIZEOF(struct task_struct, comm) + 1];
987#endif
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700988
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700989 DLOG("pid %u(%s) file %p(%ld) dev %s(id: %d)\n",
990 current->pid, get_task_comm(currtask_name, current),
991 file, file_count(file), get_name(file), id);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700992 data = kmalloc(sizeof(struct pmem_data), GFP_KERNEL);
993 if (!data) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -0700994 printk(KERN_ALERT "pmem: %s: unable to allocate memory for "
995 "pmem metadata.", __func__);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -0700996 return -1;
997 }
998 data->flags = 0;
999 data->index = -1;
1000 data->task = NULL;
1001 data->vma = NULL;
1002 data->pid = 0;
1003 data->master_file = NULL;
1004#if PMEM_DEBUG
1005 data->ref = 0;
1006#endif
1007 INIT_LIST_HEAD(&data->region_list);
1008 init_rwsem(&data->sem);
1009
1010 file->private_data = data;
1011 INIT_LIST_HEAD(&data->list);
1012
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001013 mutex_lock(&pmem[id].data_list_mutex);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001014 list_add(&data->list, &pmem[id].data_list);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001015 mutex_unlock(&pmem[id].data_list_mutex);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001016 return ret;
1017}
1018
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001019static unsigned long pmem_order(unsigned long len, int id)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001020{
1021 int i;
1022
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001023 len = (len + pmem[id].quantum - 1)/pmem[id].quantum;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001024 len--;
1025 for (i = 0; i < sizeof(len)*8; i++)
1026 if (len >> i == 0)
1027 break;
1028 return i;
1029}
1030
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001031static int pmem_allocator_all_or_nothing(const int id,
1032 const unsigned long len,
1033 const unsigned int align)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001034{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001035 /* caller should hold the lock on arena_mutex! */
1036 DLOG("all or nothing\n");
1037 if ((len > pmem[id].size) ||
1038 pmem[id].allocator.all_or_nothing.allocated)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001039 return -1;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001040 pmem[id].allocator.all_or_nothing.allocated = 1;
1041 return len;
1042}
1043
1044static int pmem_allocator_buddy_bestfit(const int id,
1045 const unsigned long len,
1046 unsigned int align)
1047{
1048 /* caller should hold the lock on arena_mutex! */
1049 int curr;
1050 int best_fit = -1;
1051 unsigned long order;
1052
1053 DLOG("buddy bestfit\n");
1054 order = pmem_order(len, id);
1055 if (order > PMEM_MAX_ORDER)
1056 goto out;
1057
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001058 DLOG("order %lx\n", order);
1059
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001060 /* Look through the bitmap.
1061 * If a free slot of the correct order is found, use it.
1062 * Otherwise, use the best fit (smallest with size > order) slot.
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001063 */
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001064 for (curr = 0;
1065 curr < pmem[id].num_entries;
1066 curr = PMEM_BUDDY_NEXT_INDEX(id, curr))
1067 if (PMEM_IS_FREE_BUDDY(id, curr)) {
1068 if (PMEM_BUDDY_ORDER(id, curr) ==
1069 (unsigned char)order) {
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001070 /* set the not free bit and clear others */
1071 best_fit = curr;
1072 break;
1073 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001074 if (PMEM_BUDDY_ORDER(id, curr) >
1075 (unsigned char)order &&
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001076 (best_fit < 0 ||
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001077 PMEM_BUDDY_ORDER(id, curr) <
1078 PMEM_BUDDY_ORDER(id, best_fit)))
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001079 best_fit = curr;
1080 }
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001081
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001082 /* if best_fit < 0, there are no suitable slots; return an error */
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001083 if (best_fit < 0) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001084#if PMEM_DEBUG
1085 printk(KERN_ALERT "pmem: %s: no space left to allocate!\n",
1086 __func__);
1087#endif
1088 goto out;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001089 }
1090
1091 /* now partition the best fit:
1092 * split the slot into 2 buddies of order - 1
1093 * repeat until the slot is of the correct order
1094 */
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001095 while (PMEM_BUDDY_ORDER(id, best_fit) > (unsigned char)order) {
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001096 int buddy;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001097 PMEM_BUDDY_ORDER(id, best_fit) -= 1;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001098 buddy = PMEM_BUDDY_INDEX(id, best_fit);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001099 PMEM_BUDDY_ORDER(id, buddy) = PMEM_BUDDY_ORDER(id, best_fit);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001100 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001101 pmem[id].allocator.buddy_bestfit.buddy_bitmap[best_fit].allocated = 1;
1102out:
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001103 return best_fit;
1104}
1105
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001106
1107static inline unsigned long paddr_from_bit(const int id, const int bitnum)
1108{
1109 return pmem[id].base + pmem[id].quantum * bitnum;
1110}
1111
1112static inline unsigned long bit_from_paddr(const int id,
1113 const unsigned long paddr)
1114{
1115 return (paddr - pmem[id].base) / pmem[id].quantum;
1116}
1117
1118static void bitmap_bits_set_all(uint32_t *bitp, int bit_start, int bit_end)
1119{
1120 int word_index = bit_start >> PMEM_32BIT_WORD_ORDER, total_words;
1121
1122 total_words = compute_total_words(bit_end, word_index);
1123 if (total_words > 0) {
1124 if (total_words == 1) {
1125 bitp[word_index] |=
1126 (start_mask(bit_start) & end_mask(bit_end));
1127 } else {
1128 bitp[word_index++] |= start_mask(bit_start);
1129 if (total_words > 2) {
1130 int total_bytes;
1131
1132 total_words -= 2;
1133 total_bytes = total_words << 2;
1134
1135 memset(&bitp[word_index], ~0, total_bytes);
1136 word_index += total_words;
1137 }
1138 bitp[word_index] |= end_mask(bit_end);
1139 }
1140 }
1141}
1142
1143static int
1144bitmap_allocate_contiguous(uint32_t *bitp, int num_bits_to_alloc,
Laura Abbott6b3eb1a2011-06-12 13:29:08 -07001145 int total_bits, int spacing, int start_bit)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001146{
1147 int bit_start, last_bit, word_index;
1148
1149 if (num_bits_to_alloc <= 0)
1150 return -1;
1151
Laura Abbott6b3eb1a2011-06-12 13:29:08 -07001152 for (bit_start = start_bit; ;
1153 bit_start = ((last_bit +
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001154 (word_index << PMEM_32BIT_WORD_ORDER) + spacing - 1)
Laura Abbott6b3eb1a2011-06-12 13:29:08 -07001155 & ~(spacing - 1)) + start_bit) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001156 int bit_end = bit_start + num_bits_to_alloc, total_words;
1157
1158 if (bit_end > total_bits)
1159 return -1; /* out of contiguous memory */
1160
1161 word_index = bit_start >> PMEM_32BIT_WORD_ORDER;
1162 total_words = compute_total_words(bit_end, word_index);
1163
1164 if (total_words <= 0)
1165 return -1;
1166
1167 if (total_words == 1) {
1168 last_bit = fls(bitp[word_index] &
1169 (start_mask(bit_start) &
1170 end_mask(bit_end)));
1171 if (last_bit)
1172 continue;
1173 } else {
1174 int end_word = word_index + (total_words - 1);
1175 last_bit =
1176 fls(bitp[word_index] & start_mask(bit_start));
1177 if (last_bit)
1178 continue;
1179
1180 for (word_index++;
1181 word_index < end_word;
1182 word_index++) {
1183 last_bit = fls(bitp[word_index]);
1184 if (last_bit)
1185 break;
1186 }
1187 if (last_bit)
1188 continue;
1189
1190 last_bit = fls(bitp[word_index] & end_mask(bit_end));
1191 if (last_bit)
1192 continue;
1193 }
1194 bitmap_bits_set_all(bitp, bit_start, bit_end);
1195 return bit_start;
1196 }
1197 return -1;
1198}
1199
1200static int reserve_quanta(const unsigned int quanta_needed,
1201 const int id,
1202 unsigned int align)
1203{
1204 /* alignment should be a valid power of 2 */
1205 int ret = -1, start_bit = 0, spacing = 1;
1206
1207 /* Sanity check */
1208 if (quanta_needed > pmem[id].allocator.bitmap.bitmap_free) {
1209#if PMEM_DEBUG
1210 printk(KERN_ALERT "pmem: %s: request (%d) too big for"
1211 " available free (%d)\n", __func__, quanta_needed,
1212 pmem[id].allocator.bitmap.bitmap_free);
1213#endif
1214 return -1;
1215 }
1216
1217 start_bit = bit_from_paddr(id,
1218 (pmem[id].base + align - 1) & ~(align - 1));
1219 if (start_bit <= -1) {
1220#if PMEM_DEBUG
1221 printk(KERN_ALERT
1222 "pmem: %s: bit_from_paddr fails for"
1223 " %u alignment.\n", __func__, align);
1224#endif
1225 return -1;
1226 }
1227 spacing = align / pmem[id].quantum;
1228 spacing = spacing > 1 ? spacing : 1;
1229
1230 ret = bitmap_allocate_contiguous(pmem[id].allocator.bitmap.bitmap,
1231 quanta_needed,
1232 (pmem[id].size + pmem[id].quantum - 1) / pmem[id].quantum,
Laura Abbott6b3eb1a2011-06-12 13:29:08 -07001233 spacing,
1234 start_bit);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001235
1236#if PMEM_DEBUG
1237 if (ret < 0)
1238 printk(KERN_ALERT "pmem: %s: not enough contiguous bits free "
1239 "in bitmap! Region memory is either too fragmented or"
1240 " request is too large for available memory.\n",
1241 __func__);
1242#endif
1243
1244 return ret;
1245}
1246
1247static int pmem_allocator_bitmap(const int id,
1248 const unsigned long len,
1249 const unsigned int align)
1250{
1251 /* caller should hold the lock on arena_mutex! */
1252 int bitnum, i;
1253 unsigned int quanta_needed;
1254
1255 DLOG("bitmap id %d, len %ld, align %u\n", id, len, align);
1256 if (!pmem[id].allocator.bitmap.bitm_alloc) {
1257#if PMEM_DEBUG
1258 printk(KERN_ALERT "pmem: bitm_alloc not present! id: %d\n",
1259 id);
1260#endif
1261 return -1;
1262 }
1263
1264 quanta_needed = (len + pmem[id].quantum - 1) / pmem[id].quantum;
1265 DLOG("quantum size %u quanta needed %u free %u id %d\n",
1266 pmem[id].quantum, quanta_needed,
1267 pmem[id].allocator.bitmap.bitmap_free, id);
1268
1269 if (pmem[id].allocator.bitmap.bitmap_free < quanta_needed) {
1270#if PMEM_DEBUG
1271 printk(KERN_ALERT "pmem: memory allocation failure. "
1272 "PMEM memory region exhausted, id %d."
1273 " Unable to comply with allocation request.\n", id);
1274#endif
1275 return -1;
1276 }
1277
1278 bitnum = reserve_quanta(quanta_needed, id, align);
1279 if (bitnum == -1)
1280 goto leave;
1281
1282 for (i = 0;
1283 i < pmem[id].allocator.bitmap.bitmap_allocs &&
1284 pmem[id].allocator.bitmap.bitm_alloc[i].bit != -1;
1285 i++)
1286 ;
1287
1288 if (i >= pmem[id].allocator.bitmap.bitmap_allocs) {
1289 void *temp;
1290 int32_t new_bitmap_allocs =
1291 pmem[id].allocator.bitmap.bitmap_allocs << 1;
1292 int j;
1293
1294 if (!new_bitmap_allocs) { /* failed sanity check!! */
1295#if PMEM_DEBUG
1296 pr_alert("pmem: bitmap_allocs number"
1297 " wrapped around to zero! Something "
1298 "is VERY wrong.\n");
1299#endif
1300 return -1;
1301 }
1302
1303 if (new_bitmap_allocs > pmem[id].num_entries) {
1304 /* failed sanity check!! */
1305#if PMEM_DEBUG
1306 pr_alert("pmem: required bitmap_allocs"
1307 " number exceeds maximum entries possible"
1308 " for current quanta\n");
1309#endif
1310 return -1;
1311 }
1312
1313 temp = krealloc(pmem[id].allocator.bitmap.bitm_alloc,
1314 new_bitmap_allocs *
1315 sizeof(*pmem[id].allocator.bitmap.bitm_alloc),
1316 GFP_KERNEL);
1317 if (!temp) {
1318#if PMEM_DEBUG
1319 pr_alert("pmem: can't realloc bitmap_allocs,"
1320 "id %d, current num bitmap allocs %d\n",
1321 id, pmem[id].allocator.bitmap.bitmap_allocs);
1322#endif
1323 return -1;
1324 }
1325 pmem[id].allocator.bitmap.bitmap_allocs = new_bitmap_allocs;
1326 pmem[id].allocator.bitmap.bitm_alloc = temp;
1327
1328 for (j = i; j < new_bitmap_allocs; j++) {
1329 pmem[id].allocator.bitmap.bitm_alloc[j].bit = -1;
1330 pmem[id].allocator.bitmap.bitm_alloc[i].quanta = 0;
1331 }
1332
1333 DLOG("increased # of allocated regions to %d for id %d\n",
1334 pmem[id].allocator.bitmap.bitmap_allocs, id);
1335 }
1336
1337 DLOG("bitnum %d, bitm_alloc index %d\n", bitnum, i);
1338
1339 pmem[id].allocator.bitmap.bitmap_free -= quanta_needed;
1340 pmem[id].allocator.bitmap.bitm_alloc[i].bit = bitnum;
1341 pmem[id].allocator.bitmap.bitm_alloc[i].quanta = quanta_needed;
1342leave:
1343 return bitnum;
1344}
1345
1346static int pmem_allocator_system(const int id,
1347 const unsigned long len,
1348 const unsigned int align)
1349{
1350 /* caller should hold the lock on arena_mutex! */
1351 struct alloc_list *list;
1352 unsigned long aligned_len;
1353 int count = SYSTEM_ALLOC_RETRY;
1354 void *buf;
1355
1356 DLOG("system id %d, len %ld, align %u\n", id, len, align);
1357
1358 if ((pmem[id].allocator.system_mem.used + len) > pmem[id].size) {
1359 DLOG("requested size would be larger than quota\n");
1360 return -1;
1361 }
1362
1363 /* Handle alignment */
1364 aligned_len = len + align;
1365
1366 /* Attempt allocation */
1367 list = kmalloc(sizeof(struct alloc_list), GFP_KERNEL);
1368 if (list == NULL) {
1369 printk(KERN_ERR "pmem: failed to allocate system metadata\n");
1370 return -1;
1371 }
1372 list->vaddr = NULL;
1373
1374 buf = NULL;
1375 while ((buf == NULL) && count--) {
1376 buf = kmalloc((aligned_len), GFP_KERNEL);
1377 if (buf == NULL) {
1378 DLOG("pmem: kmalloc %d temporarily failed len= %ld\n",
1379 count, aligned_len);
1380 }
1381 }
1382 if (!buf) {
1383 printk(KERN_CRIT "pmem: kmalloc failed for id= %d len= %ld\n",
1384 id, aligned_len);
1385 kfree(list);
1386 return -1;
1387 }
1388 list->size = aligned_len;
1389 list->addr = (void *)__pa(buf);
1390 list->aaddr = (void *)(((unsigned int)(list->addr) + (align - 1)) &
1391 ~(align - 1));
1392
1393 if (!pmem[id].cached)
1394 list->vaddr = ioremap(__pa(buf), aligned_len);
1395 else
1396 list->vaddr = ioremap_cached(__pa(buf), aligned_len);
1397
1398 INIT_LIST_HEAD(&list->allocs);
1399 list_add(&list->allocs, &pmem[id].allocator.system_mem.alist);
1400
1401 return (int)list;
1402}
1403
1404static pgprot_t pmem_phys_mem_access_prot(struct file *file, pgprot_t vma_prot)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001405{
1406 int id = get_id(file);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001407#ifdef pgprot_writecombine
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001408 if (pmem[id].cached == 0 || file->f_flags & O_SYNC)
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001409 /* on ARMv6 and ARMv7 this expands to Normal Noncached */
1410 return pgprot_writecombine(vma_prot);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001411#endif
1412#ifdef pgprot_ext_buffered
1413 else if (pmem[id].buffered)
1414 return pgprot_ext_buffered(vma_prot);
1415#endif
1416 return vma_prot;
1417}
1418
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001419static unsigned long pmem_start_addr_all_or_nothing(int id,
1420 struct pmem_data *data)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001421{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001422 return PMEM_START_ADDR(id, 0);
1423}
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001424
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001425static unsigned long pmem_start_addr_buddy_bestfit(int id,
1426 struct pmem_data *data)
1427{
1428 return PMEM_START_ADDR(id, data->index);
1429}
1430
1431static unsigned long pmem_start_addr_bitmap(int id, struct pmem_data *data)
1432{
1433 return data->index * pmem[id].quantum + pmem[id].base;
1434}
1435
1436static unsigned long pmem_start_addr_system(int id, struct pmem_data *data)
1437{
1438 return (unsigned long)(((struct alloc_list *)(data->index))->aaddr);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001439}
1440
1441static void *pmem_start_vaddr(int id, struct pmem_data *data)
1442{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001443 if (pmem[id].allocator_type == PMEM_ALLOCATORTYPE_SYSTEM)
1444 return ((struct alloc_list *)(data->index))->vaddr;
1445 else
1446 return pmem[id].start_addr(id, data) - pmem[id].base + pmem[id].vbase;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001447}
1448
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001449static unsigned long pmem_len_all_or_nothing(int id, struct pmem_data *data)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001450{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001451 return data->index;
1452}
1453
1454static unsigned long pmem_len_buddy_bestfit(int id, struct pmem_data *data)
1455{
1456 return PMEM_BUDDY_LEN(id, data->index);
1457}
1458
1459static unsigned long pmem_len_bitmap(int id, struct pmem_data *data)
1460{
1461 int i;
1462 unsigned long ret = 0;
1463
1464 mutex_lock(&pmem[id].arena_mutex);
1465
1466 for (i = 0; i < pmem[id].allocator.bitmap.bitmap_allocs; i++)
1467 if (pmem[id].allocator.bitmap.bitm_alloc[i].bit ==
1468 data->index) {
1469 ret = pmem[id].allocator.bitmap.bitm_alloc[i].quanta *
1470 pmem[id].quantum;
1471 break;
1472 }
1473
1474 mutex_unlock(&pmem[id].arena_mutex);
1475#if PMEM_DEBUG
1476 if (i >= pmem[id].allocator.bitmap.bitmap_allocs)
1477 pr_alert("pmem: %s: can't find bitnum %d in "
1478 "alloc'd array!\n", __func__, data->index);
1479#endif
1480 return ret;
1481}
1482
1483static unsigned long pmem_len_system(int id, struct pmem_data *data)
1484{
1485 unsigned long ret = 0;
1486
1487 mutex_lock(&pmem[id].arena_mutex);
1488
1489 ret = ((struct alloc_list *)data->index)->size;
1490 mutex_unlock(&pmem[id].arena_mutex);
1491
1492 return ret;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001493}
1494
1495static int pmem_map_garbage(int id, struct vm_area_struct *vma,
1496 struct pmem_data *data, unsigned long offset,
1497 unsigned long len)
1498{
1499 int i, garbage_pages = len >> PAGE_SHIFT;
1500
1501 vma->vm_flags |= VM_IO | VM_RESERVED | VM_PFNMAP | VM_SHARED | VM_WRITE;
1502 for (i = 0; i < garbage_pages; i++) {
1503 if (vm_insert_pfn(vma, vma->vm_start + offset + (i * PAGE_SIZE),
1504 pmem[id].garbage_pfn))
1505 return -EAGAIN;
1506 }
1507 return 0;
1508}
1509
1510static int pmem_unmap_pfn_range(int id, struct vm_area_struct *vma,
1511 struct pmem_data *data, unsigned long offset,
1512 unsigned long len)
1513{
1514 int garbage_pages;
1515 DLOG("unmap offset %lx len %lx\n", offset, len);
1516
1517 BUG_ON(!PMEM_IS_PAGE_ALIGNED(len));
1518
1519 garbage_pages = len >> PAGE_SHIFT;
1520 zap_page_range(vma, vma->vm_start + offset, len, NULL);
1521 pmem_map_garbage(id, vma, data, offset, len);
1522 return 0;
1523}
1524
1525static int pmem_map_pfn_range(int id, struct vm_area_struct *vma,
1526 struct pmem_data *data, unsigned long offset,
1527 unsigned long len)
1528{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001529 int ret;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001530 DLOG("map offset %lx len %lx\n", offset, len);
1531 BUG_ON(!PMEM_IS_PAGE_ALIGNED(vma->vm_start));
1532 BUG_ON(!PMEM_IS_PAGE_ALIGNED(vma->vm_end));
1533 BUG_ON(!PMEM_IS_PAGE_ALIGNED(len));
1534 BUG_ON(!PMEM_IS_PAGE_ALIGNED(offset));
1535
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001536 ret = io_remap_pfn_range(vma, vma->vm_start + offset,
1537 (pmem[id].start_addr(id, data) + offset) >> PAGE_SHIFT,
1538 len, vma->vm_page_prot);
1539 if (ret) {
1540#if PMEM_DEBUG
1541 pr_alert("pmem: %s: io_remap_pfn_range fails with "
1542 "return value: %d!\n", __func__, ret);
1543#endif
1544
1545 ret = -EAGAIN;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001546 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001547 return ret;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001548}
1549
1550static int pmem_remap_pfn_range(int id, struct vm_area_struct *vma,
1551 struct pmem_data *data, unsigned long offset,
1552 unsigned long len)
1553{
1554 /* hold the mm semp for the vma you are modifying when you call this */
1555 BUG_ON(!vma);
1556 zap_page_range(vma, vma->vm_start + offset, len, NULL);
1557 return pmem_map_pfn_range(id, vma, data, offset, len);
1558}
1559
1560static void pmem_vma_open(struct vm_area_struct *vma)
1561{
1562 struct file *file = vma->vm_file;
1563 struct pmem_data *data = file->private_data;
1564 int id = get_id(file);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001565
1566#if PMEM_DEBUG_MSGS
1567 char currtask_name[FIELD_SIZEOF(struct task_struct, comm) + 1];
1568#endif
1569 DLOG("Dev %s(id: %d) pid %u(%s) ppid %u file %p count %ld\n",
1570 get_name(file), id, current->pid,
1571 get_task_comm(currtask_name, current),
1572 current->parent->pid, file, file_count(file));
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001573 /* this should never be called as we don't support copying pmem
1574 * ranges via fork */
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001575 down_read(&data->sem);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001576 BUG_ON(!has_allocation(file));
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001577 /* remap the garbage pages, forkers don't get access to the data */
1578 pmem_unmap_pfn_range(id, vma, data, 0, vma->vm_start - vma->vm_end);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001579 up_read(&data->sem);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001580}
1581
1582static void pmem_vma_close(struct vm_area_struct *vma)
1583{
1584 struct file *file = vma->vm_file;
1585 struct pmem_data *data = file->private_data;
1586
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001587#if PMEM_DEBUG_MSGS
1588 char currtask_name[FIELD_SIZEOF(struct task_struct, comm) + 1];
1589#endif
1590 DLOG("Dev %s(id: %d) pid %u(%s) ppid %u file %p count %ld\n",
1591 get_name(file), get_id(file), current->pid,
1592 get_task_comm(currtask_name, current),
1593 current->parent->pid, file, file_count(file));
1594
1595 if (unlikely(!is_pmem_file(file))) {
1596 pr_warning("pmem: something is very wrong, you are "
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001597 "closing a vm backing an allocation that doesn't "
1598 "exist!\n");
1599 return;
1600 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001601
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001602 down_write(&data->sem);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001603 if (unlikely(!has_allocation(file))) {
1604 up_write(&data->sem);
1605 pr_warning("pmem: something is very wrong, you are "
1606 "closing a vm backing an allocation that doesn't "
1607 "exist!\n");
1608 return;
1609 }
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001610 if (data->vma == vma) {
1611 data->vma = NULL;
1612 if ((data->flags & PMEM_FLAGS_CONNECTED) &&
1613 (data->flags & PMEM_FLAGS_SUBMAP))
1614 data->flags |= PMEM_FLAGS_UNSUBMAP;
1615 }
1616 /* the kernel is going to free this vma now anyway */
1617 up_write(&data->sem);
1618}
1619
1620static struct vm_operations_struct vm_ops = {
1621 .open = pmem_vma_open,
1622 .close = pmem_vma_close,
1623};
1624
1625static int pmem_mmap(struct file *file, struct vm_area_struct *vma)
1626{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001627 struct pmem_data *data = file->private_data;
Laura Abbott1e36a022011-06-22 17:08:13 -07001628 int index = -1;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001629 unsigned long vma_size = vma->vm_end - vma->vm_start;
1630 int ret = 0, id = get_id(file);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001631#if PMEM_DEBUG_MSGS
1632 char currtask_name[FIELD_SIZEOF(struct task_struct, comm) + 1];
1633#endif
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001634
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001635 if (!data) {
1636 pr_err("pmem: Invalid file descriptor, no private data\n");
1637 return -EINVAL;
1638 }
1639 DLOG("pid %u(%s) mmap vma_size %lu on dev %s(id: %d)\n", current->pid,
1640 get_task_comm(currtask_name, current), vma_size,
1641 get_name(file), id);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001642 if (vma->vm_pgoff || !PMEM_IS_PAGE_ALIGNED(vma_size)) {
1643#if PMEM_DEBUG
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001644 pr_err("pmem: mmaps must be at offset zero, aligned"
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001645 " and a multiple of pages_size.\n");
1646#endif
1647 return -EINVAL;
1648 }
1649
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001650 down_write(&data->sem);
1651 /* check this file isn't already mmaped, for submaps check this file
1652 * has never been mmaped */
1653 if ((data->flags & PMEM_FLAGS_SUBMAP) ||
1654 (data->flags & PMEM_FLAGS_UNSUBMAP)) {
1655#if PMEM_DEBUG
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001656 pr_err("pmem: you can only mmap a pmem file once, "
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001657 "this file is already mmaped. %x\n", data->flags);
1658#endif
1659 ret = -EINVAL;
1660 goto error;
1661 }
1662 /* if file->private_data == unalloced, alloc*/
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001663 if (data->index == -1) {
1664 mutex_lock(&pmem[id].arena_mutex);
Laura Abbott1e36a022011-06-22 17:08:13 -07001665 index = pmem_allocate_from_id(id,
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001666 vma->vm_end - vma->vm_start,
1667 SZ_4K);
1668 mutex_unlock(&pmem[id].arena_mutex);
1669 /* either no space was available or an error occured */
1670 if (index == -1) {
1671 pr_err("pmem: mmap unable to allocate memory"
1672 "on %s\n", get_name(file));
1673 ret = -ENOMEM;
1674 goto error;
1675 }
1676 /* store the index of a successful allocation */
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001677 data->index = index;
1678 }
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001679
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001680 if (pmem[id].len(id, data) < vma_size) {
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001681#if PMEM_DEBUG
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001682 pr_err("pmem: mmap size [%lu] does not match"
1683 " size of backing region [%lu].\n", vma_size,
1684 pmem[id].len(id, data));
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001685#endif
1686 ret = -EINVAL;
1687 goto error;
1688 }
1689
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001690 vma->vm_pgoff = pmem[id].start_addr(id, data) >> PAGE_SHIFT;
1691
1692 vma->vm_page_prot = pmem_phys_mem_access_prot(file, vma->vm_page_prot);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001693
1694 if (data->flags & PMEM_FLAGS_CONNECTED) {
1695 struct pmem_region_node *region_node;
1696 struct list_head *elt;
1697 if (pmem_map_garbage(id, vma, data, 0, vma_size)) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001698 pr_alert("pmem: mmap failed in kernel!\n");
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001699 ret = -EAGAIN;
1700 goto error;
1701 }
1702 list_for_each(elt, &data->region_list) {
1703 region_node = list_entry(elt, struct pmem_region_node,
1704 list);
1705 DLOG("remapping file: %p %lx %lx\n", file,
1706 region_node->region.offset,
1707 region_node->region.len);
1708 if (pmem_remap_pfn_range(id, vma, data,
1709 region_node->region.offset,
1710 region_node->region.len)) {
1711 ret = -EAGAIN;
1712 goto error;
1713 }
1714 }
1715 data->flags |= PMEM_FLAGS_SUBMAP;
1716 get_task_struct(current->group_leader);
1717 data->task = current->group_leader;
1718 data->vma = vma;
1719#if PMEM_DEBUG
1720 data->pid = current->pid;
1721#endif
1722 DLOG("submmapped file %p vma %p pid %u\n", file, vma,
1723 current->pid);
1724 } else {
1725 if (pmem_map_pfn_range(id, vma, data, 0, vma_size)) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001726 pr_err("pmem: mmap failed in kernel!\n");
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001727 ret = -EAGAIN;
1728 goto error;
1729 }
1730 data->flags |= PMEM_FLAGS_MASTERMAP;
1731 data->pid = current->pid;
1732 }
1733 vma->vm_ops = &vm_ops;
1734error:
1735 up_write(&data->sem);
1736 return ret;
1737}
1738
1739/* the following are the api for accessing pmem regions by other drivers
1740 * from inside the kernel */
1741int get_pmem_user_addr(struct file *file, unsigned long *start,
1742 unsigned long *len)
1743{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001744 int ret = -1;
1745
1746 if (is_pmem_file(file)) {
1747 struct pmem_data *data = file->private_data;
1748
1749 down_read(&data->sem);
1750 if (has_allocation(file)) {
1751 if (data->vma) {
1752 *start = data->vma->vm_start;
1753 *len = data->vma->vm_end - data->vma->vm_start;
1754 } else {
1755 *start = *len = 0;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001756#if PMEM_DEBUG
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001757 pr_err("pmem: %s: no vma present.\n",
1758 __func__);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001759#endif
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001760 }
1761 ret = 0;
1762 }
1763 up_read(&data->sem);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001764 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001765
1766#if PMEM_DEBUG
1767 if (ret)
1768 pr_err("pmem: %s: requested pmem data from invalid"
1769 "file.\n", __func__);
1770#endif
1771 return ret;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001772}
1773
1774int get_pmem_addr(struct file *file, unsigned long *start,
1775 unsigned long *vstart, unsigned long *len)
1776{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001777 int ret = -1;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001778
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001779 if (is_pmem_file(file)) {
1780 struct pmem_data *data = file->private_data;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001781
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001782 down_read(&data->sem);
1783 if (has_allocation(file)) {
1784 int id = get_id(file);
1785
1786 *start = pmem[id].start_addr(id, data);
1787 *len = pmem[id].len(id, data);
1788 *vstart = (unsigned long)
1789 pmem_start_vaddr(id, data);
1790 up_read(&data->sem);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001791#if PMEM_DEBUG
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001792 down_write(&data->sem);
1793 data->ref++;
1794 up_write(&data->sem);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001795#endif
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001796 DLOG("returning start %#lx len %lu "
1797 "vstart %#lx\n",
1798 *start, *len, *vstart);
1799 ret = 0;
1800 } else {
1801 up_read(&data->sem);
1802 }
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001803 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001804 return ret;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001805}
1806
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001807int get_pmem_file(unsigned int fd, unsigned long *start, unsigned long *vstart,
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001808 unsigned long *len, struct file **filp)
1809{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001810 int ret = -1;
1811 struct file *file = fget(fd);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001812
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001813 if (unlikely(file == NULL)) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001814 pr_err("pmem: %s: requested data from file "
1815 "descriptor that doesn't exist.\n", __func__);
1816 } else {
1817#if PMEM_DEBUG_MSGS
1818 char currtask_name[FIELD_SIZEOF(struct task_struct, comm) + 1];
1819#endif
1820 DLOG("filp %p rdev %d pid %u(%s) file %p(%ld)"
1821 " dev %s(id: %d)\n", filp,
1822 file->f_dentry->d_inode->i_rdev,
1823 current->pid, get_task_comm(currtask_name, current),
1824 file, file_count(file), get_name(file), get_id(file));
1825
1826 if (!get_pmem_addr(file, start, vstart, len)) {
1827 if (filp)
1828 *filp = file;
1829 ret = 0;
1830 } else {
1831 fput(file);
1832 }
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001833 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001834 return ret;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001835}
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001836EXPORT_SYMBOL(get_pmem_file);
1837
1838int get_pmem_fd(int fd, unsigned long *start, unsigned long *len)
1839{
1840 unsigned long vstart;
1841 return get_pmem_file(fd, start, &vstart, len, NULL);
1842}
1843EXPORT_SYMBOL(get_pmem_fd);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001844
1845void put_pmem_file(struct file *file)
1846{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001847#if PMEM_DEBUG_MSGS
1848 char currtask_name[FIELD_SIZEOF(struct task_struct, comm) + 1];
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001849#endif
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001850 DLOG("rdev %d pid %u(%s) file %p(%ld)" " dev %s(id: %d)\n",
1851 file->f_dentry->d_inode->i_rdev, current->pid,
1852 get_task_comm(currtask_name, current), file,
1853 file_count(file), get_name(file), get_id(file));
1854 if (is_pmem_file(file)) {
1855#if PMEM_DEBUG
1856 struct pmem_data *data = file->private_data;
1857
1858 down_write(&data->sem);
1859 if (!data->ref--) {
1860 data->ref++;
1861 pr_alert("pmem: pmem_put > pmem_get %s "
1862 "(pid %d)\n",
1863 pmem[get_id(file)].dev.name, data->pid);
1864 BUG();
1865 }
1866 up_write(&data->sem);
1867#endif
1868 fput(file);
1869 }
1870}
1871EXPORT_SYMBOL(put_pmem_file);
1872
1873void put_pmem_fd(int fd)
1874{
1875 int put_needed;
1876 struct file *file = fget_light(fd, &put_needed);
1877
1878 if (file) {
1879 put_pmem_file(file);
1880 fput_light(file, put_needed);
1881 }
1882}
1883
1884void flush_pmem_fd(int fd, unsigned long offset, unsigned long len)
1885{
1886 int fput_needed;
1887 struct file *file = fget_light(fd, &fput_needed);
1888
1889 if (file) {
1890 flush_pmem_file(file, offset, len);
1891 fput_light(file, fput_needed);
1892 }
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001893}
1894
1895void flush_pmem_file(struct file *file, unsigned long offset, unsigned long len)
1896{
1897 struct pmem_data *data;
1898 int id;
1899 void *vaddr;
1900 struct pmem_region_node *region_node;
1901 struct list_head *elt;
1902 void *flush_start, *flush_end;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001903#ifdef CONFIG_OUTER_CACHE
1904 unsigned long phy_start, phy_end;
1905#endif
1906 if (!is_pmem_file(file))
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001907 return;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001908
1909 id = get_id(file);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001910 if (!pmem[id].cached)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001911 return;
1912
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001913 /* is_pmem_file fails if !file */
1914 data = file->private_data;
1915
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001916 down_read(&data->sem);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001917 if (!has_allocation(file))
1918 goto end;
1919
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001920 vaddr = pmem_start_vaddr(id, data);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001921
1922 if (pmem[id].allocator_type == PMEM_ALLOCATORTYPE_SYSTEM) {
1923 dmac_flush_range(vaddr,
1924 (void *)((unsigned long)vaddr +
1925 ((struct alloc_list *)(data->index))->size));
1926#ifdef CONFIG_OUTER_CACHE
1927 phy_start = pmem_start_addr_system(id, data);
1928
1929 phy_end = phy_start +
1930 ((struct alloc_list *)(data->index))->size;
1931
1932 outer_flush_range(phy_start, phy_end);
1933#endif
1934 goto end;
1935 }
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001936 /* if this isn't a submmapped file, flush the whole thing */
1937 if (unlikely(!(data->flags & PMEM_FLAGS_CONNECTED))) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001938 dmac_flush_range(vaddr, vaddr + pmem[id].len(id, data));
1939#ifdef CONFIG_OUTER_CACHE
1940 phy_start = (unsigned long)vaddr -
1941 (unsigned long)pmem[id].vbase + pmem[id].base;
1942
1943 phy_end = phy_start + pmem[id].len(id, data);
1944
1945 outer_flush_range(phy_start, phy_end);
1946#endif
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001947 goto end;
1948 }
1949 /* otherwise, flush the region of the file we are drawing */
1950 list_for_each(elt, &data->region_list) {
1951 region_node = list_entry(elt, struct pmem_region_node, list);
1952 if ((offset >= region_node->region.offset) &&
1953 ((offset + len) <= (region_node->region.offset +
1954 region_node->region.len))) {
1955 flush_start = vaddr + region_node->region.offset;
1956 flush_end = flush_start + region_node->region.len;
1957 dmac_flush_range(flush_start, flush_end);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001958#ifdef CONFIG_OUTER_CACHE
1959
1960 phy_start = (unsigned long)flush_start -
1961 (unsigned long)pmem[id].vbase + pmem[id].base;
1962
1963 phy_end = phy_start + region_node->region.len;
1964
1965 outer_flush_range(phy_start, phy_end);
1966#endif
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07001967 break;
1968 }
1969 }
1970end:
1971 up_read(&data->sem);
1972}
1973
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001974int pmem_cache_maint(struct file *file, unsigned int cmd,
1975 struct pmem_addr *pmem_addr)
1976{
1977 struct pmem_data *data;
1978 int id;
1979 unsigned long vaddr, paddr, length, offset,
1980 pmem_len, pmem_start_addr;
1981
1982 /* Called from kernel-space so file may be NULL */
1983 if (!file)
1984 return -EBADF;
1985
Shubhraprakash Das7788cad2011-11-21 13:02:22 -07001986 /*
1987 * check that the vaddr passed for flushing is valid
1988 * so that you don't crash the kernel
1989 */
1990 if (!pmem_addr->vaddr)
1991 return -EINVAL;
1992
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07001993 data = file->private_data;
1994 id = get_id(file);
1995
1996 if (!pmem[id].cached)
1997 return 0;
1998
1999 offset = pmem_addr->offset;
2000 length = pmem_addr->length;
2001
2002 down_read(&data->sem);
2003 if (!has_allocation(file)) {
2004 up_read(&data->sem);
2005 return -EINVAL;
2006 }
2007 pmem_len = pmem[id].len(id, data);
2008 pmem_start_addr = pmem[id].start_addr(id, data);
2009 up_read(&data->sem);
2010
2011 if (offset + length > pmem_len)
2012 return -EINVAL;
2013
2014 vaddr = pmem_addr->vaddr;
2015 paddr = pmem_start_addr + offset;
2016
2017 DLOG("pmem cache maint on dev %s(id: %d)"
2018 "(vaddr %lx paddr %lx len %lu bytes)\n",
2019 get_name(file), id, vaddr, paddr, length);
2020 if (cmd == PMEM_CLEAN_INV_CACHES)
2021 clean_and_invalidate_caches(vaddr,
2022 length, paddr);
2023 else if (cmd == PMEM_CLEAN_CACHES)
2024 clean_caches(vaddr, length, paddr);
2025 else if (cmd == PMEM_INV_CACHES)
2026 invalidate_caches(vaddr, length, paddr);
2027
2028 return 0;
2029}
2030EXPORT_SYMBOL(pmem_cache_maint);
2031
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002032static int pmem_connect(unsigned long connect, struct file *file)
2033{
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002034 int ret = 0, put_needed;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002035 struct file *src_file;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002036
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002037 if (!file) {
2038 pr_err("pmem: %s: NULL file pointer passed in, "
2039 "bailing out!\n", __func__);
2040 ret = -EINVAL;
2041 goto leave;
2042 }
2043
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002044 src_file = fget_light(connect, &put_needed);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002045
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002046 if (!src_file) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002047 pr_err("pmem: %s: src file not found!\n", __func__);
2048 ret = -EBADF;
2049 goto leave;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002050 }
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002051
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002052 if (src_file == file) { /* degenerative case, operator error */
2053 pr_err("pmem: %s: src_file and passed in file are "
2054 "the same; refusing to connect to self!\n", __func__);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002055 ret = -EINVAL;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002056 goto put_src_file;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002057 }
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002058
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002059 if (unlikely(!is_pmem_file(src_file))) {
2060 pr_err("pmem: %s: src file is not a pmem file!\n",
2061 __func__);
2062 ret = -EINVAL;
2063 goto put_src_file;
2064 } else {
2065 struct pmem_data *src_data = src_file->private_data;
2066
2067 if (!src_data) {
2068 pr_err("pmem: %s: src file pointer has no"
2069 "private data, bailing out!\n", __func__);
2070 ret = -EINVAL;
2071 goto put_src_file;
2072 }
2073
2074 down_read(&src_data->sem);
2075
2076 if (unlikely(!has_allocation(src_file))) {
2077 up_read(&src_data->sem);
2078 pr_err("pmem: %s: src file has no allocation!\n",
2079 __func__);
2080 ret = -EINVAL;
2081 } else {
2082 struct pmem_data *data;
2083 int src_index = src_data->index;
2084
2085 up_read(&src_data->sem);
2086
2087 data = file->private_data;
2088 if (!data) {
2089 pr_err("pmem: %s: passed in file "
2090 "pointer has no private data, bailing"
2091 " out!\n", __func__);
2092 ret = -EINVAL;
2093 goto put_src_file;
2094 }
2095
2096 down_write(&data->sem);
2097 if (has_allocation(file) &&
2098 (data->index != src_index)) {
2099 up_write(&data->sem);
2100
2101 pr_err("pmem: %s: file is already "
2102 "mapped but doesn't match this "
2103 "src_file!\n", __func__);
2104 ret = -EINVAL;
2105 } else {
2106 data->index = src_index;
2107 data->flags |= PMEM_FLAGS_CONNECTED;
2108 data->master_fd = connect;
2109 data->master_file = src_file;
2110
2111 up_write(&data->sem);
2112
2113 DLOG("connect %p to %p\n", file, src_file);
2114 }
2115 }
2116 }
2117put_src_file:
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002118 fput_light(src_file, put_needed);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002119leave:
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002120 return ret;
2121}
2122
2123static void pmem_unlock_data_and_mm(struct pmem_data *data,
2124 struct mm_struct *mm)
2125{
2126 up_write(&data->sem);
2127 if (mm != NULL) {
2128 up_write(&mm->mmap_sem);
2129 mmput(mm);
2130 }
2131}
2132
2133static int pmem_lock_data_and_mm(struct file *file, struct pmem_data *data,
2134 struct mm_struct **locked_mm)
2135{
2136 int ret = 0;
2137 struct mm_struct *mm = NULL;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002138#if PMEM_DEBUG_MSGS
2139 char currtask_name[FIELD_SIZEOF(struct task_struct, comm) + 1];
2140#endif
2141 DLOG("pid %u(%s) file %p(%ld)\n",
2142 current->pid, get_task_comm(currtask_name, current),
2143 file, file_count(file));
2144
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002145 *locked_mm = NULL;
2146lock_mm:
2147 down_read(&data->sem);
2148 if (PMEM_IS_SUBMAP(data)) {
2149 mm = get_task_mm(data->task);
2150 if (!mm) {
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002151 up_read(&data->sem);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002152#if PMEM_DEBUG
2153 pr_alert("pmem: can't remap - task is gone!\n");
2154#endif
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002155 return -1;
2156 }
2157 }
2158 up_read(&data->sem);
2159
2160 if (mm)
2161 down_write(&mm->mmap_sem);
2162
2163 down_write(&data->sem);
2164 /* check that the file didn't get mmaped before we could take the
2165 * data sem, this should be safe b/c you can only submap each file
2166 * once */
2167 if (PMEM_IS_SUBMAP(data) && !mm) {
2168 pmem_unlock_data_and_mm(data, mm);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002169 DLOG("mapping contention, repeating mmap op\n");
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002170 goto lock_mm;
2171 }
2172 /* now check that vma.mm is still there, it could have been
2173 * deleted by vma_close before we could get the data->sem */
2174 if ((data->flags & PMEM_FLAGS_UNSUBMAP) && (mm != NULL)) {
2175 /* might as well release this */
2176 if (data->flags & PMEM_FLAGS_SUBMAP) {
2177 put_task_struct(data->task);
2178 data->task = NULL;
2179 /* lower the submap flag to show the mm is gone */
2180 data->flags &= ~(PMEM_FLAGS_SUBMAP);
2181 }
2182 pmem_unlock_data_and_mm(data, mm);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002183#if PMEM_DEBUG
2184 pr_alert("pmem: vma.mm went away!\n");
2185#endif
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002186 return -1;
2187 }
2188 *locked_mm = mm;
2189 return ret;
2190}
2191
2192int pmem_remap(struct pmem_region *region, struct file *file,
2193 unsigned operation)
2194{
2195 int ret;
2196 struct pmem_region_node *region_node;
2197 struct mm_struct *mm = NULL;
2198 struct list_head *elt, *elt2;
2199 int id = get_id(file);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002200 struct pmem_data *data;
2201
2202 DLOG("operation %#x, region offset %ld, region len %ld\n",
2203 operation, region->offset, region->len);
2204
2205 if (!is_pmem_file(file)) {
2206#if PMEM_DEBUG
2207 pr_err("pmem: remap request for non-pmem file descriptor\n");
2208#endif
2209 return -EINVAL;
2210 }
2211
2212 /* is_pmem_file fails if !file */
2213 data = file->private_data;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002214
2215 /* pmem region must be aligned on a page boundry */
2216 if (unlikely(!PMEM_IS_PAGE_ALIGNED(region->offset) ||
2217 !PMEM_IS_PAGE_ALIGNED(region->len))) {
2218#if PMEM_DEBUG
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002219 pr_err("pmem: request for unaligned pmem"
2220 "suballocation %lx %lx\n",
2221 region->offset, region->len);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002222#endif
2223 return -EINVAL;
2224 }
2225
2226 /* if userspace requests a region of len 0, there's nothing to do */
2227 if (region->len == 0)
2228 return 0;
2229
2230 /* lock the mm and data */
2231 ret = pmem_lock_data_and_mm(file, data, &mm);
2232 if (ret)
2233 return 0;
2234
2235 /* only the owner of the master file can remap the client fds
2236 * that back in it */
2237 if (!is_master_owner(file)) {
2238#if PMEM_DEBUG
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002239 pr_err("pmem: remap requested from non-master process\n");
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002240#endif
2241 ret = -EINVAL;
2242 goto err;
2243 }
2244
2245 /* check that the requested range is within the src allocation */
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002246 if (unlikely((region->offset > pmem[id].len(id, data)) ||
2247 (region->len > pmem[id].len(id, data)) ||
2248 (region->offset + region->len > pmem[id].len(id, data)))) {
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002249#if PMEM_DEBUG
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002250 pr_err("pmem: suballoc doesn't fit in src_file!\n");
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002251#endif
2252 ret = -EINVAL;
2253 goto err;
2254 }
2255
2256 if (operation == PMEM_MAP) {
2257 region_node = kmalloc(sizeof(struct pmem_region_node),
2258 GFP_KERNEL);
2259 if (!region_node) {
2260 ret = -ENOMEM;
2261#if PMEM_DEBUG
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002262 pr_alert("pmem: No space to allocate remap metadata!");
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002263#endif
2264 goto err;
2265 }
2266 region_node->region = *region;
2267 list_add(&region_node->list, &data->region_list);
2268 } else if (operation == PMEM_UNMAP) {
2269 int found = 0;
2270 list_for_each_safe(elt, elt2, &data->region_list) {
2271 region_node = list_entry(elt, struct pmem_region_node,
2272 list);
2273 if (region->len == 0 ||
2274 (region_node->region.offset == region->offset &&
2275 region_node->region.len == region->len)) {
2276 list_del(elt);
2277 kfree(region_node);
2278 found = 1;
2279 }
2280 }
2281 if (!found) {
2282#if PMEM_DEBUG
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002283 pr_err("pmem: Unmap region does not map any"
2284 " mapped region!");
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002285#endif
2286 ret = -EINVAL;
2287 goto err;
2288 }
2289 }
2290
2291 if (data->vma && PMEM_IS_SUBMAP(data)) {
2292 if (operation == PMEM_MAP)
2293 ret = pmem_remap_pfn_range(id, data->vma, data,
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002294 region->offset, region->len);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002295 else if (operation == PMEM_UNMAP)
2296 ret = pmem_unmap_pfn_range(id, data->vma, data,
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002297 region->offset, region->len);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002298 }
2299
2300err:
2301 pmem_unlock_data_and_mm(data, mm);
2302 return ret;
2303}
2304
2305static void pmem_revoke(struct file *file, struct pmem_data *data)
2306{
2307 struct pmem_region_node *region_node;
2308 struct list_head *elt, *elt2;
2309 struct mm_struct *mm = NULL;
2310 int id = get_id(file);
2311 int ret = 0;
2312
2313 data->master_file = NULL;
2314 ret = pmem_lock_data_and_mm(file, data, &mm);
2315 /* if lock_data_and_mm fails either the task that mapped the fd, or
2316 * the vma that mapped it have already gone away, nothing more
2317 * needs to be done */
2318 if (ret)
2319 return;
2320 /* unmap everything */
2321 /* delete the regions and region list nothing is mapped any more */
2322 if (data->vma)
2323 list_for_each_safe(elt, elt2, &data->region_list) {
2324 region_node = list_entry(elt, struct pmem_region_node,
2325 list);
2326 pmem_unmap_pfn_range(id, data->vma, data,
2327 region_node->region.offset,
2328 region_node->region.len);
2329 list_del(elt);
2330 kfree(region_node);
2331 }
2332 /* delete the master file */
2333 pmem_unlock_data_and_mm(data, mm);
2334}
2335
2336static void pmem_get_size(struct pmem_region *region, struct file *file)
2337{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002338 /* called via ioctl file op, so file guaranteed to be not NULL */
2339 struct pmem_data *data = file->private_data;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002340 int id = get_id(file);
2341
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002342 down_read(&data->sem);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002343 if (!has_allocation(file)) {
2344 region->offset = 0;
2345 region->len = 0;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002346 } else {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002347 region->offset = pmem[id].start_addr(id, data);
2348 region->len = pmem[id].len(id, data);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002349 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002350 up_read(&data->sem);
2351 DLOG("offset 0x%lx len 0x%lx\n", region->offset, region->len);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002352}
2353
2354
2355static long pmem_ioctl(struct file *file, unsigned int cmd, unsigned long arg)
2356{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002357 /* called from user space as file op, so file guaranteed to be not
2358 * NULL
2359 */
2360 struct pmem_data *data = file->private_data;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002361 int id = get_id(file);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002362#if PMEM_DEBUG_MSGS
2363 char currtask_name[
2364 FIELD_SIZEOF(struct task_struct, comm) + 1];
2365#endif
2366
2367 DLOG("pid %u(%s) file %p(%ld) cmd %#x, dev %s(id: %d)\n",
2368 current->pid, get_task_comm(currtask_name, current),
2369 file, file_count(file), cmd, get_name(file), id);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002370
2371 switch (cmd) {
2372 case PMEM_GET_PHYS:
2373 {
2374 struct pmem_region region;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002375
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002376 DLOG("get_phys\n");
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002377 down_read(&data->sem);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002378 if (!has_allocation(file)) {
2379 region.offset = 0;
2380 region.len = 0;
2381 } else {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002382 region.offset = pmem[id].start_addr(id, data);
2383 region.len = pmem[id].len(id, data);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002384 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002385 up_read(&data->sem);
2386
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002387 if (copy_to_user((void __user *)arg, &region,
2388 sizeof(struct pmem_region)))
2389 return -EFAULT;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002390
2391 DLOG("pmem: successful request for "
2392 "physical address of pmem region id %d, "
2393 "offset 0x%lx, len 0x%lx\n",
2394 id, region.offset, region.len);
2395
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002396 break;
2397 }
2398 case PMEM_MAP:
2399 {
2400 struct pmem_region region;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002401 DLOG("map\n");
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002402 if (copy_from_user(&region, (void __user *)arg,
2403 sizeof(struct pmem_region)))
2404 return -EFAULT;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002405 return pmem_remap(&region, file, PMEM_MAP);
2406 }
2407 break;
2408 case PMEM_UNMAP:
2409 {
2410 struct pmem_region region;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002411 DLOG("unmap\n");
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002412 if (copy_from_user(&region, (void __user *)arg,
2413 sizeof(struct pmem_region)))
2414 return -EFAULT;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002415 return pmem_remap(&region, file, PMEM_UNMAP);
2416 break;
2417 }
2418 case PMEM_GET_SIZE:
2419 {
2420 struct pmem_region region;
2421 DLOG("get_size\n");
2422 pmem_get_size(&region, file);
2423 if (copy_to_user((void __user *)arg, &region,
2424 sizeof(struct pmem_region)))
2425 return -EFAULT;
2426 break;
2427 }
2428 case PMEM_GET_TOTAL_SIZE:
2429 {
2430 struct pmem_region region;
2431 DLOG("get total size\n");
2432 region.offset = 0;
2433 get_id(file);
2434 region.len = pmem[id].size;
2435 if (copy_to_user((void __user *)arg, &region,
2436 sizeof(struct pmem_region)))
2437 return -EFAULT;
2438 break;
2439 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002440 case PMEM_GET_FREE_SPACE:
2441 {
2442 struct pmem_freespace fs;
2443 DLOG("get freespace on %s(id: %d)\n",
2444 get_name(file), id);
2445
2446 mutex_lock(&pmem[id].arena_mutex);
2447 pmem[id].free_space(id, &fs);
2448 mutex_unlock(&pmem[id].arena_mutex);
2449
2450 DLOG("%s(id: %d) total free %lu, largest %lu\n",
2451 get_name(file), id, fs.total, fs.largest);
2452
2453 if (copy_to_user((void __user *)arg, &fs,
2454 sizeof(struct pmem_freespace)))
2455 return -EFAULT;
2456 break;
2457 }
2458
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002459 case PMEM_ALLOCATE:
2460 {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002461 int ret = 0;
2462 DLOG("allocate, id %d\n", id);
2463 down_write(&data->sem);
2464 if (has_allocation(file)) {
2465 pr_err("pmem: Existing allocation found on "
2466 "this file descrpitor\n");
2467 up_write(&data->sem);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002468 return -EINVAL;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002469 }
2470
2471 mutex_lock(&pmem[id].arena_mutex);
Laura Abbott1e36a022011-06-22 17:08:13 -07002472 data->index = pmem_allocate_from_id(id,
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002473 arg,
2474 SZ_4K);
2475 mutex_unlock(&pmem[id].arena_mutex);
2476 ret = data->index == -1 ? -ENOMEM :
2477 data->index;
2478 up_write(&data->sem);
2479 return ret;
2480 }
2481 case PMEM_ALLOCATE_ALIGNED:
2482 {
2483 struct pmem_allocation alloc;
2484 int ret = 0;
2485
2486 if (copy_from_user(&alloc, (void __user *)arg,
2487 sizeof(struct pmem_allocation)))
2488 return -EFAULT;
2489 DLOG("allocate id align %d %u\n", id, alloc.align);
2490 down_write(&data->sem);
2491 if (has_allocation(file)) {
2492 pr_err("pmem: Existing allocation found on "
2493 "this file descrpitor\n");
2494 up_write(&data->sem);
2495 return -EINVAL;
2496 }
2497
2498 if (alloc.align & (alloc.align - 1)) {
2499 pr_err("pmem: Alignment is not a power of 2\n");
2500 return -EINVAL;
2501 }
2502
2503 if (alloc.align != SZ_4K &&
2504 (pmem[id].allocator_type !=
2505 PMEM_ALLOCATORTYPE_BITMAP)) {
2506 pr_err("pmem: Non 4k alignment requires bitmap"
2507 " allocator on %s\n", pmem[id].name);
2508 return -EINVAL;
2509 }
2510
2511 if (alloc.align > SZ_1M ||
2512 alloc.align < SZ_4K) {
2513 pr_err("pmem: Invalid Alignment (%u) "
2514 "specified\n", alloc.align);
2515 return -EINVAL;
2516 }
2517
2518 mutex_lock(&pmem[id].arena_mutex);
Laura Abbott1e36a022011-06-22 17:08:13 -07002519 data->index = pmem_allocate_from_id(id,
2520 alloc.size,
2521 alloc.align);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002522 mutex_unlock(&pmem[id].arena_mutex);
2523 ret = data->index == -1 ? -ENOMEM :
2524 data->index;
2525 up_write(&data->sem);
2526 return ret;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002527 }
2528 case PMEM_CONNECT:
2529 DLOG("connect\n");
2530 return pmem_connect(arg, file);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002531 case PMEM_CLEAN_INV_CACHES:
2532 case PMEM_CLEAN_CACHES:
2533 case PMEM_INV_CACHES:
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002534 {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002535 struct pmem_addr pmem_addr;
2536
2537 if (copy_from_user(&pmem_addr, (void __user *)arg,
2538 sizeof(struct pmem_addr)))
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002539 return -EFAULT;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002540
2541 return pmem_cache_maint(file, cmd, &pmem_addr);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002542 }
2543 default:
2544 if (pmem[id].ioctl)
2545 return pmem[id].ioctl(file, cmd, arg);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002546
2547 DLOG("ioctl invalid (%#x)\n", cmd);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002548 return -EINVAL;
2549 }
2550 return 0;
2551}
2552
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002553static void ioremap_pmem(int id)
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002554{
Naveen Ramaraj189f1882011-08-16 17:39:22 -07002555 unsigned long addr;
2556 const struct mem_type *type;
Laura Abbott1e36a022011-06-22 17:08:13 -07002557
Naveen Ramaraj189f1882011-08-16 17:39:22 -07002558 DLOG("PMEMDEBUG: ioremaping for %s\n", pmem[id].name);
2559 if (pmem[id].map_on_demand) {
2560 addr = (unsigned long)pmem[id].area->addr;
2561 if (pmem[id].cached)
2562 type = get_mem_type(MT_DEVICE_CACHED);
2563 else
2564 type = get_mem_type(MT_DEVICE);
2565 DLOG("PMEMDEBUG: Remap phys %lx to virt %lx on %s\n",
2566 pmem[id].base, addr, pmem[id].name);
2567 if (ioremap_page_range(addr, addr + pmem[id].size,
2568 pmem[id].base, __pgprot(type->prot_pte))) {
2569 pr_err("pmem: Failed to map pages\n");
2570 BUG();
2571 }
2572 pmem[id].vbase = pmem[id].area->addr;
2573 /* Flush the cache after installing page table entries to avoid
2574 * aliasing when these pages are remapped to user space.
2575 */
2576 flush_cache_vmap(addr, addr + pmem[id].size);
2577 } else {
2578 if (pmem[id].cached)
2579 pmem[id].vbase = ioremap_cached(pmem[id].base,
2580 pmem[id].size);
2581 #ifdef ioremap_ext_buffered
2582 else if (pmem[id].buffered)
2583 pmem[id].vbase = ioremap_ext_buffered(pmem[id].base,
2584 pmem[id].size);
2585 #endif
2586 else
2587 pmem[id].vbase = ioremap(pmem[id].base, pmem[id].size);
2588 }
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002589}
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002590
2591int pmem_setup(struct android_pmem_platform_data *pdata,
2592 long (*ioctl)(struct file *, unsigned int, unsigned long),
2593 int (*release)(struct inode *, struct file *))
2594{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002595 int i, index = 0, id;
Naveen Ramaraj189f1882011-08-16 17:39:22 -07002596 struct vm_struct *pmem_vma = NULL;
Vipul Gandhif752bf62012-01-09 15:34:04 -08002597 struct page *page;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002598
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002599 if (id_count >= PMEM_MAX_DEVICES) {
2600 pr_alert("pmem: %s: unable to register driver(%s) - no more "
2601 "devices available!\n", __func__, pdata->name);
2602 goto err_no_mem;
2603 }
2604
2605 if (!pdata->size) {
2606 pr_alert("pmem: %s: unable to register pmem driver(%s) - zero "
2607 "size passed in!\n", __func__, pdata->name);
2608 goto err_no_mem;
2609 }
2610
2611 id = id_count++;
2612
2613 pmem[id].id = id;
2614
2615 if (pmem[id].allocate) {
2616 pr_alert("pmem: %s: unable to register pmem driver - "
2617 "duplicate registration of %s!\n",
2618 __func__, pdata->name);
2619 goto err_no_mem;
2620 }
2621
2622 pmem[id].allocator_type = pdata->allocator_type;
2623
2624 /* 'quantum' is a "hidden" variable that defaults to 0 in the board
2625 * files */
2626 pmem[id].quantum = pdata->quantum ?: PMEM_MIN_ALLOC;
2627 if (pmem[id].quantum < PMEM_MIN_ALLOC ||
2628 !is_power_of_2(pmem[id].quantum)) {
2629 pr_alert("pmem: %s: unable to register pmem driver %s - "
2630 "invalid quantum value (%#x)!\n",
2631 __func__, pdata->name, pmem[id].quantum);
2632 goto err_reset_pmem_info;
2633 }
2634
2635 if (pdata->size % pmem[id].quantum) {
2636 /* bad alignment for size! */
2637 pr_alert("pmem: %s: Unable to register driver %s - "
2638 "memory region size (%#lx) is not a multiple of "
2639 "quantum size(%#x)!\n", __func__, pdata->name,
2640 pdata->size, pmem[id].quantum);
2641 goto err_reset_pmem_info;
2642 }
2643
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002644 pmem[id].cached = pdata->cached;
2645 pmem[id].buffered = pdata->buffered;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002646 pmem[id].size = pdata->size;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002647 pmem[id].memory_type = pdata->memory_type;
2648 strlcpy(pmem[id].name, pdata->name, PMEM_NAME_SIZE);
2649
2650 pmem[id].num_entries = pmem[id].size / pmem[id].quantum;
2651
2652 memset(&pmem[id].kobj, 0, sizeof(pmem[0].kobj));
2653 pmem[id].kobj.kset = pmem_kset;
2654
2655 switch (pmem[id].allocator_type) {
2656 case PMEM_ALLOCATORTYPE_ALLORNOTHING:
2657 pmem[id].allocate = pmem_allocator_all_or_nothing;
2658 pmem[id].free = pmem_free_all_or_nothing;
2659 pmem[id].free_space = pmem_free_space_all_or_nothing;
2660 pmem[id].len = pmem_len_all_or_nothing;
2661 pmem[id].start_addr = pmem_start_addr_all_or_nothing;
2662 pmem[id].num_entries = 1;
2663 pmem[id].quantum = pmem[id].size;
2664 pmem[id].allocator.all_or_nothing.allocated = 0;
2665
2666 if (kobject_init_and_add(&pmem[id].kobj,
2667 &pmem_allornothing_ktype, NULL,
2668 "%s", pdata->name))
2669 goto out_put_kobj;
2670
2671 break;
2672
2673 case PMEM_ALLOCATORTYPE_BUDDYBESTFIT:
2674 pmem[id].allocator.buddy_bestfit.buddy_bitmap = kmalloc(
2675 pmem[id].num_entries * sizeof(struct pmem_bits),
2676 GFP_KERNEL);
2677 if (!pmem[id].allocator.buddy_bestfit.buddy_bitmap)
2678 goto err_reset_pmem_info;
2679
2680 memset(pmem[id].allocator.buddy_bestfit.buddy_bitmap, 0,
2681 sizeof(struct pmem_bits) * pmem[id].num_entries);
2682
2683 for (i = sizeof(pmem[id].num_entries) * 8 - 1; i >= 0; i--)
2684 if ((pmem[id].num_entries) & 1<<i) {
2685 PMEM_BUDDY_ORDER(id, index) = i;
2686 index = PMEM_BUDDY_NEXT_INDEX(id, index);
2687 }
2688 pmem[id].allocate = pmem_allocator_buddy_bestfit;
2689 pmem[id].free = pmem_free_buddy_bestfit;
2690 pmem[id].free_space = pmem_free_space_buddy_bestfit;
2691 pmem[id].len = pmem_len_buddy_bestfit;
2692 pmem[id].start_addr = pmem_start_addr_buddy_bestfit;
2693 if (kobject_init_and_add(&pmem[id].kobj,
2694 &pmem_buddy_bestfit_ktype, NULL,
2695 "%s", pdata->name))
2696 goto out_put_kobj;
2697
2698 break;
2699
2700 case PMEM_ALLOCATORTYPE_BITMAP: /* 0, default if not explicit */
2701 pmem[id].allocator.bitmap.bitm_alloc = kmalloc(
2702 PMEM_INITIAL_NUM_BITMAP_ALLOCATIONS *
2703 sizeof(*pmem[id].allocator.bitmap.bitm_alloc),
2704 GFP_KERNEL);
2705 if (!pmem[id].allocator.bitmap.bitm_alloc) {
2706 pr_alert("pmem: %s: Unable to register pmem "
2707 "driver %s - can't allocate "
2708 "bitm_alloc!\n",
2709 __func__, pdata->name);
2710 goto err_reset_pmem_info;
2711 }
2712
2713 if (kobject_init_and_add(&pmem[id].kobj,
2714 &pmem_bitmap_ktype, NULL,
2715 "%s", pdata->name))
2716 goto out_put_kobj;
2717
2718 for (i = 0; i < PMEM_INITIAL_NUM_BITMAP_ALLOCATIONS; i++) {
2719 pmem[id].allocator.bitmap.bitm_alloc[i].bit = -1;
2720 pmem[id].allocator.bitmap.bitm_alloc[i].quanta = 0;
2721 }
2722
2723 pmem[id].allocator.bitmap.bitmap_allocs =
2724 PMEM_INITIAL_NUM_BITMAP_ALLOCATIONS;
2725
2726 pmem[id].allocator.bitmap.bitmap =
2727 kcalloc((pmem[id].num_entries + 31) / 32,
2728 sizeof(unsigned int), GFP_KERNEL);
2729 if (!pmem[id].allocator.bitmap.bitmap) {
2730 pr_alert("pmem: %s: Unable to register pmem "
2731 "driver - can't allocate bitmap!\n",
2732 __func__);
2733 goto err_cant_register_device;
2734 }
2735 pmem[id].allocator.bitmap.bitmap_free = pmem[id].num_entries;
2736
2737 pmem[id].allocate = pmem_allocator_bitmap;
2738 pmem[id].free = pmem_free_bitmap;
2739 pmem[id].free_space = pmem_free_space_bitmap;
2740 pmem[id].len = pmem_len_bitmap;
2741 pmem[id].start_addr = pmem_start_addr_bitmap;
2742
2743 DLOG("bitmap allocator id %d (%s), num_entries %u, raw size "
2744 "%lu, quanta size %u\n",
2745 id, pdata->name, pmem[id].allocator.bitmap.bitmap_free,
2746 pmem[id].size, pmem[id].quantum);
2747 break;
2748
2749 case PMEM_ALLOCATORTYPE_SYSTEM:
2750
2751 INIT_LIST_HEAD(&pmem[id].allocator.system_mem.alist);
2752
2753 pmem[id].allocator.system_mem.used = 0;
2754 pmem[id].vbase = NULL;
2755
2756 if (kobject_init_and_add(&pmem[id].kobj,
2757 &pmem_system_ktype, NULL,
2758 "%s", pdata->name))
2759 goto out_put_kobj;
2760
2761 pmem[id].allocate = pmem_allocator_system;
2762 pmem[id].free = pmem_free_system;
2763 pmem[id].free_space = pmem_free_space_system;
2764 pmem[id].len = pmem_len_system;
2765 pmem[id].start_addr = pmem_start_addr_system;
2766 pmem[id].num_entries = 0;
2767 pmem[id].quantum = PAGE_SIZE;
2768
2769 DLOG("system allocator id %d (%s), raw size %lu\n",
2770 id, pdata->name, pmem[id].size);
2771 break;
2772
2773 default:
2774 pr_alert("Invalid allocator type (%d) for pmem driver\n",
2775 pdata->allocator_type);
2776 goto err_reset_pmem_info;
2777 }
2778
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002779 pmem[id].ioctl = ioctl;
2780 pmem[id].release = release;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002781 mutex_init(&pmem[id].arena_mutex);
2782 mutex_init(&pmem[id].data_list_mutex);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002783 INIT_LIST_HEAD(&pmem[id].data_list);
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002784
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002785 pmem[id].dev.name = pdata->name;
2786 pmem[id].dev.minor = id;
2787 pmem[id].dev.fops = &pmem_fops;
Laura Abbott4b392622012-01-19 16:17:02 -08002788 pmem[id].reusable = pdata->reusable;
2789 pr_info("pmem: Initializing %s as %s\n",
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002790 pdata->name, pdata->cached ? "cached" : "non-cached");
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002791
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002792 if (misc_register(&pmem[id].dev)) {
2793 pr_alert("Unable to register pmem driver!\n");
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002794 goto err_cant_register_device;
2795 }
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002796
Laura Abbott4b392622012-01-19 16:17:02 -08002797 if (!pmem[id].reusable) {
2798 pmem[id].base = allocate_contiguous_memory_nomap(pmem[id].size,
2799 pmem[id].memory_type, PAGE_SIZE);
2800 if (!pmem[id].base) {
2801 pr_err("pmem: Cannot allocate from reserved memory for %s\n",
2802 pdata->name);
2803 goto err_misc_deregister;
2804 }
Vipul Gandhif752bf62012-01-09 15:34:04 -08002805 }
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002806
Laura Abbott511edaf2011-12-14 13:34:53 -08002807 /* reusable pmem requires map on demand */
2808 pmem[id].map_on_demand = pdata->map_on_demand || pdata->reusable;
Naveen Ramaraj189f1882011-08-16 17:39:22 -07002809 if (pmem[id].map_on_demand) {
Laura Abbott511edaf2011-12-14 13:34:53 -08002810 if (pmem[id].reusable) {
2811 const struct fmem_data *fmem_info = fmem_get_info();
2812 pmem[id].area = fmem_info->area;
Laura Abbott4b392622012-01-19 16:17:02 -08002813 pmem[id].base = fmem_info->phys;
Laura Abbott511edaf2011-12-14 13:34:53 -08002814 } else {
2815 pmem_vma = get_vm_area(pmem[id].size, VM_IOREMAP);
2816 if (!pmem_vma) {
2817 pr_err("pmem: Failed to allocate virtual space for "
Naveen Ramaraj189f1882011-08-16 17:39:22 -07002818 "%s\n", pdata->name);
Vipul Gandhif752bf62012-01-09 15:34:04 -08002819 goto err_free;
Laura Abbott511edaf2011-12-14 13:34:53 -08002820 }
2821 pr_err("pmem: Reserving virtual address range %lx - %lx for"
Naveen Ramaraj189f1882011-08-16 17:39:22 -07002822 " %s\n", (unsigned long) pmem_vma->addr,
2823 (unsigned long) pmem_vma->addr + pmem[id].size,
2824 pdata->name);
Laura Abbott511edaf2011-12-14 13:34:53 -08002825 pmem[id].area = pmem_vma;
2826 }
Naveen Ramaraj189f1882011-08-16 17:39:22 -07002827 } else
2828 pmem[id].area = NULL;
2829
Vipul Gandhif752bf62012-01-09 15:34:04 -08002830 page = alloc_page(GFP_KERNEL);
2831 if (!page) {
2832 pr_err("pmem: Failed to allocate page for %s\n", pdata->name);
2833 goto cleanup_vm;
2834 }
2835 pmem[id].garbage_pfn = page_to_pfn(page);
Laura Abbott1e36a022011-06-22 17:08:13 -07002836 atomic_set(&pmem[id].allocation_cnt, 0);
Laura Abbott1e36a022011-06-22 17:08:13 -07002837
2838 if (pdata->setup_region)
2839 pmem[id].region_data = pdata->setup_region();
2840
2841 if (pdata->request_region)
2842 pmem[id].mem_request = pdata->request_region;
2843
2844 if (pdata->release_region)
2845 pmem[id].mem_release = pdata->release_region;
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002846
Laura Abbott4b392622012-01-19 16:17:02 -08002847 pr_info("allocating %lu bytes at %lx physical for %s\n",
2848 pmem[id].size, pmem[id].base, pmem[id].name);
2849
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002850 return 0;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002851
Vipul Gandhif752bf62012-01-09 15:34:04 -08002852cleanup_vm:
Laura Abbott4b392622012-01-19 16:17:02 -08002853 if (!pmem[id].reusable)
2854 remove_vm_area(pmem_vma);
Vipul Gandhif752bf62012-01-09 15:34:04 -08002855err_free:
Laura Abbott4b392622012-01-19 16:17:02 -08002856 if (!pmem[id].reusable)
2857 free_contiguous_memory_by_paddr(pmem[id].base);
Vipul Gandhif752bf62012-01-09 15:34:04 -08002858err_misc_deregister:
2859 misc_deregister(&pmem[id].dev);
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002860err_cant_register_device:
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002861out_put_kobj:
2862 kobject_put(&pmem[id].kobj);
2863 if (pmem[id].allocator_type == PMEM_ALLOCATORTYPE_BUDDYBESTFIT)
2864 kfree(pmem[id].allocator.buddy_bestfit.buddy_bitmap);
2865 else if (pmem[id].allocator_type == PMEM_ALLOCATORTYPE_BITMAP) {
2866 kfree(pmem[id].allocator.bitmap.bitmap);
2867 kfree(pmem[id].allocator.bitmap.bitm_alloc);
2868 }
2869err_reset_pmem_info:
2870 pmem[id].allocate = 0;
2871 pmem[id].dev.minor = -1;
2872err_no_mem:
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002873 return -1;
2874}
2875
2876static int pmem_probe(struct platform_device *pdev)
2877{
2878 struct android_pmem_platform_data *pdata;
2879
2880 if (!pdev || !pdev->dev.platform_data) {
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002881 pr_alert("Unable to probe pmem!\n");
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002882 return -1;
2883 }
2884 pdata = pdev->dev.platform_data;
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002885
2886 pm_runtime_set_active(&pdev->dev);
2887 pm_runtime_enable(&pdev->dev);
2888
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002889 return pmem_setup(pdata, NULL, NULL);
2890}
2891
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002892static int pmem_remove(struct platform_device *pdev)
2893{
2894 int id = pdev->id;
2895 __free_page(pfn_to_page(pmem[id].garbage_pfn));
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002896 pm_runtime_disable(&pdev->dev);
Vipul Gandhif752bf62012-01-09 15:34:04 -08002897 if (pmem[id].vbase)
2898 iounmap(pmem[id].vbase);
2899 if (pmem[id].map_on_demand && !pmem[id].reusable && pmem[id].area)
2900 free_vm_area(pmem[id].area);
2901 if (pmem[id].base)
2902 free_contiguous_memory_by_paddr(pmem[id].base);
2903 kobject_put(&pmem[id].kobj);
2904 if (pmem[id].allocator_type == PMEM_ALLOCATORTYPE_BUDDYBESTFIT)
2905 kfree(pmem[id].allocator.buddy_bestfit.buddy_bitmap);
2906 else if (pmem[id].allocator_type == PMEM_ALLOCATORTYPE_BITMAP) {
2907 kfree(pmem[id].allocator.bitmap.bitmap);
2908 kfree(pmem[id].allocator.bitmap.bitm_alloc);
2909 }
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002910 misc_deregister(&pmem[id].dev);
2911 return 0;
2912}
2913
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002914static int pmem_runtime_suspend(struct device *dev)
2915{
2916 dev_dbg(dev, "pm_runtime: suspending...\n");
2917 return 0;
2918}
2919
2920static int pmem_runtime_resume(struct device *dev)
2921{
2922 dev_dbg(dev, "pm_runtime: resuming...\n");
2923 return 0;
2924}
2925
2926static const struct dev_pm_ops pmem_dev_pm_ops = {
2927 .runtime_suspend = pmem_runtime_suspend,
2928 .runtime_resume = pmem_runtime_resume,
2929};
2930
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002931static struct platform_driver pmem_driver = {
2932 .probe = pmem_probe,
2933 .remove = pmem_remove,
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002934 .driver = { .name = "android_pmem",
2935 .pm = &pmem_dev_pm_ops,
2936 }
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002937};
2938
2939
2940static int __init pmem_init(void)
2941{
Bryan Huntsman3f2bc4d2011-08-16 17:27:22 -07002942 /* create /sys/kernel/<PMEM_SYSFS_DIR_NAME> directory */
2943 pmem_kset = kset_create_and_add(PMEM_SYSFS_DIR_NAME,
2944 NULL, kernel_kobj);
2945 if (!pmem_kset) {
2946 pr_err("pmem(%s):kset_create_and_add fail\n", __func__);
2947 return -ENOMEM;
2948 }
2949
Rebecca Schultza4ff0e82008-07-24 11:22:53 -07002950 return platform_driver_register(&pmem_driver);
2951}
2952
2953static void __exit pmem_exit(void)
2954{
2955 platform_driver_unregister(&pmem_driver);
2956}
2957
2958module_init(pmem_init);
2959module_exit(pmem_exit);
2960