blob: 03f32d0bd1d8a7be5a0fddbc058e0a7f6b059efe [file] [log] [blame]
Joonsoo Kima2541292014-08-06 16:05:25 -07001#ifndef __CMA_H__
2#define __CMA_H__
3
Thierry Redingd5e6eff2016-12-12 16:43:15 -08004#include <linux/init.h>
5#include <linux/types.h>
6
Joonsoo Kima2541292014-08-06 16:05:25 -07007/*
8 * There is always at least global CMA area and a few optional
9 * areas configured in kernel .config.
10 */
11#ifdef CONFIG_CMA_AREAS
12#define MAX_CMA_AREAS (1 + CONFIG_CMA_AREAS)
13
14#else
15#define MAX_CMA_AREAS (0)
16
17#endif
18
19struct cma;
20
Pintu Kumare48322a2014-12-18 16:17:15 -080021extern unsigned long totalcma_pages;
Sasha Levinac173822015-04-14 15:47:04 -070022extern phys_addr_t cma_get_base(const struct cma *cma);
23extern unsigned long cma_get_size(const struct cma *cma);
Joonsoo Kima2541292014-08-06 16:05:25 -070024
Weijie Yangdda02fd2014-10-24 17:47:57 +080025extern int __init cma_declare_contiguous(phys_addr_t base,
26 phys_addr_t size, phys_addr_t limit,
Joonsoo Kima2541292014-08-06 16:05:25 -070027 phys_addr_t alignment, unsigned int order_per_bit,
Joonsoo Kimc1f733aa2014-08-06 16:05:32 -070028 bool fixed, struct cma **res_cma);
Sasha Levinac173822015-04-14 15:47:04 -070029extern int cma_init_reserved_mem(phys_addr_t base, phys_addr_t size,
30 unsigned int order_per_bit,
Marek Szyprowskide9e14e2014-10-13 15:51:09 -070031 struct cma **res_cma);
Lucas Stache2f466e2017-02-24 14:58:41 -080032extern struct page *cma_alloc(struct cma *cma, size_t count, unsigned int align,
33 gfp_t gfp_mask);
Sasha Levinac173822015-04-14 15:47:04 -070034extern bool cma_release(struct cma *cma, const struct page *pages, unsigned int count);
Joonsoo Kima2541292014-08-06 16:05:25 -070035#endif