blob: d04f1789c228b323ddaecb60f48f9c65397e8ccf [file] [log] [blame]
Joonsoo Kima2541292014-08-06 16:05:25 -07001#ifndef __CMA_H__
2#define __CMA_H__
3
4/*
5 * There is always at least global CMA area and a few optional
6 * areas configured in kernel .config.
7 */
8#ifdef CONFIG_CMA_AREAS
9#define MAX_CMA_AREAS (1 + CONFIG_CMA_AREAS)
10
11#else
12#define MAX_CMA_AREAS (0)
13
14#endif
15
16struct cma;
17
Pintu Kumare48322a2014-12-18 16:17:15 -080018extern unsigned long totalcma_pages;
Sasha Levinac173822015-04-14 15:47:04 -070019extern phys_addr_t cma_get_base(const struct cma *cma);
20extern unsigned long cma_get_size(const struct cma *cma);
Laura Abbott77dbb1d2017-04-18 11:27:03 -070021extern const char *cma_get_name(const struct cma *cma);
Joonsoo Kima2541292014-08-06 16:05:25 -070022
Weijie Yangdda02fd2014-10-24 17:47:57 +080023extern int __init cma_declare_contiguous(phys_addr_t base,
24 phys_addr_t size, phys_addr_t limit,
Joonsoo Kima2541292014-08-06 16:05:25 -070025 phys_addr_t alignment, unsigned int order_per_bit,
Laura Abbott77dbb1d2017-04-18 11:27:03 -070026 bool fixed, const char *name, struct cma **res_cma);
Sasha Levinac173822015-04-14 15:47:04 -070027extern int cma_init_reserved_mem(phys_addr_t base, phys_addr_t size,
28 unsigned int order_per_bit,
Laura Abbott77dbb1d2017-04-18 11:27:03 -070029 const char *name,
Marek Szyprowskide9e14e2014-10-13 15:51:09 -070030 struct cma **res_cma);
Rohit Vaswani67a2e2132015-10-22 13:32:11 -070031extern struct page *cma_alloc(struct cma *cma, size_t count, unsigned int align);
Sasha Levinac173822015-04-14 15:47:04 -070032extern bool cma_release(struct cma *cma, const struct page *pages, unsigned int count);
Joonsoo Kima2541292014-08-06 16:05:25 -070033#endif