| /* |
| * include/linux/ion.h |
| * |
| * Copyright (C) 2011 Google, Inc. |
| * Copyright (c) 2011-2012, Code Aurora Forum. All rights reserved. |
| * |
| * This software is licensed under the terms of the GNU General Public |
| * License version 2, as published by the Free Software Foundation, and |
| * may be copied, distributed, and modified under those terms. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| * |
| */ |
| |
| #ifndef _LINUX_ION_H |
| #define _LINUX_ION_H |
| |
| #include <linux/ioctl.h> |
| #include <linux/types.h> |
| |
| |
| struct ion_handle; |
| /** |
| * enum ion_heap_types - list of all possible types of heaps |
| * @ION_HEAP_TYPE_SYSTEM: memory allocated via vmalloc |
| * @ION_HEAP_TYPE_SYSTEM_CONTIG: memory allocated via kmalloc |
| * @ION_HEAP_TYPE_CARVEOUT: memory allocated from a prereserved |
| * carveout heap, allocations are physically |
| * contiguous |
| * @ION_HEAP_TYPE_IOMMU: IOMMU memory |
| * @ION_HEAP_TYPE_CP: memory allocated from a prereserved |
| * carveout heap, allocations are physically |
| * contiguous. Used for content protection. |
| * @ION_HEAP_END: helper for iterating over heaps |
| */ |
| enum ion_heap_type { |
| ION_HEAP_TYPE_SYSTEM, |
| ION_HEAP_TYPE_SYSTEM_CONTIG, |
| ION_HEAP_TYPE_CARVEOUT, |
| ION_HEAP_TYPE_IOMMU, |
| ION_HEAP_TYPE_CP, |
| ION_HEAP_TYPE_CUSTOM, /* must be last so device specific heaps always |
| are at the end of this enum */ |
| ION_NUM_HEAPS, |
| }; |
| |
| #define ION_HEAP_SYSTEM_MASK (1 << ION_HEAP_TYPE_SYSTEM) |
| #define ION_HEAP_SYSTEM_CONTIG_MASK (1 << ION_HEAP_TYPE_SYSTEM_CONTIG) |
| #define ION_HEAP_CARVEOUT_MASK (1 << ION_HEAP_TYPE_CARVEOUT) |
| #define ION_HEAP_CP_MASK (1 << ION_HEAP_TYPE_CP) |
| |
| |
| /** |
| * These are the only ids that should be used for Ion heap ids. |
| * The ids listed are the order in which allocation will be attempted |
| * if specified. Don't swap the order of heap ids unless you know what |
| * you are doing! |
| * Id's are spaced by purpose to allow new Id's to be inserted in-between (for |
| * possible fallbacks) |
| */ |
| |
| enum ion_heap_ids { |
| INVALID_HEAP_ID = -1, |
| ION_CP_MM_HEAP_ID = 8, |
| ION_CP_MFC_HEAP_ID = 12, |
| ION_CP_WB_HEAP_ID = 16, /* 8660 only */ |
| ION_CAMERA_HEAP_ID = 20, /* 8660 only */ |
| ION_SF_HEAP_ID = 24, |
| ION_IOMMU_HEAP_ID = 25, |
| ION_QSECOM_HEAP_ID = 27, |
| ION_AUDIO_HEAP_ID = 28, |
| |
| ION_MM_FIRMWARE_HEAP_ID = 29, |
| ION_SYSTEM_HEAP_ID = 30, |
| |
| ION_HEAP_ID_RESERVED = 31 /** Bit reserved for ION_SECURE flag */ |
| }; |
| |
| enum ion_fixed_position { |
| NOT_FIXED, |
| FIXED_LOW, |
| FIXED_MIDDLE, |
| FIXED_HIGH, |
| }; |
| |
| enum cp_mem_usage { |
| VIDEO_BITSTREAM = 0x1, |
| VIDEO_PIXEL = 0x2, |
| VIDEO_NONPIXEL = 0x3, |
| MAX_USAGE = 0x4, |
| UNKNOWN = 0x7FFFFFFF, |
| }; |
| |
| /** |
| * Flag to use when allocating to indicate that a heap is secure. |
| */ |
| #define ION_SECURE (1 << ION_HEAP_ID_RESERVED) |
| |
| /** |
| * Macro should be used with ion_heap_ids defined above. |
| */ |
| #define ION_HEAP(bit) (1 << (bit)) |
| |
| #define ION_VMALLOC_HEAP_NAME "vmalloc" |
| #define ION_AUDIO_HEAP_NAME "audio" |
| #define ION_SF_HEAP_NAME "sf" |
| #define ION_MM_HEAP_NAME "mm" |
| #define ION_CAMERA_HEAP_NAME "camera_preview" |
| #define ION_IOMMU_HEAP_NAME "iommu" |
| #define ION_MFC_HEAP_NAME "mfc" |
| #define ION_WB_HEAP_NAME "wb" |
| #define ION_MM_FIRMWARE_HEAP_NAME "mm_fw" |
| #define ION_QSECOM_HEAP_NAME "qsecom" |
| #define ION_FMEM_HEAP_NAME "fmem" |
| |
| #define CACHED 1 |
| #define UNCACHED 0 |
| |
| #define ION_CACHE_SHIFT 0 |
| |
| #define ION_SET_CACHE(__cache) ((__cache) << ION_CACHE_SHIFT) |
| |
| #define ION_IS_CACHED(__flags) ((__flags) & (1 << ION_CACHE_SHIFT)) |
| |
| /* |
| * This flag allows clients when mapping into the IOMMU to specify to |
| * defer un-mapping from the IOMMU until the buffer memory is freed. |
| */ |
| #define ION_IOMMU_UNMAP_DELAYED 1 |
| |
| #ifdef __KERNEL__ |
| #include <linux/err.h> |
| #include <mach/ion.h> |
| struct ion_device; |
| struct ion_heap; |
| struct ion_mapper; |
| struct ion_client; |
| struct ion_buffer; |
| |
| /* This should be removed some day when phys_addr_t's are fully |
| plumbed in the kernel, and all instances of ion_phys_addr_t should |
| be converted to phys_addr_t. For the time being many kernel interfaces |
| do not accept phys_addr_t's that would have to */ |
| #define ion_phys_addr_t unsigned long |
| #define ion_virt_addr_t unsigned long |
| |
| /** |
| * struct ion_platform_heap - defines a heap in the given platform |
| * @type: type of the heap from ion_heap_type enum |
| * @id: unique identifier for heap. When allocating (lower numbers |
| * will be allocated from first) |
| * @name: used for debug purposes |
| * @base: base address of heap in physical memory if applicable |
| * @size: size of the heap in bytes if applicable |
| * @memory_type:Memory type used for the heap |
| * @has_outer_cache: set to 1 if outer cache is used, 0 otherwise. |
| * @extra_data: Extra data specific to each heap type |
| */ |
| struct ion_platform_heap { |
| enum ion_heap_type type; |
| unsigned int id; |
| const char *name; |
| ion_phys_addr_t base; |
| size_t size; |
| enum ion_memory_types memory_type; |
| unsigned int has_outer_cache; |
| void *extra_data; |
| }; |
| |
| /** |
| * struct ion_cp_heap_pdata - defines a content protection heap in the given |
| * platform |
| * @permission_type: Memory ID used to identify the memory to TZ |
| * @align: Alignment requirement for the memory |
| * @secure_base: Base address for securing the heap. |
| * Note: This might be different from actual base address |
| * of this heap in the case of a shared heap. |
| * @secure_size: Memory size for securing the heap. |
| * Note: This might be different from actual size |
| * of this heap in the case of a shared heap. |
| * @reusable Flag indicating whether this heap is reusable of not. |
| * (see FMEM) |
| * @mem_is_fmem Flag indicating whether this memory is coming from fmem |
| * or not. |
| * @fixed_position If nonzero, position in the fixed area. |
| * @virt_addr: Virtual address used when using fmem. |
| * @iommu_map_all: Indicates whether we should map whole heap into IOMMU. |
| * @iommu_2x_map_domain: Indicates the domain to use for overmapping. |
| * @request_region: function to be called when the number of allocations |
| * goes from 0 -> 1 |
| * @release_region: function to be called when the number of allocations |
| * goes from 1 -> 0 |
| * @setup_region: function to be called upon ion registration |
| * |
| */ |
| struct ion_cp_heap_pdata { |
| enum ion_permission_type permission_type; |
| unsigned int align; |
| ion_phys_addr_t secure_base; /* Base addr used when heap is shared */ |
| size_t secure_size; /* Size used for securing heap when heap is shared*/ |
| int reusable; |
| int mem_is_fmem; |
| enum ion_fixed_position fixed_position; |
| int iommu_map_all; |
| int iommu_2x_map_domain; |
| ion_virt_addr_t *virt_addr; |
| int (*request_region)(void *); |
| int (*release_region)(void *); |
| void *(*setup_region)(void); |
| }; |
| |
| /** |
| * struct ion_co_heap_pdata - defines a carveout heap in the given platform |
| * @adjacent_mem_id: Id of heap that this heap must be adjacent to. |
| * @align: Alignment requirement for the memory |
| * @mem_is_fmem Flag indicating whether this memory is coming from fmem |
| * or not. |
| * @fixed_position If nonzero, position in the fixed area. |
| * @request_region: function to be called when the number of allocations |
| * goes from 0 -> 1 |
| * @release_region: function to be called when the number of allocations |
| * goes from 1 -> 0 |
| * @setup_region: function to be called upon ion registration |
| * |
| */ |
| struct ion_co_heap_pdata { |
| int adjacent_mem_id; |
| unsigned int align; |
| int mem_is_fmem; |
| enum ion_fixed_position fixed_position; |
| int (*request_region)(void *); |
| int (*release_region)(void *); |
| void *(*setup_region)(void); |
| }; |
| |
| /** |
| * struct ion_platform_data - array of platform heaps passed from board file |
| * @has_outer_cache: set to 1 if outer cache is used, 0 otherwise. |
| * @nr: number of structures in the array |
| * @request_region: function to be called when the number of allocations goes |
| * from 0 -> 1 |
| * @release_region: function to be called when the number of allocations goes |
| * from 1 -> 0 |
| * @setup_region: function to be called upon ion registration |
| * @heaps: array of platform_heap structions |
| * |
| * Provided by the board file in the form of platform data to a platform device. |
| */ |
| struct ion_platform_data { |
| unsigned int has_outer_cache; |
| int nr; |
| int (*request_region)(void *); |
| int (*release_region)(void *); |
| void *(*setup_region)(void); |
| struct ion_platform_heap heaps[]; |
| }; |
| |
| #ifdef CONFIG_ION |
| |
| /** |
| * ion_client_create() - allocate a client and returns it |
| * @dev: the global ion device |
| * @heap_mask: mask of heaps this client can allocate from |
| * @name: used for debugging |
| */ |
| struct ion_client *ion_client_create(struct ion_device *dev, |
| unsigned int heap_mask, const char *name); |
| |
| /** |
| * msm_ion_client_create - allocate a client using the ion_device specified in |
| * drivers/gpu/ion/msm/msm_ion.c |
| * |
| * heap_mask and name are the same as ion_client_create, return values |
| * are the same as ion_client_create. |
| */ |
| |
| struct ion_client *msm_ion_client_create(unsigned int heap_mask, |
| const char *name); |
| |
| /** |
| * ion_client_destroy() - free's a client and all it's handles |
| * @client: the client |
| * |
| * Free the provided client and all it's resources including |
| * any handles it is holding. |
| */ |
| void ion_client_destroy(struct ion_client *client); |
| |
| /** |
| * ion_alloc - allocate ion memory |
| * @client: the client |
| * @len: size of the allocation |
| * @align: requested allocation alignment, lots of hardware blocks have |
| * alignment requirements of some kind |
| * @flags: mask of heaps to allocate from, if multiple bits are set |
| * heaps will be tried in order from lowest to highest order bit |
| * |
| * Allocate memory in one of the heaps provided in heap mask and return |
| * an opaque handle to it. |
| */ |
| struct ion_handle *ion_alloc(struct ion_client *client, size_t len, |
| size_t align, unsigned int flags); |
| |
| /** |
| * ion_free - free a handle |
| * @client: the client |
| * @handle: the handle to free |
| * |
| * Free the provided handle. |
| */ |
| void ion_free(struct ion_client *client, struct ion_handle *handle); |
| |
| /** |
| * ion_phys - returns the physical address and len of a handle |
| * @client: the client |
| * @handle: the handle |
| * @addr: a pointer to put the address in |
| * @len: a pointer to put the length in |
| * |
| * This function queries the heap for a particular handle to get the |
| * handle's physical address. It't output is only correct if |
| * a heap returns physically contiguous memory -- in other cases |
| * this api should not be implemented -- ion_map_dma should be used |
| * instead. Returns -EINVAL if the handle is invalid. This has |
| * no implications on the reference counting of the handle -- |
| * the returned value may not be valid if the caller is not |
| * holding a reference. |
| */ |
| int ion_phys(struct ion_client *client, struct ion_handle *handle, |
| ion_phys_addr_t *addr, size_t *len); |
| |
| /** |
| * ion_map_kernel - create mapping for the given handle |
| * @client: the client |
| * @handle: handle to map |
| * @flags: flags for this mapping |
| * |
| * Map the given handle into the kernel and return a kernel address that |
| * can be used to access this address. If no flags are specified, this |
| * will return a non-secure uncached mapping. |
| */ |
| void *ion_map_kernel(struct ion_client *client, struct ion_handle *handle, |
| unsigned long flags); |
| |
| /** |
| * ion_unmap_kernel() - destroy a kernel mapping for a handle |
| * @client: the client |
| * @handle: handle to unmap |
| */ |
| void ion_unmap_kernel(struct ion_client *client, struct ion_handle *handle); |
| |
| /** |
| * ion_map_dma - create a dma mapping for a given handle |
| * @client: the client |
| * @handle: handle to map |
| * |
| * Return an sglist describing the given handle |
| */ |
| struct scatterlist *ion_map_dma(struct ion_client *client, |
| struct ion_handle *handle, |
| unsigned long flags); |
| |
| /** |
| * ion_unmap_dma() - destroy a dma mapping for a handle |
| * @client: the client |
| * @handle: handle to unmap |
| */ |
| void ion_unmap_dma(struct ion_client *client, struct ion_handle *handle); |
| |
| /** |
| * ion_share() - given a handle, obtain a buffer to pass to other clients |
| * @client: the client |
| * @handle: the handle to share |
| * |
| * Given a handle, return a buffer, which exists in a global name |
| * space, and can be passed to other clients. Should be passed into ion_import |
| * to obtain a new handle for this buffer. |
| * |
| * NOTE: This function does do not an extra reference. The burden is on the |
| * caller to make sure the buffer doesn't go away while it's being passed to |
| * another client. That is, ion_free should not be called on this handle until |
| * the buffer has been imported into the other client. |
| */ |
| struct ion_buffer *ion_share(struct ion_client *client, |
| struct ion_handle *handle); |
| |
| /** |
| * ion_import() - given an buffer in another client, import it |
| * @client: this blocks client |
| * @buffer: the buffer to import (as obtained from ion_share) |
| * |
| * Given a buffer, add it to the client and return the handle to use to refer |
| * to it further. This is called to share a handle from one kernel client to |
| * another. |
| */ |
| struct ion_handle *ion_import(struct ion_client *client, |
| struct ion_buffer *buffer); |
| |
| /** |
| * ion_import_fd() - given an fd obtained via ION_IOC_SHARE ioctl, import it |
| * @client: this blocks client |
| * @fd: the fd |
| * |
| * A helper function for drivers that will be recieving ion buffers shared |
| * with them from userspace. These buffers are represented by a file |
| * descriptor obtained as the return from the ION_IOC_SHARE ioctl. |
| * This function coverts that fd into the underlying buffer, and returns |
| * the handle to use to refer to it further. |
| */ |
| struct ion_handle *ion_import_fd(struct ion_client *client, int fd); |
| |
| /** |
| * ion_handle_get_flags - get the flags for a given handle |
| * |
| * @client - client who allocated the handle |
| * @handle - handle to get the flags |
| * @flags - pointer to store the flags |
| * |
| * Gets the current flags for a handle. These flags indicate various options |
| * of the buffer (caching, security, etc.) |
| */ |
| int ion_handle_get_flags(struct ion_client *client, struct ion_handle *handle, |
| unsigned long *flags); |
| |
| |
| /** |
| * ion_map_iommu - map the given handle into an iommu |
| * |
| * @client - client who allocated the handle |
| * @handle - handle to map |
| * @domain_num - domain number to map to |
| * @partition_num - partition number to allocate iova from |
| * @align - alignment for the iova |
| * @iova_length - length of iova to map. If the iova length is |
| * greater than the handle length, the remaining |
| * address space will be mapped to a dummy buffer. |
| * @iova - pointer to store the iova address |
| * @buffer_size - pointer to store the size of the buffer |
| * @flags - flags for options to map |
| * @iommu_flags - flags specific to the iommu. |
| * |
| * Maps the handle into the iova space specified via domain number. Iova |
| * will be allocated from the partition specified via partition_num. |
| * Returns 0 on success, negative value on error. |
| */ |
| int ion_map_iommu(struct ion_client *client, struct ion_handle *handle, |
| int domain_num, int partition_num, unsigned long align, |
| unsigned long iova_length, unsigned long *iova, |
| unsigned long *buffer_size, |
| unsigned long flags, unsigned long iommu_flags); |
| |
| |
| /** |
| * ion_handle_get_size - get the allocated size of a given handle |
| * |
| * @client - client who allocated the handle |
| * @handle - handle to get the size |
| * @size - pointer to store the size |
| * |
| * gives the allocated size of a handle. returns 0 on success, negative |
| * value on error |
| * |
| * NOTE: This is intended to be used only to get a size to pass to map_iommu. |
| * You should *NOT* rely on this for any other usage. |
| */ |
| |
| int ion_handle_get_size(struct ion_client *client, struct ion_handle *handle, |
| unsigned long *size); |
| |
| /** |
| * ion_unmap_iommu - unmap the handle from an iommu |
| * |
| * @client - client who allocated the handle |
| * @handle - handle to unmap |
| * @domain_num - domain to unmap from |
| * @partition_num - partition to unmap from |
| * |
| * Decrement the reference count on the iommu mapping. If the count is |
| * 0, the mapping will be removed from the iommu. |
| */ |
| void ion_unmap_iommu(struct ion_client *client, struct ion_handle *handle, |
| int domain_num, int partition_num); |
| |
| |
| /** |
| * ion_secure_heap - secure a heap |
| * |
| * @client - a client that has allocated from the heap heap_id |
| * @heap_id - heap id to secure. |
| * |
| * Secure a heap |
| * Returns 0 on success |
| */ |
| int ion_secure_heap(struct ion_device *dev, int heap_id); |
| |
| /** |
| * ion_unsecure_heap - un-secure a heap |
| * |
| * @client - a client that has allocated from the heap heap_id |
| * @heap_id - heap id to un-secure. |
| * |
| * Un-secure a heap |
| * Returns 0 on success |
| */ |
| int ion_unsecure_heap(struct ion_device *dev, int heap_id); |
| |
| /** |
| * msm_ion_secure_heap - secure a heap. Wrapper around ion_secure_heap. |
| * |
| * @heap_id - heap id to secure. |
| * |
| * Secure a heap |
| * Returns 0 on success |
| */ |
| int msm_ion_secure_heap(int heap_id); |
| |
| /** |
| * msm_ion_unsecure_heap - unsecure a heap. Wrapper around ion_unsecure_heap. |
| * |
| * @heap_id - heap id to secure. |
| * |
| * Un-secure a heap |
| * Returns 0 on success |
| */ |
| int msm_ion_unsecure_heap(int heap_id); |
| |
| /** |
| * msm_ion_do_cache_op - do cache operations. |
| * |
| * @client - pointer to ION client. |
| * @handle - pointer to buffer handle. |
| * @vaddr - virtual address to operate on. |
| * @len - Length of data to do cache operation on. |
| * @cmd - Cache operation to perform: |
| * ION_IOC_CLEAN_CACHES |
| * ION_IOC_INV_CACHES |
| * ION_IOC_CLEAN_INV_CACHES |
| * |
| * Returns 0 on success |
| */ |
| int msm_ion_do_cache_op(struct ion_client *client, struct ion_handle *handle, |
| void *vaddr, unsigned long len, unsigned int cmd); |
| |
| #else |
| static inline struct ion_client *ion_client_create(struct ion_device *dev, |
| unsigned int heap_mask, const char *name) |
| { |
| return ERR_PTR(-ENODEV); |
| } |
| |
| static inline struct ion_client *msm_ion_client_create(unsigned int heap_mask, |
| const char *name) |
| { |
| return ERR_PTR(-ENODEV); |
| } |
| |
| static inline void ion_client_destroy(struct ion_client *client) { } |
| |
| static inline struct ion_handle *ion_alloc(struct ion_client *client, |
| size_t len, size_t align, unsigned int flags) |
| { |
| return ERR_PTR(-ENODEV); |
| } |
| |
| static inline void ion_free(struct ion_client *client, |
| struct ion_handle *handle) { } |
| |
| |
| static inline int ion_phys(struct ion_client *client, |
| struct ion_handle *handle, ion_phys_addr_t *addr, size_t *len) |
| { |
| return -ENODEV; |
| } |
| |
| static inline void *ion_map_kernel(struct ion_client *client, |
| struct ion_handle *handle, unsigned long flags) |
| { |
| return ERR_PTR(-ENODEV); |
| } |
| |
| static inline void ion_unmap_kernel(struct ion_client *client, |
| struct ion_handle *handle) { } |
| |
| static inline struct scatterlist *ion_map_dma(struct ion_client *client, |
| struct ion_handle *handle, unsigned long flags) |
| { |
| return ERR_PTR(-ENODEV); |
| } |
| |
| static inline void ion_unmap_dma(struct ion_client *client, |
| struct ion_handle *handle) { } |
| |
| static inline struct ion_buffer *ion_share(struct ion_client *client, |
| struct ion_handle *handle) |
| { |
| return ERR_PTR(-ENODEV); |
| } |
| |
| static inline struct ion_handle *ion_import(struct ion_client *client, |
| struct ion_buffer *buffer) |
| { |
| return ERR_PTR(-ENODEV); |
| } |
| |
| static inline struct ion_handle *ion_import_fd(struct ion_client *client, |
| int fd) |
| { |
| return ERR_PTR(-ENODEV); |
| } |
| |
| static inline int ion_handle_get_flags(struct ion_client *client, |
| struct ion_handle *handle, unsigned long *flags) |
| { |
| return -ENODEV; |
| } |
| |
| static inline int ion_map_iommu(struct ion_client *client, |
| struct ion_handle *handle, int domain_num, |
| int partition_num, unsigned long align, |
| unsigned long iova_length, unsigned long *iova, |
| unsigned long *buffer_size, |
| unsigned long flags, |
| unsigned long iommu_flags) |
| { |
| return -ENODEV; |
| } |
| |
| static inline void ion_unmap_iommu(struct ion_client *client, |
| struct ion_handle *handle, int domain_num, |
| int partition_num) |
| { |
| return; |
| } |
| |
| static inline int ion_secure_heap(struct ion_device *dev, int heap_id) |
| { |
| return -ENODEV; |
| |
| } |
| |
| static inline int ion_unsecure_heap(struct ion_device *dev, int heap_id) |
| { |
| return -ENODEV; |
| } |
| |
| static inline int msm_ion_secure_heap(int heap_id) |
| { |
| return -ENODEV; |
| |
| } |
| |
| static inline int msm_ion_unsecure_heap(int heap_id) |
| { |
| return -ENODEV; |
| } |
| |
| static inline int msm_ion_do_cache_op(struct ion_client *client, |
| struct ion_handle *handle, void *vaddr, |
| unsigned long len, unsigned int cmd) |
| { |
| return -ENODEV; |
| } |
| |
| #endif /* CONFIG_ION */ |
| #endif /* __KERNEL__ */ |
| |
| /** |
| * DOC: Ion Userspace API |
| * |
| * create a client by opening /dev/ion |
| * most operations handled via following ioctls |
| * |
| */ |
| |
| /** |
| * struct ion_allocation_data - metadata passed from userspace for allocations |
| * @len: size of the allocation |
| * @align: required alignment of the allocation |
| * @flags: flags passed to heap |
| * @handle: pointer that will be populated with a cookie to use to refer |
| * to this allocation |
| * |
| * Provided by userspace as an argument to the ioctl |
| */ |
| struct ion_allocation_data { |
| size_t len; |
| size_t align; |
| unsigned int flags; |
| struct ion_handle *handle; |
| }; |
| |
| /** |
| * struct ion_fd_data - metadata passed to/from userspace for a handle/fd pair |
| * @handle: a handle |
| * @fd: a file descriptor representing that handle |
| * |
| * For ION_IOC_SHARE or ION_IOC_MAP userspace populates the handle field with |
| * the handle returned from ion alloc, and the kernel returns the file |
| * descriptor to share or map in the fd field. For ION_IOC_IMPORT, userspace |
| * provides the file descriptor and the kernel returns the handle. |
| */ |
| struct ion_fd_data { |
| struct ion_handle *handle; |
| int fd; |
| }; |
| |
| /** |
| * struct ion_handle_data - a handle passed to/from the kernel |
| * @handle: a handle |
| */ |
| struct ion_handle_data { |
| struct ion_handle *handle; |
| }; |
| |
| /** |
| * struct ion_custom_data - metadata passed to/from userspace for a custom ioctl |
| * @cmd: the custom ioctl function to call |
| * @arg: additional data to pass to the custom ioctl, typically a user |
| * pointer to a predefined structure |
| * |
| * This works just like the regular cmd and arg fields of an ioctl. |
| */ |
| struct ion_custom_data { |
| unsigned int cmd; |
| unsigned long arg; |
| }; |
| |
| |
| /* struct ion_flush_data - data passed to ion for flushing caches |
| * |
| * @handle: handle with data to flush |
| * @fd: fd to flush |
| * @vaddr: userspace virtual address mapped with mmap |
| * @offset: offset into the handle to flush |
| * @length: length of handle to flush |
| * |
| * Performs cache operations on the handle. If p is the start address |
| * of the handle, p + offset through p + offset + length will have |
| * the cache operations performed |
| */ |
| struct ion_flush_data { |
| struct ion_handle *handle; |
| int fd; |
| void *vaddr; |
| unsigned int offset; |
| unsigned int length; |
| }; |
| |
| /* struct ion_flag_data - information about flags for this buffer |
| * |
| * @handle: handle to get flags from |
| * @flags: flags of this handle |
| * |
| * Takes handle as an input and outputs the flags from the handle |
| * in the flag field. |
| */ |
| struct ion_flag_data { |
| struct ion_handle *handle; |
| unsigned long flags; |
| }; |
| |
| #define ION_IOC_MAGIC 'I' |
| |
| /** |
| * DOC: ION_IOC_ALLOC - allocate memory |
| * |
| * Takes an ion_allocation_data struct and returns it with the handle field |
| * populated with the opaque handle for the allocation. |
| */ |
| #define ION_IOC_ALLOC _IOWR(ION_IOC_MAGIC, 0, \ |
| struct ion_allocation_data) |
| |
| /** |
| * DOC: ION_IOC_FREE - free memory |
| * |
| * Takes an ion_handle_data struct and frees the handle. |
| */ |
| #define ION_IOC_FREE _IOWR(ION_IOC_MAGIC, 1, struct ion_handle_data) |
| |
| /** |
| * DOC: ION_IOC_MAP - get a file descriptor to mmap |
| * |
| * Takes an ion_fd_data struct with the handle field populated with a valid |
| * opaque handle. Returns the struct with the fd field set to a file |
| * descriptor open in the current address space. This file descriptor |
| * can then be used as an argument to mmap. |
| */ |
| #define ION_IOC_MAP _IOWR(ION_IOC_MAGIC, 2, struct ion_fd_data) |
| |
| /** |
| * DOC: ION_IOC_SHARE - creates a file descriptor to use to share an allocation |
| * |
| * Takes an ion_fd_data struct with the handle field populated with a valid |
| * opaque handle. Returns the struct with the fd field set to a file |
| * descriptor open in the current address space. This file descriptor |
| * can then be passed to another process. The corresponding opaque handle can |
| * be retrieved via ION_IOC_IMPORT. |
| */ |
| #define ION_IOC_SHARE _IOWR(ION_IOC_MAGIC, 4, struct ion_fd_data) |
| |
| /** |
| * DOC: ION_IOC_IMPORT - imports a shared file descriptor |
| * |
| * Takes an ion_fd_data struct with the fd field populated with a valid file |
| * descriptor obtained from ION_IOC_SHARE and returns the struct with the handle |
| * filed set to the corresponding opaque handle. |
| */ |
| #define ION_IOC_IMPORT _IOWR(ION_IOC_MAGIC, 5, int) |
| |
| /** |
| * DOC: ION_IOC_CUSTOM - call architecture specific ion ioctl |
| * |
| * Takes the argument of the architecture specific ioctl to call and |
| * passes appropriate userdata for that ioctl |
| */ |
| #define ION_IOC_CUSTOM _IOWR(ION_IOC_MAGIC, 6, struct ion_custom_data) |
| |
| |
| /** |
| * DOC: ION_IOC_CLEAN_CACHES - clean the caches |
| * |
| * Clean the caches of the handle specified. |
| */ |
| #define ION_IOC_CLEAN_CACHES _IOWR(ION_IOC_MAGIC, 7, \ |
| struct ion_flush_data) |
| /** |
| * DOC: ION_MSM_IOC_INV_CACHES - invalidate the caches |
| * |
| * Invalidate the caches of the handle specified. |
| */ |
| #define ION_IOC_INV_CACHES _IOWR(ION_IOC_MAGIC, 8, \ |
| struct ion_flush_data) |
| /** |
| * DOC: ION_MSM_IOC_CLEAN_CACHES - clean and invalidate the caches |
| * |
| * Clean and invalidate the caches of the handle specified. |
| */ |
| #define ION_IOC_CLEAN_INV_CACHES _IOWR(ION_IOC_MAGIC, 9, \ |
| struct ion_flush_data) |
| |
| /** |
| * DOC: ION_IOC_GET_FLAGS - get the flags of the handle |
| * |
| * Gets the flags of the current handle which indicate cachability, |
| * secure state etc. |
| */ |
| #define ION_IOC_GET_FLAGS _IOWR(ION_IOC_MAGIC, 10, \ |
| struct ion_flag_data) |
| #endif /* _LINUX_ION_H */ |