/* * include/linux/ion.h * * Copyright (C) 2011 Google, Inc. * Copyright (c) 2011-2012, Code Aurora Forum. All rights reserved. * * This software is licensed under the terms of the GNU General Public * License version 2, as published by the Free Software Foundation, and * may be copied, distributed, and modified under those terms. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * */ #ifndef _LINUX_ION_H #define _LINUX_ION_H #include #include struct ion_handle; /** * enum ion_heap_types - list of all possible types of heaps * @ION_HEAP_TYPE_SYSTEM: memory allocated via vmalloc * @ION_HEAP_TYPE_SYSTEM_CONTIG: memory allocated via kmalloc * @ION_HEAP_TYPE_CARVEOUT: memory allocated from a prereserved * carveout heap, allocations are physically * contiguous * @ION_HEAP_TYPE_IOMMU: IOMMU memory * @ION_HEAP_TYPE_CP: memory allocated from a prereserved * carveout heap, allocations are physically * contiguous. Used for content protection. * @ION_HEAP_END: helper for iterating over heaps */ enum ion_heap_type { ION_HEAP_TYPE_SYSTEM, ION_HEAP_TYPE_SYSTEM_CONTIG, ION_HEAP_TYPE_CARVEOUT, ION_HEAP_TYPE_IOMMU, ION_HEAP_TYPE_CP, ION_HEAP_TYPE_CUSTOM, /* must be last so device specific heaps always are at the end of this enum */ ION_NUM_HEAPS, }; #define ION_HEAP_SYSTEM_MASK (1 << ION_HEAP_TYPE_SYSTEM) #define ION_HEAP_SYSTEM_CONTIG_MASK (1 << ION_HEAP_TYPE_SYSTEM_CONTIG) #define ION_HEAP_CARVEOUT_MASK (1 << ION_HEAP_TYPE_CARVEOUT) #define ION_HEAP_CP_MASK (1 << ION_HEAP_TYPE_CP) /** * These are the only ids that should be used for Ion heap ids. * The ids listed are the order in which allocation will be attempted * if specified. Don't swap the order of heap ids unless you know what * you are doing! * Id's are spaced by purpose to allow new Id's to be inserted in-between (for * possible fallbacks) */ enum ion_heap_ids { INVALID_HEAP_ID = -1, ION_CP_MM_HEAP_ID = 8, ION_CP_MFC_HEAP_ID = 12, ION_CP_WB_HEAP_ID = 16, /* 8660 only */ ION_CAMERA_HEAP_ID = 20, /* 8660 only */ ION_SF_HEAP_ID = 24, ION_IOMMU_HEAP_ID = 25, ION_QSECOM_HEAP_ID = 27, ION_AUDIO_HEAP_ID = 28, ION_MM_FIRMWARE_HEAP_ID = 29, ION_SYSTEM_HEAP_ID = 30, ION_HEAP_ID_RESERVED = 31 /** Bit reserved for ION_SECURE flag */ }; enum ion_fixed_position { NOT_FIXED, FIXED_LOW, FIXED_MIDDLE, FIXED_HIGH, }; /** * Flag to use when allocating to indicate that a heap is secure. */ #define ION_SECURE (1 << ION_HEAP_ID_RESERVED) /** * Macro should be used with ion_heap_ids defined above. */ #define ION_HEAP(bit) (1 << (bit)) #define ION_VMALLOC_HEAP_NAME "vmalloc" #define ION_AUDIO_HEAP_NAME "audio" #define ION_SF_HEAP_NAME "sf" #define ION_MM_HEAP_NAME "mm" #define ION_CAMERA_HEAP_NAME "camera_preview" #define ION_IOMMU_HEAP_NAME "iommu" #define ION_MFC_HEAP_NAME "mfc" #define ION_WB_HEAP_NAME "wb" #define ION_MM_FIRMWARE_HEAP_NAME "mm_fw" #define ION_QSECOM_HEAP_NAME "qsecom" #define ION_FMEM_HEAP_NAME "fmem" #define CACHED 1 #define UNCACHED 0 #define ION_CACHE_SHIFT 0 #define ION_SET_CACHE(__cache) ((__cache) << ION_CACHE_SHIFT) #define ION_IS_CACHED(__flags) ((__flags) & (1 << ION_CACHE_SHIFT)) /* * This flag allows clients when mapping into the IOMMU to specify to * defer un-mapping from the IOMMU until the buffer memory is freed. */ #define ION_IOMMU_UNMAP_DELAYED 1 #ifdef __KERNEL__ #include #include struct ion_device; struct ion_heap; struct ion_mapper; struct ion_client; struct ion_buffer; /* This should be removed some day when phys_addr_t's are fully plumbed in the kernel, and all instances of ion_phys_addr_t should be converted to phys_addr_t. For the time being many kernel interfaces do not accept phys_addr_t's that would have to */ #define ion_phys_addr_t unsigned long #define ion_virt_addr_t unsigned long /** * struct ion_platform_heap - defines a heap in the given platform * @type: type of the heap from ion_heap_type enum * @id: unique identifier for heap. When allocating (lower numbers * will be allocated from first) * @name: used for debug purposes * @base: base address of heap in physical memory if applicable * @size: size of the heap in bytes if applicable * @memory_type:Memory type used for the heap * @extra_data: Extra data specific to each heap type */ struct ion_platform_heap { enum ion_heap_type type; unsigned int id; const char *name; ion_phys_addr_t base; size_t size; enum ion_memory_types memory_type; void *extra_data; }; /** * struct ion_cp_heap_pdata - defines a content protection heap in the given * platform * @permission_type: Memory ID used to identify the memory to TZ * @align: Alignment requirement for the memory * @secure_base: Base address for securing the heap. * Note: This might be different from actual base address * of this heap in the case of a shared heap. * @secure_size: Memory size for securing the heap. * Note: This might be different from actual size * of this heap in the case of a shared heap. * @reusable Flag indicating whether this heap is reusable of not. * (see FMEM) * @mem_is_fmem Flag indicating whether this memory is coming from fmem * or not. * @fixed_position If nonzero, position in the fixed area. * @virt_addr: Virtual address used when using fmem. * @request_region: function to be called when the number of allocations * goes from 0 -> 1 * @release_region: function to be called when the number of allocations * goes from 1 -> 0 * @setup_region: function to be called upon ion registration * */ struct ion_cp_heap_pdata { enum ion_permission_type permission_type; unsigned int align; ion_phys_addr_t secure_base; /* Base addr used when heap is shared */ size_t secure_size; /* Size used for securing heap when heap is shared*/ int reusable; int mem_is_fmem; enum ion_fixed_position fixed_position; ion_virt_addr_t *virt_addr; int (*request_region)(void *); int (*release_region)(void *); void *(*setup_region)(void); }; /** * struct ion_co_heap_pdata - defines a carveout heap in the given platform * @adjacent_mem_id: Id of heap that this heap must be adjacent to. * @align: Alignment requirement for the memory * @mem_is_fmem Flag indicating whether this memory is coming from fmem * or not. * @fixed_position If nonzero, position in the fixed area. * @request_region: function to be called when the number of allocations * goes from 0 -> 1 * @release_region: function to be called when the number of allocations * goes from 1 -> 0 * @setup_region: function to be called upon ion registration * */ struct ion_co_heap_pdata { int adjacent_mem_id; unsigned int align; int mem_is_fmem; enum ion_fixed_position fixed_position; int (*request_region)(void *); int (*release_region)(void *); void *(*setup_region)(void); }; /** * struct ion_platform_data - array of platform heaps passed from board file * @nr: number of structures in the array * @request_region: function to be called when the number of allocations goes * from 0 -> 1 * @release_region: function to be called when the number of allocations goes * from 1 -> 0 * @setup_region: function to be called upon ion registration * @heaps: array of platform_heap structions * * Provided by the board file in the form of platform data to a platform device. */ struct ion_platform_data { int nr; int (*request_region)(void *); int (*release_region)(void *); void *(*setup_region)(void); struct ion_platform_heap heaps[]; }; #ifdef CONFIG_ION /** * ion_client_create() - allocate a client and returns it * @dev: the global ion device * @heap_mask: mask of heaps this client can allocate from * @name: used for debugging */ struct ion_client *ion_client_create(struct ion_device *dev, unsigned int heap_mask, const char *name); /** * msm_ion_client_create - allocate a client using the ion_device specified in * drivers/gpu/ion/msm/msm_ion.c * * heap_mask and name are the same as ion_client_create, return values * are the same as ion_client_create. */ struct ion_client *msm_ion_client_create(unsigned int heap_mask, const char *name); /** * ion_client_destroy() - free's a client and all it's handles * @client: the client * * Free the provided client and all it's resources including * any handles it is holding. */ void ion_client_destroy(struct ion_client *client); /** * ion_alloc - allocate ion memory * @client: the client * @len: size of the allocation * @align: requested allocation alignment, lots of hardware blocks have * alignment requirements of some kind * @flags: mask of heaps to allocate from, if multiple bits are set * heaps will be tried in order from lowest to highest order bit * * Allocate memory in one of the heaps provided in heap mask and return * an opaque handle to it. */ struct ion_handle *ion_alloc(struct ion_client *client, size_t len, size_t align, unsigned int flags); /** * ion_free - free a handle * @client: the client * @handle: the handle to free * * Free the provided handle. */ void ion_free(struct ion_client *client, struct ion_handle *handle); /** * ion_phys - returns the physical address and len of a handle * @client: the client * @handle: the handle * @addr: a pointer to put the address in * @len: a pointer to put the length in * * This function queries the heap for a particular handle to get the * handle's physical address. It't output is only correct if * a heap returns physically contiguous memory -- in other cases * this api should not be implemented -- ion_map_dma should be used * instead. Returns -EINVAL if the handle is invalid. This has * no implications on the reference counting of the handle -- * the returned value may not be valid if the caller is not * holding a reference. */ int ion_phys(struct ion_client *client, struct ion_handle *handle, ion_phys_addr_t *addr, size_t *len); /** * ion_map_kernel - create mapping for the given handle * @client: the client * @handle: handle to map * @flags: flags for this mapping * * Map the given handle into the kernel and return a kernel address that * can be used to access this address. If no flags are specified, this * will return a non-secure uncached mapping. */ void *ion_map_kernel(struct ion_client *client, struct ion_handle *handle, unsigned long flags); /** * ion_unmap_kernel() - destroy a kernel mapping for a handle * @client: the client * @handle: handle to unmap */ void ion_unmap_kernel(struct ion_client *client, struct ion_handle *handle); /** * ion_map_dma - create a dma mapping for a given handle * @client: the client * @handle: handle to map * * Return an sglist describing the given handle */ struct scatterlist *ion_map_dma(struct ion_client *client, struct ion_handle *handle, unsigned long flags); /** * ion_unmap_dma() - destroy a dma mapping for a handle * @client: the client * @handle: handle to unmap */ void ion_unmap_dma(struct ion_client *client, struct ion_handle *handle); /** * ion_share() - given a handle, obtain a buffer to pass to other clients * @client: the client * @handle: the handle to share * * Given a handle, return a buffer, which exists in a global name * space, and can be passed to other clients. Should be passed into ion_import * to obtain a new handle for this buffer. * * NOTE: This function does do not an extra reference. The burden is on the * caller to make sure the buffer doesn't go away while it's being passed to * another client. That is, ion_free should not be called on this handle until * the buffer has been imported into the other client. */ struct ion_buffer *ion_share(struct ion_client *client, struct ion_handle *handle); /** * ion_import() - given an buffer in another client, import it * @client: this blocks client * @buffer: the buffer to import (as obtained from ion_share) * * Given a buffer, add it to the client and return the handle to use to refer * to it further. This is called to share a handle from one kernel client to * another. */ struct ion_handle *ion_import(struct ion_client *client, struct ion_buffer *buffer); /** * ion_import_fd() - given an fd obtained via ION_IOC_SHARE ioctl, import it * @client: this blocks client * @fd: the fd * * A helper function for drivers that will be recieving ion buffers shared * with them from userspace. These buffers are represented by a file * descriptor obtained as the return from the ION_IOC_SHARE ioctl. * This function coverts that fd into the underlying buffer, and returns * the handle to use to refer to it further. */ struct ion_handle *ion_import_fd(struct ion_client *client, int fd); /** * ion_handle_get_flags - get the flags for a given handle * * @client - client who allocated the handle * @handle - handle to get the flags * @flags - pointer to store the flags * * Gets the current flags for a handle. These flags indicate various options * of the buffer (caching, security, etc.) */ int ion_handle_get_flags(struct ion_client *client, struct ion_handle *handle, unsigned long *flags); /** * ion_map_iommu - map the given handle into an iommu * * @client - client who allocated the handle * @handle - handle to map * @domain_num - domain number to map to * @partition_num - partition number to allocate iova from * @align - alignment for the iova * @iova_length - length of iova to map. If the iova length is * greater than the handle length, the remaining * address space will be mapped to a dummy buffer. * @iova - pointer to store the iova address * @buffer_size - pointer to store the size of the buffer * @flags - flags for options to map * @iommu_flags - flags specific to the iommu. * * Maps the handle into the iova space specified via domain number. Iova * will be allocated from the partition specified via partition_num. * Returns 0 on success, negative value on error. */ int ion_map_iommu(struct ion_client *client, struct ion_handle *handle, int domain_num, int partition_num, unsigned long align, unsigned long iova_length, unsigned long *iova, unsigned long *buffer_size, unsigned long flags, unsigned long iommu_flags); /** * ion_handle_get_size - get the allocated size of a given handle * * @client - client who allocated the handle * @handle - handle to get the size * @size - pointer to store the size * * gives the allocated size of a handle. returns 0 on success, negative * value on error * * NOTE: This is intended to be used only to get a size to pass to map_iommu. * You should *NOT* rely on this for any other usage. */ int ion_handle_get_size(struct ion_client *client, struct ion_handle *handle, unsigned long *size); /** * ion_unmap_iommu - unmap the handle from an iommu * * @client - client who allocated the handle * @handle - handle to unmap * @domain_num - domain to unmap from * @partition_num - partition to unmap from * * Decrement the reference count on the iommu mapping. If the count is * 0, the mapping will be removed from the iommu. */ void ion_unmap_iommu(struct ion_client *client, struct ion_handle *handle, int domain_num, int partition_num); /** * ion_secure_heap - secure a heap * * @client - a client that has allocated from the heap heap_id * @heap_id - heap id to secure. * * Secure a heap * Returns 0 on success */ int ion_secure_heap(struct ion_device *dev, int heap_id); /** * ion_unsecure_heap - un-secure a heap * * @client - a client that has allocated from the heap heap_id * @heap_id - heap id to un-secure. * * Un-secure a heap * Returns 0 on success */ int ion_unsecure_heap(struct ion_device *dev, int heap_id); /** * msm_ion_secure_heap - secure a heap. Wrapper around ion_secure_heap. * * @heap_id - heap id to secure. * * Secure a heap * Returns 0 on success */ int msm_ion_secure_heap(int heap_id); /** * msm_ion_unsecure_heap - unsecure a heap. Wrapper around ion_unsecure_heap. * * @heap_id - heap id to secure. * * Un-secure a heap * Returns 0 on success */ int msm_ion_unsecure_heap(int heap_id); /** * msm_ion_do_cache_op - do cache operations. * * @client - pointer to ION client. * @handle - pointer to buffer handle. * @vaddr - virtual address to operate on. * @len - Length of data to do cache operation on. * @cmd - Cache operation to perform: * ION_IOC_CLEAN_CACHES * ION_IOC_INV_CACHES * ION_IOC_CLEAN_INV_CACHES * * Returns 0 on success */ int msm_ion_do_cache_op(struct ion_client *client, struct ion_handle *handle, void *vaddr, unsigned long len, unsigned int cmd); #else static inline struct ion_client *ion_client_create(struct ion_device *dev, unsigned int heap_mask, const char *name) { return ERR_PTR(-ENODEV); } static inline struct ion_client *msm_ion_client_create(unsigned int heap_mask, const char *name) { return ERR_PTR(-ENODEV); } static inline void ion_client_destroy(struct ion_client *client) { } static inline struct ion_handle *ion_alloc(struct ion_client *client, size_t len, size_t align, unsigned int flags) { return ERR_PTR(-ENODEV); } static inline void ion_free(struct ion_client *client, struct ion_handle *handle) { } static inline int ion_phys(struct ion_client *client, struct ion_handle *handle, ion_phys_addr_t *addr, size_t *len) { return -ENODEV; } static inline void *ion_map_kernel(struct ion_client *client, struct ion_handle *handle, unsigned long flags) { return ERR_PTR(-ENODEV); } static inline void ion_unmap_kernel(struct ion_client *client, struct ion_handle *handle) { } static inline struct scatterlist *ion_map_dma(struct ion_client *client, struct ion_handle *handle, unsigned long flags) { return ERR_PTR(-ENODEV); } static inline void ion_unmap_dma(struct ion_client *client, struct ion_handle *handle) { } static inline struct ion_buffer *ion_share(struct ion_client *client, struct ion_handle *handle) { return ERR_PTR(-ENODEV); } static inline struct ion_handle *ion_import(struct ion_client *client, struct ion_buffer *buffer) { return ERR_PTR(-ENODEV); } static inline struct ion_handle *ion_import_fd(struct ion_client *client, int fd) { return ERR_PTR(-ENODEV); } static inline int ion_handle_get_flags(struct ion_client *client, struct ion_handle *handle, unsigned long *flags) { return -ENODEV; } static inline int ion_map_iommu(struct ion_client *client, struct ion_handle *handle, int domain_num, int partition_num, unsigned long align, unsigned long iova_length, unsigned long *iova, unsigned long *buffer_size, unsigned long flags, unsigned long iommu_flags) { return -ENODEV; } static inline void ion_unmap_iommu(struct ion_client *client, struct ion_handle *handle, int domain_num, int partition_num) { return; } static inline int ion_secure_heap(struct ion_device *dev, int heap_id) { return -ENODEV; } static inline int ion_unsecure_heap(struct ion_device *dev, int heap_id) { return -ENODEV; } static inline int msm_ion_secure_heap(int heap_id) { return -ENODEV; } static inline int msm_ion_unsecure_heap(int heap_id) { return -ENODEV; } static inline int msm_ion_do_cache_op(struct ion_client *client, struct ion_handle *handle, void *vaddr, unsigned long len, unsigned int cmd) { return -ENODEV; } #endif /* CONFIG_ION */ #endif /* __KERNEL__ */ /** * DOC: Ion Userspace API * * create a client by opening /dev/ion * most operations handled via following ioctls * */ /** * struct ion_allocation_data - metadata passed from userspace for allocations * @len: size of the allocation * @align: required alignment of the allocation * @flags: flags passed to heap * @handle: pointer that will be populated with a cookie to use to refer * to this allocation * * Provided by userspace as an argument to the ioctl */ struct ion_allocation_data { size_t len; size_t align; unsigned int flags; struct ion_handle *handle; }; /** * struct ion_fd_data - metadata passed to/from userspace for a handle/fd pair * @handle: a handle * @fd: a file descriptor representing that handle * * For ION_IOC_SHARE or ION_IOC_MAP userspace populates the handle field with * the handle returned from ion alloc, and the kernel returns the file * descriptor to share or map in the fd field. For ION_IOC_IMPORT, userspace * provides the file descriptor and the kernel returns the handle. */ struct ion_fd_data { struct ion_handle *handle; int fd; }; /** * struct ion_handle_data - a handle passed to/from the kernel * @handle: a handle */ struct ion_handle_data { struct ion_handle *handle; }; /** * struct ion_custom_data - metadata passed to/from userspace for a custom ioctl * @cmd: the custom ioctl function to call * @arg: additional data to pass to the custom ioctl, typically a user * pointer to a predefined structure * * This works just like the regular cmd and arg fields of an ioctl. */ struct ion_custom_data { unsigned int cmd; unsigned long arg; }; /* struct ion_flush_data - data passed to ion for flushing caches * * @handle: handle with data to flush * @fd: fd to flush * @vaddr: userspace virtual address mapped with mmap * @offset: offset into the handle to flush * @length: length of handle to flush * * Performs cache operations on the handle. If p is the start address * of the handle, p + offset through p + offset + length will have * the cache operations performed */ struct ion_flush_data { struct ion_handle *handle; int fd; void *vaddr; unsigned int offset; unsigned int length; }; /* struct ion_flag_data - information about flags for this buffer * * @handle: handle to get flags from * @flags: flags of this handle * * Takes handle as an input and outputs the flags from the handle * in the flag field. */ struct ion_flag_data { struct ion_handle *handle; unsigned long flags; }; #define ION_IOC_MAGIC 'I' /** * DOC: ION_IOC_ALLOC - allocate memory * * Takes an ion_allocation_data struct and returns it with the handle field * populated with the opaque handle for the allocation. */ #define ION_IOC_ALLOC _IOWR(ION_IOC_MAGIC, 0, \ struct ion_allocation_data) /** * DOC: ION_IOC_FREE - free memory * * Takes an ion_handle_data struct and frees the handle. */ #define ION_IOC_FREE _IOWR(ION_IOC_MAGIC, 1, struct ion_handle_data) /** * DOC: ION_IOC_MAP - get a file descriptor to mmap * * Takes an ion_fd_data struct with the handle field populated with a valid * opaque handle. Returns the struct with the fd field set to a file * descriptor open in the current address space. This file descriptor * can then be used as an argument to mmap. */ #define ION_IOC_MAP _IOWR(ION_IOC_MAGIC, 2, struct ion_fd_data) /** * DOC: ION_IOC_SHARE - creates a file descriptor to use to share an allocation * * Takes an ion_fd_data struct with the handle field populated with a valid * opaque handle. Returns the struct with the fd field set to a file * descriptor open in the current address space. This file descriptor * can then be passed to another process. The corresponding opaque handle can * be retrieved via ION_IOC_IMPORT. */ #define ION_IOC_SHARE _IOWR(ION_IOC_MAGIC, 4, struct ion_fd_data) /** * DOC: ION_IOC_IMPORT - imports a shared file descriptor * * Takes an ion_fd_data struct with the fd field populated with a valid file * descriptor obtained from ION_IOC_SHARE and returns the struct with the handle * filed set to the corresponding opaque handle. */ #define ION_IOC_IMPORT _IOWR(ION_IOC_MAGIC, 5, int) /** * DOC: ION_IOC_CUSTOM - call architecture specific ion ioctl * * Takes the argument of the architecture specific ioctl to call and * passes appropriate userdata for that ioctl */ #define ION_IOC_CUSTOM _IOWR(ION_IOC_MAGIC, 6, struct ion_custom_data) /** * DOC: ION_IOC_CLEAN_CACHES - clean the caches * * Clean the caches of the handle specified. */ #define ION_IOC_CLEAN_CACHES _IOWR(ION_IOC_MAGIC, 7, \ struct ion_flush_data) /** * DOC: ION_MSM_IOC_INV_CACHES - invalidate the caches * * Invalidate the caches of the handle specified. */ #define ION_IOC_INV_CACHES _IOWR(ION_IOC_MAGIC, 8, \ struct ion_flush_data) /** * DOC: ION_MSM_IOC_CLEAN_CACHES - clean and invalidate the caches * * Clean and invalidate the caches of the handle specified. */ #define ION_IOC_CLEAN_INV_CACHES _IOWR(ION_IOC_MAGIC, 9, \ struct ion_flush_data) /** * DOC: ION_IOC_GET_FLAGS - get the flags of the handle * * Gets the flags of the current handle which indicate cachability, * secure state etc. */ #define ION_IOC_GET_FLAGS _IOWR(ION_IOC_MAGIC, 10, \ struct ion_flag_data) #endif /* _LINUX_ION_H */