/* * Copyright (c) 2011-2012, Code Aurora Forum. All rights reserved. * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are * met: * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials provided * with the distribution. * * Neither the name of Code Aurora Forum, Inc. nor the names of its * contributors may be used to endorse or promote products derived * from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED "AS IS" AND ANY EXPRESS OR IMPLIED * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE * OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN * IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include #include #include "gralloc_priv.h" #include "alloc_controller.h" #include "memalloc.h" #ifdef USE_ION #include "ionalloc.h" #else #include "pmemalloc.h" #include "ashmemalloc.h" #endif #include "gr.h" #include "comptype.h" using namespace gralloc; using namespace qdutils; //Common functions static bool canFallback(int usage, bool triedSystem) { // Fallback to system heap when alloc fails unless // 1. Composition type is MDP // 2. Alloc from system heap was already tried // 3. The heap type is requsted explicitly // 4. The heap type is protected // 5. The buffer is meant for external display only if(QCCompositionType::getInstance().getCompositionType() & COMPOSITION_TYPE_MDP) return false; if(triedSystem) return false; if(usage & (GRALLOC_HEAP_MASK | GRALLOC_USAGE_PROTECTED | GRALLOC_USAGE_PRIVATE_CP_BUFFER)) return false; if(usage & (GRALLOC_HEAP_MASK | GRALLOC_USAGE_PRIVATE_EXTERNAL_ONLY)) return false; //Return true by default return true; } static bool useUncached(int usage) { // System heaps cannot be uncached if(usage & (GRALLOC_USAGE_PRIVATE_SYSTEM_HEAP | GRALLOC_USAGE_PRIVATE_IOMMU_HEAP)) return false; if (usage & GRALLOC_USAGE_PRIVATE_UNCACHED) return true; return false; } IAllocController* IAllocController::sController = NULL; IAllocController* IAllocController::getInstance(bool useMasterHeap) { if(sController == NULL) { #ifdef USE_ION sController = new IonController(); #else if(useMasterHeap) sController = new PmemAshmemController(); else sController = new PmemKernelController(); #endif } return sController; } #ifdef USE_ION //-------------- IonController-----------------------// IonController::IonController() { mIonAlloc = new IonAlloc(); } int IonController::allocate(alloc_data& data, int usage) { int ionFlags = 0; int ret; bool noncontig = false; data.uncached = useUncached(usage); data.allocType = 0; if(usage & GRALLOC_USAGE_PRIVATE_UI_CONTIG_HEAP) ionFlags |= ION_HEAP(ION_SF_HEAP_ID); if(usage & GRALLOC_USAGE_PRIVATE_SYSTEM_HEAP) { ionFlags |= ION_HEAP(ION_SYSTEM_HEAP_ID); noncontig = true; } if(usage & GRALLOC_USAGE_PRIVATE_IOMMU_HEAP) ionFlags |= ION_HEAP(ION_IOMMU_HEAP_ID); if(usage & GRALLOC_USAGE_PRIVATE_MM_HEAP) ionFlags |= ION_HEAP(ION_CP_MM_HEAP_ID); if(usage & GRALLOC_USAGE_PRIVATE_CAMERA_HEAP) ionFlags |= ION_HEAP(ION_CAMERA_HEAP_ID); if(usage & GRALLOC_USAGE_PRIVATE_CP_BUFFER) ionFlags |= ION_SECURE; // if no flags are set, default to // SF + IOMMU heaps, so that bypass can work // we can fall back to system heap if // we run out. if(!ionFlags) ionFlags = ION_HEAP(ION_SF_HEAP_ID) | ION_HEAP(ION_IOMMU_HEAP_ID); data.flags = ionFlags; ret = mIonAlloc->alloc_buffer(data); // Fallback if(ret < 0 && canFallback(usage, (ionFlags & ION_SYSTEM_HEAP_ID))) { ALOGW("Falling back to system heap"); data.flags = ION_HEAP(ION_SYSTEM_HEAP_ID); noncontig = true; ret = mIonAlloc->alloc_buffer(data); } if(ret >= 0 ) { data.allocType |= private_handle_t::PRIV_FLAGS_USES_ION; if(noncontig) data.allocType |= private_handle_t::PRIV_FLAGS_NONCONTIGUOUS_MEM; if(ionFlags & ION_SECURE) data.allocType |= private_handle_t::PRIV_FLAGS_SECURE_BUFFER; } return ret; } IMemAlloc* IonController::getAllocator(int flags) { IMemAlloc* memalloc = NULL; if (flags & private_handle_t::PRIV_FLAGS_USES_ION) { memalloc = mIonAlloc; } else { ALOGE("%s: Invalid flags passed: 0x%x", __FUNCTION__, flags); } return memalloc; } #else //-------------- PmemKernelController-----------------------// PmemKernelController::PmemKernelController() { mPmemAdspAlloc = new PmemKernelAlloc(DEVICE_PMEM_ADSP); // XXX: Right now, there is no need to maintain an instance // of the SMI allocator as we need it only in a few cases } PmemKernelController::~PmemKernelController() { } int PmemKernelController::allocate(alloc_data& data, int usage) { int ret = 0; bool adspFallback = false; if (!(usage & GRALLOC_USAGE_PRIVATE_SMI_HEAP)) adspFallback = true; // Try SMI first if ((usage & GRALLOC_USAGE_PRIVATE_SMI_HEAP) || (usage & GRALLOC_USAGE_EXTERNAL_DISP) || (usage & GRALLOC_USAGE_PROTECTED)) { int tempFd = open(DEVICE_PMEM_SMIPOOL, O_RDWR, 0); if(tempFd > 0) { close(tempFd); IMemAlloc* memalloc; memalloc = new PmemKernelAlloc(DEVICE_PMEM_SMIPOOL); ret = memalloc->alloc_buffer(data); if(ret >= 0) return ret; else { if(adspFallback) ALOGW("Allocation from SMI failed, trying ADSP"); } } } if ((usage & GRALLOC_USAGE_PRIVATE_ADSP_HEAP) || adspFallback) { ret = mPmemAdspAlloc->alloc_buffer(data); } return ret; } IMemAlloc* PmemKernelController::getAllocator(int flags) { IMemAlloc* memalloc; if (flags & private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP) memalloc = mPmemAdspAlloc; else { ALOGE("%s: Invalid flags passed: 0x%x", __FUNCTION__, flags); memalloc = NULL; } return memalloc; } //-------------- PmemAshmmemController-----------------------// PmemAshmemController::PmemAshmemController() { mPmemUserspaceAlloc = new PmemUserspaceAlloc(); mAshmemAlloc = new AshmemAlloc(); mPmemKernelCtrl = new PmemKernelController(); } PmemAshmemController::~PmemAshmemController() { } int PmemAshmemController::allocate(alloc_data& data, int usage) { int ret = 0; data.allocType = 0; // Make buffers cacheable by default data.uncached = false; // Override if we explicitly need uncached buffers if (usage & GRALLOC_USAGE_PRIVATE_UNCACHED) data.uncached = true; // If ADSP or SMI is requested use the kernel controller if(usage & (GRALLOC_USAGE_PRIVATE_ADSP_HEAP| GRALLOC_USAGE_PRIVATE_SMI_HEAP)) { ret = mPmemKernelCtrl->allocate(data, usage); if(ret < 0) ALOGE("%s: Failed to allocate ADSP/SMI memory", __func__); else data.allocType = private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP; return ret; } if(usage & GRALLOC_USAGE_PRIVATE_SYSTEM_HEAP) { ret = mAshmemAlloc->alloc_buffer(data); if(ret >= 0) { data.allocType = private_handle_t::PRIV_FLAGS_USES_ASHMEM; data.allocType |= private_handle_t::PRIV_FLAGS_NONCONTIGUOUS_MEM; } return ret; } // if no memory specific flags are set, // default to EBI heap, so that bypass // can work. We can fall back to system // heap if we run out. ret = mPmemUserspaceAlloc->alloc_buffer(data); // Fallback if(ret >= 0 ) { data.allocType = private_handle_t::PRIV_FLAGS_USES_PMEM; } else if(ret < 0 && canFallback(usage, false)) { ALOGW("Falling back to ashmem"); ret = mAshmemAlloc->alloc_buffer(data); if(ret >= 0) { data.allocType = private_handle_t::PRIV_FLAGS_USES_ASHMEM; data.allocType |= private_handle_t::PRIV_FLAGS_NONCONTIGUOUS_MEM; } } return ret; } IMemAlloc* PmemAshmemController::getAllocator(int flags) { IMemAlloc* memalloc; if (flags & private_handle_t::PRIV_FLAGS_USES_PMEM) memalloc = mPmemUserspaceAlloc; else if (flags & private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP) memalloc = mPmemKernelCtrl->getAllocator(flags); else if (flags & private_handle_t::PRIV_FLAGS_USES_ASHMEM) memalloc = mAshmemAlloc; else { ALOGE("%s: Invalid flags passed: 0x%x", __FUNCTION__, flags); memalloc = NULL; } return memalloc; } #endif size_t getBufferSizeAndDimensions(int width, int height, int format, int& alignedw, int &alignedh) { size_t size; alignedw = ALIGN(width, 32); alignedh = ALIGN(height, 32); switch (format) { case HAL_PIXEL_FORMAT_RGBA_8888: case HAL_PIXEL_FORMAT_RGBX_8888: case HAL_PIXEL_FORMAT_BGRA_8888: size = alignedw * alignedh * 4; break; case HAL_PIXEL_FORMAT_RGB_888: size = alignedw * alignedh * 3; break; case HAL_PIXEL_FORMAT_RGB_565: case HAL_PIXEL_FORMAT_RGBA_5551: case HAL_PIXEL_FORMAT_RGBA_4444: size = alignedw * alignedh * 2; break; // adreno formats case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: // NV21 size = ALIGN(alignedw*alignedh, 4096); size += ALIGN(2 * ALIGN(width/2, 32) * ALIGN(height/2, 32), 4096); break; case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: // NV12 // The chroma plane is subsampled, // but the pitch in bytes is unchanged // The GPU needs 4K alignment, but the video decoder needs 8K alignedw = ALIGN(width, 128); size = ALIGN( alignedw * alignedh, 8192); size += ALIGN( alignedw * ALIGN(height/2, 32), 8192); break; case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: case HAL_PIXEL_FORMAT_YCbCr_420_SP: case HAL_PIXEL_FORMAT_YCrCb_420_SP: case HAL_PIXEL_FORMAT_YV12: if ((format == HAL_PIXEL_FORMAT_YV12) && ((width&1) || (height&1))) { ALOGE("w or h is odd for the YV12 format"); return -EINVAL; } alignedw = ALIGN(width, 16); alignedh = height; if (HAL_PIXEL_FORMAT_NV12_ENCODEABLE == format) { // The encoder requires a 2K aligned chroma offset. size = ALIGN(alignedw*alignedh, 2048) + (ALIGN(alignedw/2, 16) * (alignedh/2))*2; } else { size = alignedw*alignedh + (ALIGN(alignedw/2, 16) * (alignedh/2))*2; } size = ALIGN(size, 4096); break; case HAL_PIXEL_FORMAT_YCbCr_422_SP: case HAL_PIXEL_FORMAT_YCrCb_422_SP: if(width & 1) { ALOGE("width is odd for the YUV422_SP format"); return -EINVAL; } alignedw = ALIGN(width, 16); alignedh = height; size = ALIGN(alignedw * alignedh * 2, 4096); break; default: ALOGE("unrecognized pixel format: 0x%x", format); return -EINVAL; } return size; } // Allocate buffer from width, height and format into a // private_handle_t. It is the responsibility of the caller // to free the buffer using the free_buffer function int alloc_buffer(private_handle_t **pHnd, int w, int h, int format, int usage) { alloc_data data; int alignedw, alignedh; gralloc::IAllocController* sAlloc = gralloc::IAllocController::getInstance(false); data.base = 0; data.fd = -1; data.offset = 0; data.size = getBufferSizeAndDimensions(w, h, format, alignedw, alignedh); data.align = getpagesize(); data.uncached = useUncached(usage); int allocFlags = usage; int err = sAlloc->allocate(data, allocFlags); if (0 != err) { ALOGE("%s: allocate failed", __FUNCTION__); return -ENOMEM; } private_handle_t* hnd = new private_handle_t(data.fd, data.size, data.allocType, 0, format, alignedw, alignedh); hnd->base = (int) data.base; hnd->offset = data.offset; hnd->gpuaddr = 0; *pHnd = hnd; return 0; } void free_buffer(private_handle_t *hnd) { gralloc::IAllocController* sAlloc = gralloc::IAllocController::getInstance(false); if (hnd && hnd->fd > 0) { IMemAlloc* memalloc = sAlloc->getAllocator(hnd->flags); memalloc->free_buffer((void*)hnd->base, hnd->size, hnd->offset, hnd->fd); } if(hnd) delete hnd; }