/* * Copyright (c) 2011-2012, Code Aurora Forum. All rights reserved. * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are * met: * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials provided * with the distribution. * * Neither the name of Code Aurora Forum, Inc. nor the names of its * contributors may be used to endorse or promote products derived * from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED "AS IS" AND ANY EXPRESS OR IMPLIED * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE * OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN * IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include <cutils/log.h> #include <fcntl.h> #include "gralloc_priv.h" #include "alloc_controller.h" #include "memalloc.h" #include "ionalloc.h" #include "gr.h" #include "comptype.h" using namespace gralloc; using namespace qdutils; //Common functions static bool canFallback(int usage, bool triedSystem) { // Fallback to system heap when alloc fails unless // 1. Composition type is MDP // 2. Alloc from system heap was already tried // 3. The heap type is requsted explicitly // 4. The heap type is protected // 5. The buffer is meant for external display only if(QCCompositionType::getInstance().getCompositionType() & COMPOSITION_TYPE_MDP) return false; if(triedSystem) return false; if(usage & (GRALLOC_HEAP_MASK | GRALLOC_USAGE_PROTECTED | GRALLOC_USAGE_PRIVATE_CP_BUFFER)) return false; if(usage & (GRALLOC_HEAP_MASK | GRALLOC_USAGE_PRIVATE_EXTERNAL_ONLY)) return false; //Return true by default return true; } static bool useUncached(int usage) { // System heaps cannot be uncached if(usage & (GRALLOC_USAGE_PRIVATE_SYSTEM_HEAP | GRALLOC_USAGE_PRIVATE_IOMMU_HEAP)) return false; if (usage & GRALLOC_USAGE_PRIVATE_UNCACHED) return true; return false; } IAllocController* IAllocController::sController = NULL; IAllocController* IAllocController::getInstance(void) { if(sController == NULL) { sController = new IonController(); } return sController; } //-------------- IonController-----------------------// IonController::IonController() { mIonAlloc = new IonAlloc(); } int IonController::allocate(alloc_data& data, int usage) { int ionFlags = 0; int ret; bool noncontig = false; data.uncached = useUncached(usage); data.allocType = 0; if(usage & GRALLOC_USAGE_PRIVATE_UI_CONTIG_HEAP) ionFlags |= ION_HEAP(ION_SF_HEAP_ID); if(usage & GRALLOC_USAGE_PRIVATE_SYSTEM_HEAP) { ionFlags |= ION_HEAP(ION_SYSTEM_HEAP_ID); noncontig = true; } if(usage & GRALLOC_USAGE_PRIVATE_IOMMU_HEAP) ionFlags |= ION_HEAP(ION_IOMMU_HEAP_ID); if(usage & GRALLOC_USAGE_PRIVATE_MM_HEAP) ionFlags |= ION_HEAP(ION_CP_MM_HEAP_ID); if(usage & GRALLOC_USAGE_PRIVATE_CAMERA_HEAP) ionFlags |= ION_HEAP(ION_CAMERA_HEAP_ID); if(usage & GRALLOC_USAGE_PRIVATE_CP_BUFFER) ionFlags |= ION_SECURE; // if no flags are set, default to // SF + IOMMU heaps, so that bypass can work // we can fall back to system heap if // we run out. if(!ionFlags) ionFlags = ION_HEAP(ION_SF_HEAP_ID) | ION_HEAP(ION_IOMMU_HEAP_ID); data.flags = ionFlags; ret = mIonAlloc->alloc_buffer(data); // Fallback if(ret < 0 && canFallback(usage, (ionFlags & ION_SYSTEM_HEAP_ID))) { ALOGW("Falling back to system heap"); data.flags = ION_HEAP(ION_SYSTEM_HEAP_ID); noncontig = true; ret = mIonAlloc->alloc_buffer(data); } if(ret >= 0 ) { data.allocType |= private_handle_t::PRIV_FLAGS_USES_ION; if(noncontig) data.allocType |= private_handle_t::PRIV_FLAGS_NONCONTIGUOUS_MEM; if(ionFlags & ION_SECURE) data.allocType |= private_handle_t::PRIV_FLAGS_SECURE_BUFFER; } return ret; } IMemAlloc* IonController::getAllocator(int flags) { IMemAlloc* memalloc = NULL; if (flags & private_handle_t::PRIV_FLAGS_USES_ION) { memalloc = mIonAlloc; } else { ALOGE("%s: Invalid flags passed: 0x%x", __FUNCTION__, flags); } return memalloc; } size_t getBufferSizeAndDimensions(int width, int height, int format, int& alignedw, int &alignedh) { size_t size; alignedw = ALIGN(width, 32); alignedh = ALIGN(height, 32); switch (format) { case HAL_PIXEL_FORMAT_RGBA_8888: case HAL_PIXEL_FORMAT_RGBX_8888: case HAL_PIXEL_FORMAT_BGRA_8888: size = alignedw * alignedh * 4; break; case HAL_PIXEL_FORMAT_RGB_888: size = alignedw * alignedh * 3; break; case HAL_PIXEL_FORMAT_RGB_565: case HAL_PIXEL_FORMAT_RGBA_5551: case HAL_PIXEL_FORMAT_RGBA_4444: size = alignedw * alignedh * 2; break; // adreno formats case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: // NV21 size = ALIGN(alignedw*alignedh, 4096); size += ALIGN(2 * ALIGN(width/2, 32) * ALIGN(height/2, 32), 4096); break; case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: // NV12 // The chroma plane is subsampled, // but the pitch in bytes is unchanged // The GPU needs 4K alignment, but the video decoder needs 8K alignedw = ALIGN(width, 128); size = ALIGN( alignedw * alignedh, 8192); size += ALIGN( alignedw * ALIGN(height/2, 32), 8192); break; case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: case HAL_PIXEL_FORMAT_YCbCr_420_SP: case HAL_PIXEL_FORMAT_YCrCb_420_SP: case HAL_PIXEL_FORMAT_YV12: if ((format == HAL_PIXEL_FORMAT_YV12) && ((width&1) || (height&1))) { ALOGE("w or h is odd for the YV12 format"); return -EINVAL; } alignedw = ALIGN(width, 16); alignedh = height; if (HAL_PIXEL_FORMAT_NV12_ENCODEABLE == format) { // The encoder requires a 2K aligned chroma offset. size = ALIGN(alignedw*alignedh, 2048) + (ALIGN(alignedw/2, 16) * (alignedh/2))*2; } else { size = alignedw*alignedh + (ALIGN(alignedw/2, 16) * (alignedh/2))*2; } size = ALIGN(size, 4096); break; case HAL_PIXEL_FORMAT_YCbCr_422_SP: case HAL_PIXEL_FORMAT_YCrCb_422_SP: if(width & 1) { ALOGE("width is odd for the YUV422_SP format"); return -EINVAL; } alignedw = ALIGN(width, 16); alignedh = height; size = ALIGN(alignedw * alignedh * 2, 4096); break; default: ALOGE("unrecognized pixel format: 0x%x", format); return -EINVAL; } return size; } // Allocate buffer from width, height and format into a // private_handle_t. It is the responsibility of the caller // to free the buffer using the free_buffer function int alloc_buffer(private_handle_t **pHnd, int w, int h, int format, int usage) { alloc_data data; int alignedw, alignedh; gralloc::IAllocController* sAlloc = gralloc::IAllocController::getInstance(); data.base = 0; data.fd = -1; data.offset = 0; data.size = getBufferSizeAndDimensions(w, h, format, alignedw, alignedh); data.align = getpagesize(); data.uncached = useUncached(usage); int allocFlags = usage; int err = sAlloc->allocate(data, allocFlags); if (0 != err) { ALOGE("%s: allocate failed", __FUNCTION__); return -ENOMEM; } private_handle_t* hnd = new private_handle_t(data.fd, data.size, data.allocType, 0, format, alignedw, alignedh); hnd->base = (int) data.base; hnd->offset = data.offset; hnd->gpuaddr = 0; *pHnd = hnd; return 0; } void free_buffer(private_handle_t *hnd) { gralloc::IAllocController* sAlloc = gralloc::IAllocController::getInstance(); if (hnd && hnd->fd > 0) { IMemAlloc* memalloc = sAlloc->getAllocator(hnd->flags); memalloc->free_buffer((void*)hnd->base, hnd->size, hnd->offset, hnd->fd); } if(hnd) delete hnd; }