8d111e9b7f
Conflicts: common.mk Change-Id: I512d7e14264192b716e0c5921b685e995853d4f6
438 lines
14 KiB
C++
438 lines
14 KiB
C++
/*
|
|
* Copyright (c) 2011-2012, Code Aurora Forum. All rights reserved.
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions are
|
|
* met:
|
|
* * Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* * Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer in the documentation and/or other materials provided
|
|
* with the distribution.
|
|
* * Neither the name of Code Aurora Forum, Inc. nor the names of its
|
|
* contributors may be used to endorse or promote products derived
|
|
* from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED "AS IS" AND ANY EXPRESS OR IMPLIED
|
|
* WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS
|
|
* BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
|
|
* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY,
|
|
* WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE
|
|
* OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN
|
|
* IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <cutils/log.h>
|
|
#include <fcntl.h>
|
|
#include "gralloc_priv.h"
|
|
#include "alloc_controller.h"
|
|
#include "memalloc.h"
|
|
#ifdef USE_ION
|
|
#include "ionalloc.h"
|
|
#else
|
|
#include "pmemalloc.h"
|
|
#include "ashmemalloc.h"
|
|
#endif
|
|
#include "gr.h"
|
|
#include "comptype.h"
|
|
|
|
using namespace gralloc;
|
|
using namespace qdutils;
|
|
|
|
//Common functions
|
|
static bool canFallback(int usage, bool triedSystem)
|
|
{
|
|
// Fallback to system heap when alloc fails unless
|
|
// 1. Composition type is MDP
|
|
// 2. Alloc from system heap was already tried
|
|
// 3. The heap type is requsted explicitly
|
|
// 4. The heap type is protected
|
|
// 5. The buffer is meant for external display only
|
|
|
|
if(QCCompositionType::getInstance().getCompositionType() &
|
|
COMPOSITION_TYPE_MDP)
|
|
return false;
|
|
if(triedSystem)
|
|
return false;
|
|
if(usage & (GRALLOC_HEAP_MASK | GRALLOC_USAGE_PROTECTED |
|
|
GRALLOC_USAGE_PRIVATE_CP_BUFFER))
|
|
return false;
|
|
if(usage & (GRALLOC_HEAP_MASK | GRALLOC_USAGE_PRIVATE_EXTERNAL_ONLY))
|
|
return false;
|
|
//Return true by default
|
|
return true;
|
|
}
|
|
|
|
static bool useUncached(int usage)
|
|
{
|
|
// System heaps cannot be uncached
|
|
if(usage & (GRALLOC_USAGE_PRIVATE_SYSTEM_HEAP |
|
|
GRALLOC_USAGE_PRIVATE_IOMMU_HEAP))
|
|
return false;
|
|
if (usage & GRALLOC_USAGE_PRIVATE_UNCACHED)
|
|
return true;
|
|
return false;
|
|
}
|
|
|
|
IAllocController* IAllocController::sController = NULL;
|
|
IAllocController* IAllocController::getInstance(bool useMasterHeap)
|
|
{
|
|
if(sController == NULL) {
|
|
#ifdef USE_ION
|
|
sController = new IonController();
|
|
#else
|
|
if(useMasterHeap)
|
|
sController = new PmemAshmemController();
|
|
else
|
|
sController = new PmemKernelController();
|
|
#endif
|
|
}
|
|
return sController;
|
|
}
|
|
|
|
|
|
#ifdef USE_ION
|
|
//-------------- IonController-----------------------//
|
|
IonController::IonController()
|
|
{
|
|
mIonAlloc = new IonAlloc();
|
|
}
|
|
|
|
int IonController::allocate(alloc_data& data, int usage)
|
|
{
|
|
int ionFlags = 0;
|
|
int ret;
|
|
bool noncontig = false;
|
|
|
|
data.uncached = useUncached(usage);
|
|
data.allocType = 0;
|
|
|
|
if(usage & GRALLOC_USAGE_PRIVATE_UI_CONTIG_HEAP)
|
|
ionFlags |= ION_HEAP(ION_SF_HEAP_ID);
|
|
|
|
if(usage & GRALLOC_USAGE_PRIVATE_SYSTEM_HEAP) {
|
|
ionFlags |= ION_HEAP(ION_SYSTEM_HEAP_ID);
|
|
noncontig = true;
|
|
}
|
|
|
|
if(usage & GRALLOC_USAGE_PRIVATE_IOMMU_HEAP)
|
|
ionFlags |= ION_HEAP(ION_IOMMU_HEAP_ID);
|
|
|
|
if(usage & GRALLOC_USAGE_PRIVATE_MM_HEAP)
|
|
ionFlags |= ION_HEAP(ION_CP_MM_HEAP_ID);
|
|
|
|
if(usage & GRALLOC_USAGE_PRIVATE_CAMERA_HEAP)
|
|
ionFlags |= ION_HEAP(ION_CAMERA_HEAP_ID);
|
|
|
|
if(usage & GRALLOC_USAGE_PRIVATE_CP_BUFFER)
|
|
ionFlags |= ION_SECURE;
|
|
|
|
// if no flags are set, default to
|
|
// SF + IOMMU heaps, so that bypass can work
|
|
// we can fall back to system heap if
|
|
// we run out.
|
|
if(!ionFlags)
|
|
ionFlags = ION_HEAP(ION_SF_HEAP_ID) | ION_HEAP(ION_IOMMU_HEAP_ID);
|
|
|
|
data.flags = ionFlags;
|
|
ret = mIonAlloc->alloc_buffer(data);
|
|
|
|
// Fallback
|
|
if(ret < 0 && canFallback(usage,
|
|
(ionFlags & ION_SYSTEM_HEAP_ID)))
|
|
{
|
|
ALOGW("Falling back to system heap");
|
|
data.flags = ION_HEAP(ION_SYSTEM_HEAP_ID);
|
|
noncontig = true;
|
|
ret = mIonAlloc->alloc_buffer(data);
|
|
}
|
|
|
|
if(ret >= 0 ) {
|
|
data.allocType |= private_handle_t::PRIV_FLAGS_USES_ION;
|
|
if(noncontig)
|
|
data.allocType |= private_handle_t::PRIV_FLAGS_NONCONTIGUOUS_MEM;
|
|
if(ionFlags & ION_SECURE)
|
|
data.allocType |= private_handle_t::PRIV_FLAGS_SECURE_BUFFER;
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
IMemAlloc* IonController::getAllocator(int flags)
|
|
{
|
|
IMemAlloc* memalloc = NULL;
|
|
if (flags & private_handle_t::PRIV_FLAGS_USES_ION) {
|
|
memalloc = mIonAlloc;
|
|
} else {
|
|
ALOGE("%s: Invalid flags passed: 0x%x", __FUNCTION__, flags);
|
|
}
|
|
|
|
return memalloc;
|
|
}
|
|
#else
|
|
//-------------- PmemKernelController-----------------------//
|
|
PmemKernelController::PmemKernelController()
|
|
{
|
|
mPmemAdspAlloc = new PmemKernelAlloc(DEVICE_PMEM_ADSP);
|
|
// XXX: Right now, there is no need to maintain an instance
|
|
// of the SMI allocator as we need it only in a few cases
|
|
}
|
|
|
|
PmemKernelController::~PmemKernelController()
|
|
{
|
|
}
|
|
|
|
int PmemKernelController::allocate(alloc_data& data, int usage)
|
|
{
|
|
int ret = 0;
|
|
bool adspFallback = false;
|
|
if (!(usage & GRALLOC_USAGE_PRIVATE_SMI_HEAP))
|
|
adspFallback = true;
|
|
|
|
// Try SMI first
|
|
if ((usage & GRALLOC_USAGE_PRIVATE_SMI_HEAP) ||
|
|
(usage & GRALLOC_USAGE_EXTERNAL_DISP) ||
|
|
(usage & GRALLOC_USAGE_PROTECTED))
|
|
{
|
|
int tempFd = open(DEVICE_PMEM_SMIPOOL, O_RDWR, 0);
|
|
if(tempFd > 0) {
|
|
close(tempFd);
|
|
IMemAlloc* memalloc;
|
|
memalloc = new PmemKernelAlloc(DEVICE_PMEM_SMIPOOL);
|
|
ret = memalloc->alloc_buffer(data);
|
|
if(ret >= 0)
|
|
return ret;
|
|
else {
|
|
if(adspFallback)
|
|
ALOGW("Allocation from SMI failed, trying ADSP");
|
|
}
|
|
}
|
|
}
|
|
|
|
if ((usage & GRALLOC_USAGE_PRIVATE_ADSP_HEAP) || adspFallback) {
|
|
ret = mPmemAdspAlloc->alloc_buffer(data);
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
IMemAlloc* PmemKernelController::getAllocator(int flags)
|
|
{
|
|
IMemAlloc* memalloc;
|
|
if (flags & private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP)
|
|
memalloc = mPmemAdspAlloc;
|
|
else {
|
|
ALOGE("%s: Invalid flags passed: 0x%x", __FUNCTION__, flags);
|
|
memalloc = NULL;
|
|
}
|
|
|
|
return memalloc;
|
|
}
|
|
|
|
//-------------- PmemAshmmemController-----------------------//
|
|
|
|
PmemAshmemController::PmemAshmemController()
|
|
{
|
|
mPmemUserspaceAlloc = new PmemUserspaceAlloc();
|
|
mAshmemAlloc = new AshmemAlloc();
|
|
mPmemKernelCtrl = new PmemKernelController();
|
|
}
|
|
|
|
PmemAshmemController::~PmemAshmemController()
|
|
{
|
|
}
|
|
|
|
int PmemAshmemController::allocate(alloc_data& data, int usage)
|
|
{
|
|
int ret = 0;
|
|
data.allocType = 0;
|
|
|
|
// Make buffers cacheable by default
|
|
data.uncached = false;
|
|
|
|
// Override if we explicitly need uncached buffers
|
|
if (usage & GRALLOC_USAGE_PRIVATE_UNCACHED)
|
|
data.uncached = true;
|
|
|
|
// If ADSP or SMI is requested use the kernel controller
|
|
if(usage & (GRALLOC_USAGE_PRIVATE_ADSP_HEAP|
|
|
GRALLOC_USAGE_PRIVATE_SMI_HEAP)) {
|
|
ret = mPmemKernelCtrl->allocate(data, usage);
|
|
if(ret < 0)
|
|
ALOGE("%s: Failed to allocate ADSP/SMI memory", __func__);
|
|
else
|
|
data.allocType = private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP;
|
|
return ret;
|
|
}
|
|
|
|
if(usage & GRALLOC_USAGE_PRIVATE_SYSTEM_HEAP) {
|
|
ret = mAshmemAlloc->alloc_buffer(data);
|
|
if(ret >= 0) {
|
|
data.allocType = private_handle_t::PRIV_FLAGS_USES_ASHMEM;
|
|
data.allocType |= private_handle_t::PRIV_FLAGS_NONCONTIGUOUS_MEM;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
// if no memory specific flags are set,
|
|
// default to EBI heap, so that bypass
|
|
// can work. We can fall back to system
|
|
// heap if we run out.
|
|
ret = mPmemUserspaceAlloc->alloc_buffer(data);
|
|
|
|
// Fallback
|
|
if(ret >= 0 ) {
|
|
data.allocType = private_handle_t::PRIV_FLAGS_USES_PMEM;
|
|
} else if(ret < 0 && canFallback(usage, false)) {
|
|
ALOGW("Falling back to ashmem");
|
|
ret = mAshmemAlloc->alloc_buffer(data);
|
|
if(ret >= 0) {
|
|
data.allocType = private_handle_t::PRIV_FLAGS_USES_ASHMEM;
|
|
data.allocType |= private_handle_t::PRIV_FLAGS_NONCONTIGUOUS_MEM;
|
|
}
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
IMemAlloc* PmemAshmemController::getAllocator(int flags)
|
|
{
|
|
IMemAlloc* memalloc;
|
|
if (flags & private_handle_t::PRIV_FLAGS_USES_PMEM)
|
|
memalloc = mPmemUserspaceAlloc;
|
|
else if (flags & private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP)
|
|
memalloc = mPmemKernelCtrl->getAllocator(flags);
|
|
else if (flags & private_handle_t::PRIV_FLAGS_USES_ASHMEM)
|
|
memalloc = mAshmemAlloc;
|
|
else {
|
|
ALOGE("%s: Invalid flags passed: 0x%x", __FUNCTION__, flags);
|
|
memalloc = NULL;
|
|
}
|
|
|
|
return memalloc;
|
|
}
|
|
#endif
|
|
|
|
size_t getBufferSizeAndDimensions(int width, int height, int format,
|
|
int& alignedw, int &alignedh)
|
|
{
|
|
size_t size;
|
|
|
|
alignedw = ALIGN(width, 32);
|
|
alignedh = ALIGN(height, 32);
|
|
switch (format) {
|
|
case HAL_PIXEL_FORMAT_RGBA_8888:
|
|
case HAL_PIXEL_FORMAT_RGBX_8888:
|
|
case HAL_PIXEL_FORMAT_BGRA_8888:
|
|
size = alignedw * alignedh * 4;
|
|
break;
|
|
case HAL_PIXEL_FORMAT_RGB_888:
|
|
size = alignedw * alignedh * 3;
|
|
break;
|
|
case HAL_PIXEL_FORMAT_RGB_565:
|
|
case HAL_PIXEL_FORMAT_RGBA_5551:
|
|
case HAL_PIXEL_FORMAT_RGBA_4444:
|
|
size = alignedw * alignedh * 2;
|
|
break;
|
|
|
|
// adreno formats
|
|
case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: // NV21
|
|
size = ALIGN(alignedw*alignedh, 4096);
|
|
size += ALIGN(2 * ALIGN(width/2, 32) * ALIGN(height/2, 32), 4096);
|
|
break;
|
|
case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: // NV12
|
|
// The chroma plane is subsampled,
|
|
// but the pitch in bytes is unchanged
|
|
// The GPU needs 4K alignment, but the video decoder needs 8K
|
|
alignedw = ALIGN(width, 128);
|
|
size = ALIGN( alignedw * alignedh, 8192);
|
|
size += ALIGN( alignedw * ALIGN(height/2, 32), 8192);
|
|
break;
|
|
case HAL_PIXEL_FORMAT_NV12_ENCODEABLE:
|
|
case HAL_PIXEL_FORMAT_YCbCr_420_SP:
|
|
case HAL_PIXEL_FORMAT_YCrCb_420_SP:
|
|
case HAL_PIXEL_FORMAT_YV12:
|
|
if ((format == HAL_PIXEL_FORMAT_YV12) && ((width&1) || (height&1))) {
|
|
ALOGE("w or h is odd for the YV12 format");
|
|
return -EINVAL;
|
|
}
|
|
alignedw = ALIGN(width, 16);
|
|
alignedh = height;
|
|
if (HAL_PIXEL_FORMAT_NV12_ENCODEABLE == format) {
|
|
// The encoder requires a 2K aligned chroma offset.
|
|
size = ALIGN(alignedw*alignedh, 2048) +
|
|
(ALIGN(alignedw/2, 16) * (alignedh/2))*2;
|
|
} else {
|
|
size = alignedw*alignedh +
|
|
(ALIGN(alignedw/2, 16) * (alignedh/2))*2;
|
|
}
|
|
size = ALIGN(size, 4096);
|
|
break;
|
|
case HAL_PIXEL_FORMAT_YCbCr_422_SP:
|
|
case HAL_PIXEL_FORMAT_YCrCb_422_SP:
|
|
if(width & 1) {
|
|
ALOGE("width is odd for the YUV422_SP format");
|
|
return -EINVAL;
|
|
}
|
|
alignedw = ALIGN(width, 16);
|
|
alignedh = height;
|
|
size = ALIGN(alignedw * alignedh * 2, 4096);
|
|
break;
|
|
default:
|
|
ALOGE("unrecognized pixel format: 0x%x", format);
|
|
return -EINVAL;
|
|
}
|
|
|
|
return size;
|
|
}
|
|
|
|
// Allocate buffer from width, height and format into a
|
|
// private_handle_t. It is the responsibility of the caller
|
|
// to free the buffer using the free_buffer function
|
|
int alloc_buffer(private_handle_t **pHnd, int w, int h, int format, int usage)
|
|
{
|
|
alloc_data data;
|
|
int alignedw, alignedh;
|
|
gralloc::IAllocController* sAlloc =
|
|
gralloc::IAllocController::getInstance(false);
|
|
data.base = 0;
|
|
data.fd = -1;
|
|
data.offset = 0;
|
|
data.size = getBufferSizeAndDimensions(w, h, format, alignedw, alignedh);
|
|
data.align = getpagesize();
|
|
data.uncached = useUncached(usage);
|
|
int allocFlags = usage;
|
|
|
|
int err = sAlloc->allocate(data, allocFlags);
|
|
if (0 != err) {
|
|
ALOGE("%s: allocate failed", __FUNCTION__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
private_handle_t* hnd = new private_handle_t(data.fd, data.size,
|
|
data.allocType, 0, format,
|
|
alignedw, alignedh);
|
|
hnd->base = (int) data.base;
|
|
hnd->offset = data.offset;
|
|
hnd->gpuaddr = 0;
|
|
*pHnd = hnd;
|
|
return 0;
|
|
}
|
|
|
|
void free_buffer(private_handle_t *hnd)
|
|
{
|
|
gralloc::IAllocController* sAlloc =
|
|
gralloc::IAllocController::getInstance(false);
|
|
if (hnd && hnd->fd > 0) {
|
|
IMemAlloc* memalloc = sAlloc->getAllocator(hnd->flags);
|
|
memalloc->free_buffer((void*)hnd->base, hnd->size, hnd->offset, hnd->fd);
|
|
}
|
|
if(hnd)
|
|
delete hnd;
|
|
|
|
}
|