3ae7f52726
Add support for the GRALLOC_USAGE_EXTERNAL_DISP and the GRALLOC_USAGE_PROTECTED usage bits. If any of the two usage bits are set, we need to send the buffers to the overlay. Since we send the buffers to the overlay, we need to allocate these buffers from the PMEM region. (cherry picked from commit 195345cd63f0e19cfee4cf055587d5fffe76d0ef) Change-Id: I64116b9be07c8ddadbd5a132ef21b19e5326737b
454 lines
15 KiB
C++
Executable File
454 lines
15 KiB
C++
Executable File
/*
|
|
* Copyright (C) 2010 The Android Open Source Project
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#include <limits.h>
|
|
#include <unistd.h>
|
|
#include <fcntl.h>
|
|
|
|
#include <sys/mman.h>
|
|
|
|
#include "gr.h"
|
|
#include "gpu.h"
|
|
|
|
static const int OMX_QCOM_COLOR_FormatYVU420SemiPlanar = 0x7FA30C00;
|
|
static const int QOMX_COLOR_FormatYUV420PackedSemiPlanar64x32Tile2m8ka = 0x7FA30C03;
|
|
static const int QOMX_INTERLACE_FLAG = 0x49283654;
|
|
static const int QOMX_3D_VIDEO_FLAG = 0x23784238;
|
|
|
|
gpu_context_t::gpu_context_t(Deps& deps, PmemAllocator& pmemAllocator,
|
|
PmemAllocator& pmemAdspAllocator, const private_module_t* module) :
|
|
deps(deps),
|
|
pmemAllocator(pmemAllocator),
|
|
pmemAdspAllocator(pmemAdspAllocator)
|
|
{
|
|
// Zero out the alloc_device_t
|
|
memset(static_cast<alloc_device_t*>(this), 0, sizeof(alloc_device_t));
|
|
|
|
// Initialize the procs
|
|
common.tag = HARDWARE_DEVICE_TAG;
|
|
common.version = 0;
|
|
common.module = const_cast<hw_module_t*>(&module->base.common);
|
|
common.close = gralloc_close;
|
|
alloc = gralloc_alloc;
|
|
allocSize = gralloc_alloc_size;
|
|
free = gralloc_free;
|
|
}
|
|
|
|
int gpu_context_t::gralloc_alloc_framebuffer_locked(size_t size, int usage,
|
|
buffer_handle_t* pHandle)
|
|
{
|
|
private_module_t* m = reinterpret_cast<private_module_t*>(common.module);
|
|
|
|
// we don't support allocations with both the FB and PMEM_ADSP flags
|
|
if (usage & GRALLOC_USAGE_PRIVATE_PMEM_ADSP) {
|
|
return -EINVAL;
|
|
}
|
|
|
|
// allocate the framebuffer
|
|
if (m->framebuffer == NULL) {
|
|
// initialize the framebuffer, the framebuffer is mapped once
|
|
// and forever.
|
|
int err = deps.mapFrameBufferLocked(m);
|
|
if (err < 0) {
|
|
return err;
|
|
}
|
|
}
|
|
|
|
const uint32_t bufferMask = m->bufferMask;
|
|
const uint32_t numBuffers = m->numBuffers;
|
|
const size_t bufferSize = m->finfo.line_length * m->info.yres;
|
|
if (numBuffers == 1) {
|
|
// If we have only one buffer, we never use page-flipping. Instead,
|
|
// we return a regular buffer which will be memcpy'ed to the main
|
|
// screen when post is called.
|
|
int newUsage = (usage & ~GRALLOC_USAGE_HW_FB) | GRALLOC_USAGE_HW_2D;
|
|
return gralloc_alloc_buffer(bufferSize, newUsage, pHandle, BUFFER_TYPE_UI, m->fbFormat, m->info.xres, m->info.yres);
|
|
}
|
|
|
|
if (bufferMask >= ((1LU<<numBuffers)-1)) {
|
|
// We ran out of buffers.
|
|
return -ENOMEM;
|
|
}
|
|
|
|
// create a "fake" handles for it
|
|
intptr_t vaddr = intptr_t(m->framebuffer->base);
|
|
private_handle_t* hnd = new private_handle_t(dup(m->framebuffer->fd), bufferSize,
|
|
private_handle_t::PRIV_FLAGS_USES_PMEM |
|
|
private_handle_t::PRIV_FLAGS_FRAMEBUFFER,
|
|
BUFFER_TYPE_UI, m->fbFormat, m->info.xres, m->info.yres);
|
|
|
|
// find a free slot
|
|
for (uint32_t i=0 ; i<numBuffers ; i++) {
|
|
if ((bufferMask & (1LU<<i)) == 0) {
|
|
m->bufferMask |= (1LU<<i);
|
|
break;
|
|
}
|
|
vaddr += bufferSize;
|
|
}
|
|
|
|
hnd->base = vaddr;
|
|
hnd->offset = vaddr - intptr_t(m->framebuffer->base);
|
|
*pHandle = hnd;
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
int gpu_context_t::gralloc_alloc_framebuffer(size_t size, int usage,
|
|
buffer_handle_t* pHandle)
|
|
{
|
|
private_module_t* m = reinterpret_cast<private_module_t*>(common.module);
|
|
pthread_mutex_lock(&m->lock);
|
|
int err = gralloc_alloc_framebuffer_locked(size, usage, pHandle);
|
|
pthread_mutex_unlock(&m->lock);
|
|
return err;
|
|
}
|
|
|
|
int gpu_context_t::alloc_ashmem_buffer(size_t size, unsigned int postfix, void** pBase,
|
|
int* pOffset, int* pFd)
|
|
{
|
|
int err = 0;
|
|
int fd = -1;
|
|
void* base = 0;
|
|
int offset = 0;
|
|
|
|
char name[ASHMEM_NAME_LEN];
|
|
snprintf(name, ASHMEM_NAME_LEN, "gralloc-buffer-%x", postfix);
|
|
int prot = PROT_READ | PROT_WRITE;
|
|
fd = ashmem_create_region(name, size);
|
|
if (fd < 0) {
|
|
LOGE("couldn't create ashmem (%s)", strerror(errno));
|
|
err = -errno;
|
|
} else {
|
|
if (ashmem_set_prot_region(fd, prot) < 0) {
|
|
LOGE("ashmem_set_prot_region(fd=%d, prot=%x) failed (%s)",
|
|
fd, prot, strerror(errno));
|
|
close(fd);
|
|
err = -errno;
|
|
} else {
|
|
base = mmap(0, size, prot, MAP_SHARED|MAP_POPULATE|MAP_LOCKED, fd, 0);
|
|
if (base == MAP_FAILED) {
|
|
LOGE("alloc mmap(fd=%d, size=%d, prot=%x) failed (%s)",
|
|
fd, size, prot, strerror(errno));
|
|
close(fd);
|
|
err = -errno;
|
|
} else {
|
|
memset((char*)base + offset, 0, size);
|
|
}
|
|
}
|
|
}
|
|
if(err == 0) {
|
|
*pFd = fd;
|
|
*pBase = base;
|
|
*pOffset = offset;
|
|
}
|
|
return err;
|
|
}
|
|
|
|
int gpu_context_t::gralloc_alloc_buffer(size_t size, int usage, buffer_handle_t* pHandle,
|
|
int bufferType, int format, int width, int height)
|
|
{
|
|
int err = 0;
|
|
int flags = 0;
|
|
|
|
int fd = -1;
|
|
void* base = 0; // XXX JMG: This should change to just get an address from
|
|
// the PmemAllocator rather than getting the base & offset separately
|
|
int offset = 0;
|
|
int lockState = 0;
|
|
|
|
size = roundUpToPageSize(size);
|
|
#ifndef USE_ASHMEM
|
|
if (usage & GRALLOC_USAGE_HW_TEXTURE) {
|
|
// enable pmem in that case, so our software GL can fallback to
|
|
// the copybit module.
|
|
flags |= private_handle_t::PRIV_FLAGS_USES_PMEM;
|
|
}
|
|
|
|
if (usage & GRALLOC_USAGE_HW_2D) {
|
|
flags |= private_handle_t::PRIV_FLAGS_USES_PMEM;
|
|
}
|
|
#endif
|
|
if ((usage & GRALLOC_USAGE_PRIVATE_PMEM_ADSP) || (usage & GRALLOC_USAGE_PRIVATE_PMEM_SMIPOOL)
|
|
|| (usage & GRALLOC_USAGE_EXTERNAL_DISP) || (usage & GRALLOC_USAGE_PROTECTED)) {
|
|
flags |= private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP;
|
|
flags &= ~private_handle_t::PRIV_FLAGS_USES_PMEM;
|
|
}
|
|
|
|
private_module_t* m = reinterpret_cast<private_module_t*>(common.module);
|
|
if((flags & private_handle_t::PRIV_FLAGS_USES_PMEM) == 0 &&
|
|
(flags & private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP) == 0) {
|
|
flags |= private_handle_t::PRIV_FLAGS_USES_ASHMEM;
|
|
err = alloc_ashmem_buffer(size, (unsigned int)pHandle, &base, &offset, &fd);
|
|
if(err >= 0)
|
|
lockState |= private_handle_t::LOCK_STATE_MAPPED;
|
|
}
|
|
else if ((flags & private_handle_t::PRIV_FLAGS_USES_PMEM) != 0 ||
|
|
(flags & private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP) != 0) {
|
|
|
|
PmemAllocator* pma = 0;
|
|
|
|
if ((flags & private_handle_t::PRIV_FLAGS_USES_PMEM) != 0) {
|
|
if ((flags & private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP) != 0) {
|
|
LOGE("attempting to allocate a gralloc buffer with both the "
|
|
"USES_PMEM and USES_PMEM_ADSP flags. Unsetting the "
|
|
"USES_PMEM_ADSP flag.");
|
|
flags &= ~private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP;
|
|
}
|
|
pma = &pmemAllocator;
|
|
} else { // (flags & private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP) != 0
|
|
pma = &pmemAdspAllocator;
|
|
}
|
|
|
|
// PMEM buffers are always mmapped
|
|
lockState |= private_handle_t::LOCK_STATE_MAPPED;
|
|
|
|
err = pma->alloc_pmem_buffer(size, usage, &base, &offset, &fd, format);
|
|
if (err < 0) {
|
|
if (((usage & GRALLOC_USAGE_HW_MASK) == 0) &&
|
|
((usage & GRALLOC_USAGE_PRIVATE_PMEM_ADSP) == 0)) {
|
|
// the caller didn't request PMEM, so we can try something else
|
|
flags &= ~private_handle_t::PRIV_FLAGS_USES_PMEM;
|
|
err = 0;
|
|
goto try_ashmem;
|
|
} else {
|
|
LOGE("couldn't open pmem (%s)", strerror(errno));
|
|
}
|
|
}
|
|
} else {
|
|
try_ashmem:
|
|
fd = deps.ashmem_create_region("gralloc-buffer", size);
|
|
if (fd < 0) {
|
|
LOGE("couldn't create ashmem (%s)", strerror(errno));
|
|
err = -errno;
|
|
}
|
|
}
|
|
|
|
if (err == 0) {
|
|
private_handle_t* hnd = new private_handle_t(fd, size, flags, bufferType, format, width, height);
|
|
hnd->offset = offset;
|
|
hnd->base = int(base)+offset;
|
|
hnd->lockState = lockState;
|
|
*pHandle = hnd;
|
|
}
|
|
|
|
LOGE_IF(err, "gralloc failed err=%s", strerror(-err));
|
|
|
|
return err;
|
|
}
|
|
|
|
static inline size_t ALIGN(size_t x, size_t align) {
|
|
return (x + align-1) & ~(align-1);
|
|
}
|
|
|
|
void gpu_context_t::getGrallocInformationFromFormat(int inputFormat, int *colorFormat, int *bufferType)
|
|
{
|
|
*bufferType = BUFFER_TYPE_VIDEO;
|
|
*colorFormat = inputFormat;
|
|
|
|
if (inputFormat == HAL_PIXEL_FORMAT_YV12) {
|
|
*bufferType = BUFFER_TYPE_VIDEO;
|
|
} else if (inputFormat & S3D_FORMAT_MASK) {
|
|
// S3D format
|
|
*colorFormat = COLOR_FORMAT(inputFormat);
|
|
} else if (inputFormat & INTERLACE_MASK) {
|
|
// Interlaced
|
|
*colorFormat = inputFormat ^ HAL_PIXEL_FORMAT_INTERLACE;
|
|
} else if (inputFormat < 0x7) {
|
|
// RGB formats
|
|
*colorFormat = inputFormat;
|
|
*bufferType = BUFFER_TYPE_UI;
|
|
} else if ((inputFormat == HAL_PIXEL_FORMAT_R_8) ||
|
|
(inputFormat == HAL_PIXEL_FORMAT_RG_88)) {
|
|
*colorFormat = inputFormat;
|
|
*bufferType = BUFFER_TYPE_UI;
|
|
}
|
|
}
|
|
|
|
int gpu_context_t::alloc_impl(int w, int h, int format, int usage,
|
|
buffer_handle_t* pHandle, int* pStride, int bufferSize) {
|
|
if (!pHandle || !pStride)
|
|
return -EINVAL;
|
|
|
|
size_t size, alignedw, alignedh;
|
|
|
|
alignedw = ALIGN(w, 32);
|
|
alignedh = ALIGN(h, 32);
|
|
int colorFormat, bufferType;
|
|
getGrallocInformationFromFormat(format, &colorFormat, &bufferType);
|
|
|
|
switch (colorFormat) {
|
|
case HAL_PIXEL_FORMAT_RGBA_8888:
|
|
case HAL_PIXEL_FORMAT_RGBX_8888:
|
|
case HAL_PIXEL_FORMAT_BGRA_8888:
|
|
size = alignedw * alignedh * 4;
|
|
break;
|
|
case HAL_PIXEL_FORMAT_RGB_888:
|
|
size = alignedw * alignedh * 3;
|
|
break;
|
|
case HAL_PIXEL_FORMAT_RGB_565:
|
|
case HAL_PIXEL_FORMAT_RGBA_5551:
|
|
case HAL_PIXEL_FORMAT_RGBA_4444:
|
|
size = alignedw * alignedh * 2;
|
|
break;
|
|
|
|
// adreno formats
|
|
case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: // NV21
|
|
size = ALIGN(alignedw*alignedh, 4096);
|
|
size += ALIGN(2 * ALIGN(w/2, 32) * ALIGN(h/2, 32), 4096);
|
|
break;
|
|
case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: // NV12
|
|
// The chroma plane is subsampled,
|
|
// but the pitch in bytes is unchanged
|
|
// The GPU needs 4K alignment, but the video decoder needs 8K
|
|
alignedw = ALIGN(w, 128);
|
|
size = ALIGN( alignedw * alignedh, 8192);
|
|
size += ALIGN( alignedw * ALIGN(h/2, 32), 8192);
|
|
break;
|
|
case HAL_PIXEL_FORMAT_YCbCr_420_SP:
|
|
case HAL_PIXEL_FORMAT_YCrCb_420_SP:
|
|
case HAL_PIXEL_FORMAT_YV12:
|
|
if ((w&1) || (h&1)) {
|
|
LOGE("w or h is odd for the YUV format");
|
|
return -EINVAL;
|
|
}
|
|
alignedw = ALIGN(w, 16);
|
|
alignedh = h;
|
|
size = alignedw*alignedh +
|
|
(ALIGN(alignedw/2, 16) * (alignedh/2))*2;
|
|
size = ALIGN(size, 4096);
|
|
break;
|
|
|
|
default:
|
|
LOGE("unrecognized pixel format: %d", format);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if ((ssize_t)size <= 0)
|
|
return -EINVAL;
|
|
|
|
size = (bufferSize >= size)? bufferSize : size;
|
|
|
|
// All buffers marked as protected or for external
|
|
// display need to go to overlay
|
|
if ((usage & GRALLOC_USAGE_EXTERNAL_DISP) ||
|
|
(usage & GRALLOC_USAGE_PROTECTED)) {
|
|
bufferType = BUFFER_TYPE_VIDEO;
|
|
}
|
|
int err;
|
|
if (usage & GRALLOC_USAGE_HW_FB) {
|
|
err = gralloc_alloc_framebuffer(size, usage, pHandle);
|
|
} else {
|
|
err = gralloc_alloc_buffer(size, usage, pHandle, bufferType, format, w, h);
|
|
}
|
|
|
|
if (err < 0) {
|
|
return err;
|
|
}
|
|
|
|
*pStride = alignedw;
|
|
return 0;
|
|
}
|
|
|
|
int gpu_context_t::free_impl(private_handle_t const* hnd) {
|
|
private_module_t* m = reinterpret_cast<private_module_t*>(common.module);
|
|
if (hnd->flags & private_handle_t::PRIV_FLAGS_FRAMEBUFFER) {
|
|
// free this buffer
|
|
const size_t bufferSize = m->finfo.line_length * m->info.yres;
|
|
int index = (hnd->base - m->framebuffer->base) / bufferSize;
|
|
m->bufferMask &= ~(1<<index);
|
|
} else {
|
|
PmemAllocator* pmem_allocator = 0;
|
|
if (hnd->flags & private_handle_t::PRIV_FLAGS_USES_PMEM) {
|
|
pmem_allocator = &pmemAllocator;
|
|
} else if (hnd->flags & private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP) {
|
|
pmem_allocator = &pmemAdspAllocator;
|
|
} else if (hnd->flags & private_handle_t::PRIV_FLAGS_USES_ASHMEM) {
|
|
// free ashmem
|
|
if (hnd->fd >= 0) {
|
|
if (hnd->base) {
|
|
int err = munmap((void*)hnd->base, hnd->size);
|
|
LOGE_IF(err<0, "ASHMEM_UNMAP failed (%s), "
|
|
"fd=%d, sub.offset=%d, sub.size=%d",
|
|
strerror(errno), hnd->fd, hnd->offset, hnd->size);
|
|
}
|
|
}
|
|
}
|
|
if (pmem_allocator) {
|
|
pmem_allocator->free_pmem_buffer(hnd->size, (void*)hnd->base,
|
|
hnd->offset, hnd->fd);
|
|
}
|
|
|
|
deps.terminateBuffer(&m->base, const_cast<private_handle_t*>(hnd));
|
|
}
|
|
|
|
deps.close(hnd->fd);
|
|
delete hnd; // XXX JMG: move this to the deps
|
|
return 0;
|
|
}
|
|
|
|
/******************************************************************************
|
|
* Static functions
|
|
*****************************************************************************/
|
|
|
|
int gpu_context_t::gralloc_alloc(alloc_device_t* dev, int w, int h, int format,
|
|
int usage, buffer_handle_t* pHandle, int* pStride)
|
|
{
|
|
if (!dev) {
|
|
return -EINVAL;
|
|
}
|
|
gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev);
|
|
return gpu->alloc_impl(w, h, format, usage, pHandle, pStride, 0);
|
|
}
|
|
|
|
int gpu_context_t::gralloc_alloc_size(alloc_device_t* dev, int w, int h, int format,
|
|
int usage, buffer_handle_t* pHandle, int* pStride, int bufferSize)
|
|
{
|
|
if (!dev) {
|
|
return -EINVAL;
|
|
}
|
|
gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev);
|
|
return gpu->alloc_impl(w, h, format, usage, pHandle, pStride, bufferSize);
|
|
}
|
|
|
|
int gpu_context_t::gralloc_free(alloc_device_t* dev,
|
|
buffer_handle_t handle)
|
|
{
|
|
if (private_handle_t::validate(handle) < 0)
|
|
return -EINVAL;
|
|
|
|
private_handle_t const* hnd = reinterpret_cast<private_handle_t const*>(handle);
|
|
gpu_context_t* gpu = reinterpret_cast<gpu_context_t*>(dev);
|
|
return gpu->free_impl(hnd);
|
|
}
|
|
|
|
/*****************************************************************************/
|
|
|
|
int gpu_context_t::gralloc_close(struct hw_device_t *dev)
|
|
{
|
|
gpu_context_t* ctx = reinterpret_cast<gpu_context_t*>(dev);
|
|
if (ctx) {
|
|
/* TODO: keep a list of all buffer_handle_t created, and free them
|
|
* all here.
|
|
*/
|
|
delete ctx;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
|
|
gpu_context_t::Deps::~Deps() {}
|