/* * Copyright (C) 2010 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ //#define LOG_NDEBUG 0 #include #include #include #include #include #include #include #include #include #include "gralloc_priv.h" #include "pmemalloc.h" #define BEGIN_FUNC LOGV("%s begin", __PRETTY_FUNCTION__) #define END_FUNC LOGV("%s end", __PRETTY_FUNCTION__) static int get_open_flags(int usage) { int openFlags = O_RDWR | O_SYNC; uint32_t uread = usage & GRALLOC_USAGE_SW_READ_MASK; uint32_t uwrite = usage & GRALLOC_USAGE_SW_WRITE_MASK; if (uread == GRALLOC_USAGE_SW_READ_OFTEN || uwrite == GRALLOC_USAGE_SW_WRITE_OFTEN) { openFlags &= ~O_SYNC; } return openFlags; } PmemAllocator::~PmemAllocator() { BEGIN_FUNC; END_FUNC; } PmemUserspaceAllocator::PmemUserspaceAllocator(Deps& deps, Deps::Allocator& allocator, const char* pmemdev): deps(deps), allocator(allocator), pmemdev(pmemdev), master_fd(MASTER_FD_INIT) { BEGIN_FUNC; pthread_mutex_init(&lock, NULL); END_FUNC; } PmemUserspaceAllocator::~PmemUserspaceAllocator() { BEGIN_FUNC; END_FUNC; } void* PmemUserspaceAllocator::get_base_address() { BEGIN_FUNC; END_FUNC; return master_base; } int PmemUserspaceAllocator::init_pmem_area_locked() { BEGIN_FUNC; int err = 0; int fd = deps.open(pmemdev, O_RDWR, 0); if (fd >= 0) { size_t size = 0; err = deps.getPmemTotalSize(fd, &size); if (err < 0) { LOGE("%s: PMEM_GET_TOTAL_SIZE failed (%d), limp mode", pmemdev, err); size = 8<<20; // 8 MiB } allocator.setSize(size); void* base = deps.mmap(0, size, PROT_READ|PROT_WRITE, MAP_SHARED, fd, 0); if (base == MAP_FAILED) { LOGE("%s: failed to map pmem master fd: %s", pmemdev, strerror(deps.getErrno())); err = -deps.getErrno(); base = 0; deps.close(fd); fd = -1; } else { master_fd = fd; master_base = base; } } else { LOGE("%s: failed to open pmem device: %s", pmemdev, strerror(deps.getErrno())); err = -deps.getErrno(); } END_FUNC; return err; } int PmemUserspaceAllocator::init_pmem_area() { BEGIN_FUNC; pthread_mutex_lock(&lock); int err = master_fd; if (err == MASTER_FD_INIT) { // first time, try to initialize pmem err = init_pmem_area_locked(); if (err) { LOGE("%s: failed to initialize pmem area", pmemdev); master_fd = err; } } else if (err < 0) { // pmem couldn't be initialized, never use it } else { // pmem OK err = 0; } pthread_mutex_unlock(&lock); END_FUNC; return err; } int PmemUserspaceAllocator::alloc_pmem_buffer(size_t size, int usage, void** pBase, int* pOffset, int* pFd, int format) { BEGIN_FUNC; int err = init_pmem_area(); if (err == 0) { void* base = master_base; int offset = allocator.allocate(size); if (offset < 0) { // no more pmem memory LOGE("%s: no more pmem available", pmemdev); err = -ENOMEM; } else { int openFlags = get_open_flags(usage); //LOGD("%s: allocating pmem at offset 0x%p", pmemdev, offset); // now create the "sub-heap" int fd = deps.open(pmemdev, openFlags, 0); err = fd < 0 ? fd : 0; // and connect to it if (err == 0) err = deps.connectPmem(fd, master_fd); // and make it available to the client process if (err == 0) err = deps.mapPmem(fd, offset, size); if (err < 0) { LOGE("%s: failed to initialize pmem sub-heap: %d", pmemdev, err); err = -deps.getErrno(); deps.close(fd); allocator.deallocate(offset); fd = -1; } else { LOGV("%s: mapped fd %d at offset %d, size %d", pmemdev, fd, offset, size); memset((char*)base + offset, 0, size); //Clean cache before flushing to ensure pmem is properly flushed err = deps.cleanPmem(fd, (unsigned long) base + offset, offset, size); if (err < 0) { LOGE("cleanPmem failed: (%s)", strerror(deps.getErrno())); } #ifdef HOST cacheflush(intptr_t(base) + offset, intptr_t(base) + offset + size, 0); #endif *pBase = base; *pOffset = offset; *pFd = fd; } //LOGD_IF(!err, "%s: allocating pmem size=%d, offset=%d", pmemdev, size, offset); } } END_FUNC; return err; } int PmemUserspaceAllocator::free_pmem_buffer(size_t size, void* base, int offset, int fd) { BEGIN_FUNC; int err = 0; if (fd >= 0) { int err = deps.unmapPmem(fd, offset, size); LOGE_IF(err<0, "PMEM_UNMAP failed (%s), fd=%d, sub.offset=%u, " "sub.size=%u", strerror(deps.getErrno()), fd, offset, size); if (err == 0) { // we can't deallocate the memory in case of UNMAP failure // because it would give that process access to someone else's // surfaces, which would be a security breach. allocator.deallocate(offset); } } END_FUNC; return err; } PmemUserspaceAllocator::Deps::Allocator::~Allocator() { BEGIN_FUNC; END_FUNC; } PmemUserspaceAllocator::Deps::~Deps() { BEGIN_FUNC; END_FUNC; } PmemKernelAllocator::PmemKernelAllocator(Deps& deps): deps(deps) { BEGIN_FUNC; END_FUNC; } PmemKernelAllocator::~PmemKernelAllocator() { BEGIN_FUNC; END_FUNC; } void* PmemKernelAllocator::get_base_address() { BEGIN_FUNC; END_FUNC; return 0; } static unsigned clp2(unsigned x) { x = x - 1; x = x | (x >> 1); x = x | (x >> 2); x = x | (x >> 4); x = x | (x >> 8); x = x | (x >>16); return x + 1; } int PmemKernelAllocator::alloc_pmem_buffer(size_t size, int usage, void** pBase,int* pOffset, int* pFd, int format) { BEGIN_FUNC; *pBase = 0; *pOffset = 0; *pFd = -1; int err, offset = 0; int openFlags = get_open_flags(usage); const char *device; if (usage & GRALLOC_USAGE_PRIVATE_PMEM_ADSP) { device = DEVICE_PMEM_ADSP; } else if (usage & GRALLOC_USAGE_PRIVATE_PMEM_SMIPOOL) { device = DEVICE_PMEM_SMIPOOL; } else if ((usage & GRALLOC_USAGE_EXTERNAL_DISP) || (usage & GRALLOC_USAGE_PROTECTED)) { int tempFd = deps.open(DEVICE_PMEM_SMIPOOL, openFlags, 0); if (tempFd < 0) { device = DEVICE_PMEM_ADSP; } else { close(tempFd); device = DEVICE_PMEM_SMIPOOL; } } else { LOGE("Invalid device"); return -EINVAL; } int fd = deps.open(device, openFlags, 0); if (fd < 0) { err = -deps.getErrno(); END_FUNC; LOGE("Error opening %s", device); return err; } // The size should already be page aligned, now round it up to a power of 2. //size = clp2(size); if (format == HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED) { // Tile format buffers need physical alignment to 8K err = deps.alignPmem(fd, size, 8192); if (err < 0) { LOGE("alignPmem failed"); } } void* base = deps.mmap(0, size, PROT_READ|PROT_WRITE, MAP_SHARED, fd, 0); if (base == MAP_FAILED) { LOGE("%s: failed to map pmem fd: %s", device, strerror(deps.getErrno())); err = -deps.getErrno(); deps.close(fd); END_FUNC; return err; } memset(base, 0, size); *pBase = base; *pOffset = 0; *pFd = fd; END_FUNC; return 0; } int PmemKernelAllocator::free_pmem_buffer(size_t size, void* base, int offset, int fd) { BEGIN_FUNC; // The size should already be page aligned, now round it up to a power of 2 // like we did when allocating. //size = clp2(size); int err = deps.munmap(base, size); if (err < 0) { err = deps.getErrno(); LOGW("error unmapping pmem fd: %s", strerror(err)); return -err; } END_FUNC; return 0; } PmemKernelAllocator::Deps::~Deps() { BEGIN_FUNC; END_FUNC; }