/* * Copyright (C) 2010 ARM Limited. All rights reserved. * * Copyright (C) 2008 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include #include #include #include #include #include #include #include #include #include #include #include #include "alloc_device.h" #include "gralloc_priv.h" #include "gralloc_helper.h" #include "framebuffer_device.h" #if GRALLOC_ARM_UMP_MODULE #include #include #endif #if GRALLOC_ARM_DMA_BUF_MODULE #include #include "ion_4.12.h" #include "dma-heap.h" #define ION_SYSTEM (char*)"ion_system_heap" #define ION_CMA (char*)"linux,cma" #define DMABUF_SYSTEM (char*)"system" #define DMABUF_CMA (char*)"linux,cma" static enum { INTERFACE_UNKNOWN, INTERFACE_ION_LEGACY, INTERFACE_ION_MODERN, INTERFACE_DMABUF_HEAPS } interface_ver; static int system_heap_id; static int cma_heap_id; #endif #if GRALLOC_SIMULATE_FAILURES #include /* system property keys for controlling simulated UMP allocation failures */ #define PROP_MALI_TEST_GRALLOC_FAIL_FIRST "mali.test.gralloc.fail_first" #define PROP_MALI_TEST_GRALLOC_FAIL_INTERVAL "mali.test.gralloc.fail_interval" static int __ump_alloc_should_fail() { static unsigned int call_count = 0; unsigned int first_fail = 0; int fail_period = 0; int fail = 0; ++call_count; /* read the system properties that control failure simulation */ { char prop_value[PROPERTY_VALUE_MAX]; if (property_get(PROP_MALI_TEST_GRALLOC_FAIL_FIRST, prop_value, "0") > 0) { sscanf(prop_value, "%11u", &first_fail); } if (property_get(PROP_MALI_TEST_GRALLOC_FAIL_INTERVAL, prop_value, "0") > 0) { sscanf(prop_value, "%11u", &fail_period); } } /* failure simulation is enabled by setting the first_fail property to non-zero */ if (first_fail > 0) { LOGI("iteration %u (fail=%u, period=%u)\n", call_count, first_fail, fail_period); fail = (call_count == first_fail) || (call_count > first_fail && fail_period > 0 && 0 == (call_count - first_fail) % fail_period); if (fail) { AERR("failed ump_ref_drv_allocate on iteration #%d\n", call_count); } } return fail; } #endif #ifdef FBIOGET_DMABUF static int fb_get_framebuffer_dmabuf(private_module_t *m, private_handle_t *hnd) { struct fb_dmabuf_export fb_dma_buf; int res; res = ioctl(m->framebuffer->fd, FBIOGET_DMABUF, &fb_dma_buf); if (res == 0) { hnd->share_fd = fb_dma_buf.fd; return 0; } else { AINF("FBIOGET_DMABUF ioctl failed(%d). See gralloc_priv.h and the integration manual for vendor framebuffer " "integration", res); return -1; } } #endif #if GRALLOC_ARM_DMA_BUF_MODULE #define DEVPATH "/dev/dma_heap" int dma_heap_open(const char* name) { int ret, fd; char buf[256]; ret = sprintf(buf, "%s/%s", DEVPATH, name); if (ret < 0) { AERR("sprintf failed!\n"); return ret; } fd = open(buf, O_RDONLY); if (fd < 0) AERR("open %s failed!\n", buf); return fd; } int dma_heap_alloc(int fd, size_t len, unsigned int flags, int *dmabuf_fd) { struct dma_heap_allocation_data data = { .len = len, .fd_flags = O_RDWR | O_CLOEXEC, .heap_flags = flags, }; int ret; if (dmabuf_fd == NULL) return -EINVAL; ret = ioctl(fd, DMA_HEAP_IOCTL_ALLOC, &data); if (ret < 0) return ret; *dmabuf_fd = (int)data.fd; return ret; } static int alloc_ion_fd(int ion_fd, size_t size, unsigned int heap_mask, unsigned int flags, int *shared_fd) { int heap; if (interface_ver == INTERFACE_DMABUF_HEAPS) { int fd = system_heap_id; unsigned long flg = 0; if (heap_mask == ION_HEAP_TYPE_DMA_MASK) fd = cma_heap_id; return dma_heap_alloc(fd, size, flg, shared_fd); } if (interface_ver == INTERFACE_ION_MODERN) { heap = 1 << system_heap_id; if (heap_mask == ION_HEAP_TYPE_DMA_MASK) heap = 1 << cma_heap_id; } else { heap = heap_mask; } return ion_alloc_fd(ion_fd, size, 0, heap, flags, shared_fd); } #endif static int gralloc_alloc_buffer(alloc_device_t *dev, size_t size, int usage, buffer_handle_t *pHandle) { #if GRALLOC_ARM_DMA_BUF_MODULE { private_module_t *m = reinterpret_cast(dev->common.module); void *cpu_ptr = MAP_FAILED; int shared_fd; int ret; unsigned int heap_mask; int lock_state = 0; int map_mask = 0; if (usage & GRALLOC_USAGE_PROTECTED) { #if defined(ION_HEAP_SECURE_MASK) heap_mask = ION_HEAP_SECURE_MASK; #else AERR("The platform does NOT support protected ION memory."); return -1; #endif } else if (usage & GRALLOC_USAGE_HW_FB) { heap_mask = ION_HEAP_TYPE_DMA_MASK; } else { heap_mask = ION_HEAP_SYSTEM_MASK; } ret = alloc_ion_fd(m->ion_client, size, heap_mask, 0, &shared_fd); if (ret != 0) { AERR("Failed to ion_alloc_fd from ion_client:%d", m->ion_client); return -1; } if (!(usage & GRALLOC_USAGE_PROTECTED)) { map_mask = PROT_READ | PROT_WRITE; } else { map_mask = PROT_WRITE; } cpu_ptr = mmap(NULL, size, map_mask, MAP_SHARED, shared_fd, 0); if (MAP_FAILED == cpu_ptr) { AERR("ion_map( %d ) failed", m->ion_client); close(shared_fd); return -1; } lock_state = private_handle_t::LOCK_STATE_MAPPED; private_handle_t *hnd = new private_handle_t(private_handle_t::PRIV_FLAGS_USES_ION, usage, size, cpu_ptr, lock_state); if (NULL != hnd) { hnd->share_fd = shared_fd; *pHandle = hnd; return 0; } else { AERR("Gralloc out of mem for ion_client:%d", m->ion_client); } close(shared_fd); ret = munmap(cpu_ptr, size); if (0 != ret) { AERR("munmap failed for base:%p size: %lu", cpu_ptr, (unsigned long)size); } return -1; } #endif #if GRALLOC_ARM_UMP_MODULE MALI_IGNORE(dev); { ump_handle ump_mem_handle; void *cpu_ptr; ump_secure_id ump_id; ump_alloc_constraints constraints; size = round_up_to_page_size(size); if ((usage & GRALLOC_USAGE_SW_READ_MASK) == GRALLOC_USAGE_SW_READ_OFTEN) { constraints = UMP_REF_DRV_CONSTRAINT_USE_CACHE; } else { constraints = UMP_REF_DRV_CONSTRAINT_NONE; } #ifdef GRALLOC_SIMULATE_FAILURES /* if the failure condition matches, fail this iteration */ if (__ump_alloc_should_fail()) { ump_mem_handle = UMP_INVALID_MEMORY_HANDLE; } else #endif { if (usage & GRALLOC_USAGE_PROTECTED) { AERR("gralloc_alloc_buffer() does not support to allocate protected UMP memory."); } else { ump_mem_handle = ump_ref_drv_allocate(size, constraints); if (UMP_INVALID_MEMORY_HANDLE != ump_mem_handle) { cpu_ptr = ump_mapped_pointer_get(ump_mem_handle); if (NULL != cpu_ptr) { ump_id = ump_secure_id_get(ump_mem_handle); if (UMP_INVALID_SECURE_ID != ump_id) { private_handle_t *hnd = new private_handle_t(private_handle_t::PRIV_FLAGS_USES_UMP, usage, size, cpu_ptr, private_handle_t::LOCK_STATE_MAPPED, ump_id, ump_mem_handle); if (NULL != hnd) { *pHandle = hnd; return 0; } else { AERR("gralloc_alloc_buffer() failed to allocate handle. ump_handle = %p, ump_id = %d", ump_mem_handle, ump_id); } } else { AERR("gralloc_alloc_buffer() failed to retrieve valid secure id. ump_handle = %p", ump_mem_handle); } ump_mapped_pointer_release(ump_mem_handle); } else { AERR("gralloc_alloc_buffer() failed to map UMP memory. ump_handle = %p", ump_mem_handle); } ump_reference_release(ump_mem_handle); } else { AERR("gralloc_alloc_buffer() failed to allocate UMP memory. size:%d constraints: %d", size, constraints); } } } return -1; } #endif } #ifndef DISABLE_FRAMEBUFFER_HAL static int gralloc_alloc_framebuffer_locked(alloc_device_t *dev, size_t size, int usage, buffer_handle_t *pHandle) { private_module_t *m = reinterpret_cast(dev->common.module); // allocate the framebuffer if (m->framebuffer == NULL) { // initialize the framebuffer, the framebuffer is mapped once and forever. int err = init_frame_buffer_locked(m); if (err < 0) { return err; } } uint32_t bufferMask = m->bufferMask; const uint32_t numBuffers = m->numBuffers; const size_t bufferSize = m->finfo.line_length * m->info.yres; if (numBuffers == 1) { // If we have only one buffer, we never use page-flipping. Instead, // we return a regular buffer which will be memcpy'ed to the main // screen when post is called. int newUsage = (usage & ~GRALLOC_USAGE_HW_FB) | GRALLOC_USAGE_HW_2D; AERR("fallback to single buffering. Virtual Y-res too small %d", m->info.yres); return gralloc_alloc_buffer(dev, bufferSize, newUsage, pHandle); } if (bufferMask >= ((1LU << numBuffers) - 1)) { // We ran out of buffers, reset bufferMask. bufferMask = 0; m->bufferMask = 0; } void *vaddr = m->framebuffer->base; // find a free slot for (uint32_t i = 0 ; i < numBuffers ; i++) { if ((bufferMask & (1LU << i)) == 0) { m->bufferMask |= (1LU << i); break; } vaddr = (void *)((uintptr_t)vaddr + bufferSize); } // The entire framebuffer memory is already mapped, now create a buffer object for parts of this memory private_handle_t *hnd = new private_handle_t(private_handle_t::PRIV_FLAGS_FRAMEBUFFER, usage, size, vaddr, 0, m->framebuffer->fd, (uintptr_t)vaddr - (uintptr_t) m->framebuffer->base, m->framebuffer->fb_paddr); #if GRALLOC_ARM_UMP_MODULE hnd->ump_id = m->framebuffer->ump_id; /* create a backing ump memory handle if the framebuffer is exposed as a secure ID */ if ((int)UMP_INVALID_SECURE_ID != hnd->ump_id) { hnd->ump_mem_handle = (int)ump_handle_create_from_secure_id(hnd->ump_id); if ((int)UMP_INVALID_MEMORY_HANDLE == hnd->ump_mem_handle) { AINF("warning: unable to create UMP handle from secure ID %i\n", hnd->ump_id); } } #endif #if GRALLOC_ARM_DMA_BUF_MODULE { #ifdef FBIOGET_DMABUF /* * Perform allocator specific actions. If these fail we fall back to a regular buffer * which will be memcpy'ed to the main screen when fb_post is called. */ if (fb_get_framebuffer_dmabuf(m, hnd) == -1) { int newUsage = (usage & ~GRALLOC_USAGE_HW_FB) | GRALLOC_USAGE_HW_2D; AINF("Fallback to single buffering. Unable to map framebuffer memory to handle:%p", hnd); return gralloc_alloc_buffer(dev, bufferSize, newUsage, pHandle); } #endif } // correct numFds/numInts when there is no dmabuf fd if (hnd->share_fd < 0) { hnd->numFds--; hnd->numInts++; } #endif *pHandle = hnd; return 0; } static int gralloc_alloc_framebuffer(alloc_device_t *dev, size_t size, int usage, buffer_handle_t *pHandle) { private_module_t *m = reinterpret_cast(dev->common.module); pthread_mutex_lock(&m->lock); int err = gralloc_alloc_framebuffer_locked(dev, size, usage, pHandle); pthread_mutex_unlock(&m->lock); return err; } #endif /* DISABLE_FRAMEBUFFER_HAL */ static int alloc_device_alloc(alloc_device_t *dev, int w, int h, int format, int usage, buffer_handle_t *pHandle, int *pStride) { if (!pHandle || !pStride) { return -EINVAL; } size_t size; size_t stride; int bpp = 1; if (format == HAL_PIXEL_FORMAT_YCrCb_420_SP || format == HAL_PIXEL_FORMAT_YV12 /* HAL_PIXEL_FORMAT_YCbCr_420_SP, HAL_PIXEL_FORMAT_YCbCr_420_P, HAL_PIXEL_FORMAT_YCbCr_422_I are not defined in Android. * To enable Mali DDK EGLImage support for those formats, firstly, you have to add them in Android system/core/include/system/graphics.h. * Then, define SUPPORT_LEGACY_FORMAT in the same header file(Mali DDK will also check this definition). */ #ifdef SUPPORT_LEGACY_FORMAT || format == HAL_PIXEL_FORMAT_YCbCr_420_SP || format == HAL_PIXEL_FORMAT_YCbCr_420_P || format == HAL_PIXEL_FORMAT_YCbCr_422_I #endif ) { switch (format) { case HAL_PIXEL_FORMAT_YCrCb_420_SP: stride = GRALLOC_ALIGN(w, 16); size = GRALLOC_ALIGN(h, 16) * (stride + GRALLOC_ALIGN(stride / 2, 16)); break; case HAL_PIXEL_FORMAT_YV12: #ifdef SUPPORT_LEGACY_FORMAT case HAL_PIXEL_FORMAT_YCbCr_420_P: #endif /* * Since Utgard has limitation that "64-byte alignment is enforced on texture and mipmap addresses", here to make sure * the v, u plane start addresses are 64-byte aligned. */ stride = GRALLOC_ALIGN(w, (h % 8 == 0) ? GRALLOC_ALIGN_BASE_16 : ((h % 4 == 0) ? GRALLOC_ALIGN_BASE_64 : GRALLOC_ALIGN_BASE_128)); size = GRALLOC_ALIGN(h, 2) * (stride + GRALLOC_ALIGN(stride / 2, 16)); break; #ifdef SUPPORT_LEGACY_FORMAT case HAL_PIXEL_FORMAT_YCbCr_420_SP: stride = GRALLOC_ALIGN(w, 16); size = GRALLOC_ALIGN(h, 16) * (stride + GRALLOC_ALIGN(stride / 2, 16)); break; case HAL_PIXEL_FORMAT_YCbCr_422_I: stride = GRALLOC_ALIGN(w, 16); size = h * stride * 2; break; #endif default: return -EINVAL; } } else { switch (format) { case HAL_PIXEL_FORMAT_RGBA_8888: case HAL_PIXEL_FORMAT_RGBX_8888: case HAL_PIXEL_FORMAT_BGRA_8888: bpp = 4; break; case HAL_PIXEL_FORMAT_RGB_888: bpp = 3; break; case HAL_PIXEL_FORMAT_RGB_565: #if PLATFORM_SDK_VERSION < 19 case HAL_PIXEL_FORMAT_RGBA_5551: case HAL_PIXEL_FORMAT_RGBA_4444: #endif bpp = 2; break; case HAL_PIXEL_FORMAT_BLOB: if (h != 1) { AERR("Height for HAL_PIXEL_FORMAT_BLOB must be 1. h=%d", h); return -EINVAL; } break; default: AERR("The format is not supported yet: format=%d\n", format); return -EINVAL; } if (format == HAL_PIXEL_FORMAT_BLOB) { stride = 0; /* No 'rows', it's effectively a long one dimensional array */ size = w; }else{ size_t bpr = GRALLOC_ALIGN(w * bpp, 64); size = bpr * h; stride = bpr / bpp; } } int err; #ifndef DISABLE_FRAMEBUFFER_HAL if (usage & GRALLOC_USAGE_HW_FB) { err = gralloc_alloc_framebuffer(dev, size, usage, pHandle); } else #endif { err = gralloc_alloc_buffer(dev, size, usage, pHandle); } if (err < 0) { return err; } /* match the framebuffer format */ if (usage & GRALLOC_USAGE_HW_FB) { #ifdef GRALLOC_16_BITS format = HAL_PIXEL_FORMAT_RGB_565; #else format = HAL_PIXEL_FORMAT_BGRA_8888; #endif } private_handle_t *hnd = (private_handle_t *)*pHandle; int private_usage = usage & (GRALLOC_USAGE_PRIVATE_0 | GRALLOC_USAGE_PRIVATE_1); switch (private_usage) { case 0: hnd->yuv_info = MALI_YUV_BT601_NARROW; break; case GRALLOC_USAGE_PRIVATE_1: hnd->yuv_info = MALI_YUV_BT601_WIDE; break; case GRALLOC_USAGE_PRIVATE_0: hnd->yuv_info = MALI_YUV_BT709_NARROW; break; case (GRALLOC_USAGE_PRIVATE_0 | GRALLOC_USAGE_PRIVATE_1): hnd->yuv_info = MALI_YUV_BT709_WIDE; break; } hnd->width = w; hnd->height = h; hnd->format = format; hnd->stride = stride; hnd->byte_stride = GRALLOC_ALIGN(w*bpp,64); *pStride = stride; return 0; } static int alloc_device_free(alloc_device_t __unused *dev, buffer_handle_t handle) { if (private_handle_t::validate(handle) < 0) { return -EINVAL; } private_handle_t const *hnd = reinterpret_cast(handle); if (hnd->flags & private_handle_t::PRIV_FLAGS_FRAMEBUFFER) { #if GRALLOC_ARM_UMP_MODULE if ((int)UMP_INVALID_MEMORY_HANDLE != hnd->ump_mem_handle) { ump_reference_release((ump_handle)hnd->ump_mem_handle); } #endif } else if (hnd->flags & private_handle_t::PRIV_FLAGS_USES_UMP) { #if GRALLOC_ARM_UMP_MODULE /* Buffer might be unregistered so we need to check for invalid ump handle*/ if ((int)UMP_INVALID_MEMORY_HANDLE != hnd->ump_mem_handle) { ump_mapped_pointer_release((ump_handle)hnd->ump_mem_handle); ump_reference_release((ump_handle)hnd->ump_mem_handle); } #else AERR("Can't free ump memory for handle:%p. Not supported.", hnd); #endif } else if (hnd->flags & private_handle_t::PRIV_FLAGS_USES_ION) { #if GRALLOC_ARM_DMA_BUF_MODULE /* Buffer might be unregistered so we need to check for invalid ump handle*/ if (0 != hnd->base) { if (0 != munmap((void *)hnd->base, hnd->size)) { AERR("Failed to munmap handle %p", hnd); } } close(hnd->share_fd); memset((void *)hnd, 0, sizeof(*hnd)); #else AERR("Can't free dma_buf memory for handle:0x%x. Not supported.", (unsigned int)hnd); #endif } delete hnd; return 0; } static int alloc_device_close(struct hw_device_t *device) { alloc_device_t *dev = reinterpret_cast(device); if (dev) { #if GRALLOC_ARM_DMA_BUF_MODULE private_module_t *m = reinterpret_cast(device); if (0 != ion_close(m->ion_client)) { AERR("Failed to close ion_client: %d", m->ion_client); } close(m->ion_client); #endif delete dev; #if GRALLOC_ARM_UMP_MODULE ump_close(); // Our UMP memory refs will be released automatically here... #endif } return 0; } #if GRALLOC_ARM_DMA_BUF_MODULE static int find_heap_id(int ion_client, char* name) { int i, ret, cnt, heap_id = -1; struct ion_heap_data *data; ret = ion_query_heap_cnt(ion_client, &cnt); if (ret) { AERR("ion count query failed with %s", strerror(errno)); return -1; } data = (struct ion_heap_data *)malloc(cnt * sizeof(*data)); if (!data) { AERR("Error allocating data %s\n", strerror(errno)); return -1; } ret = ion_query_get_heaps(ion_client, cnt, data); if (ret) { AERR("Error querying heaps from ion %s", strerror(errno)); } else { for (i = 0; i < cnt; i++) { struct ion_heap_data *dat = (struct ion_heap_data *)data; if (strcmp(dat[i].name, name) == 0) { heap_id = dat[i].heap_id; break; } } if (i > cnt) { AERR("No System Heap Found amongst %d heaps\n", cnt); heap_id = -1; } } free(data); return heap_id; } #endif static int initialize_interface(private_module_t *m) { int fd; if (interface_ver != INTERFACE_UNKNOWN) return 0; /* test for dma-heaps*/ fd = dma_heap_open(DMABUF_SYSTEM); if (fd >= 0) { AINF("Using DMA-BUF Heaps.\n"); interface_ver = INTERFACE_DMABUF_HEAPS; system_heap_id = fd; cma_heap_id = dma_heap_open(DMABUF_CMA); /* Open other dma heaps here */ return 0; } /* test for modern vs legacy ION */ m->ion_client = ion_open(); if (m->ion_client < 0) { AERR("ion_open failed with %s", strerror(errno)); return -1; } if (!ion_is_legacy(m->ion_client)) { system_heap_id = find_heap_id(m->ion_client, ION_SYSTEM); cma_heap_id = find_heap_id(m->ion_client, ION_CMA); if (system_heap_id < 0) { ion_close(m->ion_client); m->ion_client = -1; AERR( "ion_open failed: no system heap found" ); return -1; } if (cma_heap_id < 0) { AERR("No cma heap found, falling back to system"); cma_heap_id = system_heap_id; } AINF("Using ION Modern interface.\n"); interface_ver = INTERFACE_ION_MODERN; } else { AINF("Using ION Legacy interface.\n"); interface_ver = INTERFACE_ION_LEGACY; } return 0; } int alloc_device_open(hw_module_t const *module, const char *name, hw_device_t **device) { MALI_IGNORE(name); alloc_device_t *dev; dev = new alloc_device_t; if (NULL == dev) { return -1; } #if GRALLOC_ARM_UMP_MODULE ump_result ump_res = ump_open(); if (UMP_OK != ump_res) { AERR("UMP open failed with %d", ump_res); delete dev; return -1; } #endif /* initialize our state here */ memset(dev, 0, sizeof(*dev)); /* initialize the procs */ dev->common.tag = HARDWARE_DEVICE_TAG; dev->common.version = 0; dev->common.module = const_cast(module); dev->common.close = alloc_device_close; dev->alloc = alloc_device_alloc; dev->free = alloc_device_free; #if GRALLOC_ARM_DMA_BUF_MODULE private_module_t *m = reinterpret_cast(dev->common.module); if (initialize_interface(m) < 0) { delete dev; return -1; } #endif *device = &dev->common; return 0; }