| /* |
| * Copyright (c) 2011-2021, The Linux Foundation. All rights reserved. |
| |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions are |
| * met: |
| * * Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * * Redistributions in binary form must reproduce the above |
| * copyright notice, this list of conditions and the following |
| * disclaimer in the documentation and/or other materials provided |
| * with the distribution. |
| * * Neither the name of The Linux Foundation nor the names of its |
| * contributors may be used to endorse or promote products derived |
| * from this software without specific prior written permission. |
| * |
| * THIS SOFTWARE IS PROVIDED "AS IS" AND ANY EXPRESS OR IMPLIED |
| * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF |
| * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT |
| * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS |
| * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
| * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
| * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR |
| * BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, |
| * WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE |
| * OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN |
| * IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #ifndef QMAA |
| #include <media/msm_media_info.h> |
| #endif |
| |
| #include <drm/drm_fourcc.h> |
| |
| #include <cutils/properties.h> |
| #include <algorithm> |
| |
| #include "gr_adreno_info.h" |
| #include "gr_camera_info.h" |
| #include "gr_utils.h" |
| |
| #define ASTC_BLOCK_SIZE 16 |
| |
| namespace gralloc { |
| |
| bool IsYuvFormat(int format) { |
| switch (format) { |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP: |
| case HAL_PIXEL_FORMAT_YCbCr_422_SP: |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: // Same as YCbCr_420_SP_VENUS |
| case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_422_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_NV21_ZSL: |
| case HAL_PIXEL_FORMAT_RAW16: |
| case HAL_PIXEL_FORMAT_Y16: |
| case HAL_PIXEL_FORMAT_RAW12: |
| case HAL_PIXEL_FORMAT_RAW10: |
| case HAL_PIXEL_FORMAT_YV12: |
| case HAL_PIXEL_FORMAT_Y8: |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010: |
| case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: |
| // Below formats used by camera and VR |
| case HAL_PIXEL_FORMAT_BLOB: |
| case HAL_PIXEL_FORMAT_RAW_OPAQUE: |
| case HAL_PIXEL_FORMAT_NV12_HEIF: |
| case HAL_PIXEL_FORMAT_CbYCrY_422_I: |
| case HAL_PIXEL_FORMAT_NV12_LINEAR_FLEX : |
| case HAL_PIXEL_FORMAT_NV12_UBWC_FLEX: |
| case HAL_PIXEL_FORMAT_NV12_UBWC_FLEX_2_BATCH: |
| case HAL_PIXEL_FORMAT_NV12_UBWC_FLEX_4_BATCH: |
| case HAL_PIXEL_FORMAT_NV12_UBWC_FLEX_8_BATCH: |
| case HAL_PIXEL_FORMAT_MULTIPLANAR_FLEX: |
| return true; |
| default: |
| return false; |
| } |
| } |
| |
| bool IsUncompressedRGBFormat(int format) { |
| switch (format) { |
| case HAL_PIXEL_FORMAT_RGBA_8888: |
| case HAL_PIXEL_FORMAT_RGBX_8888: |
| case HAL_PIXEL_FORMAT_RGB_888: |
| case HAL_PIXEL_FORMAT_RGB_565: |
| case HAL_PIXEL_FORMAT_BGR_565: |
| case HAL_PIXEL_FORMAT_BGRA_8888: |
| case HAL_PIXEL_FORMAT_RGBA_5551: |
| case HAL_PIXEL_FORMAT_RGBA_4444: |
| case HAL_PIXEL_FORMAT_R_8: |
| case HAL_PIXEL_FORMAT_RG_88: |
| case HAL_PIXEL_FORMAT_BGRX_8888: |
| case HAL_PIXEL_FORMAT_RGBA_1010102: |
| case HAL_PIXEL_FORMAT_ARGB_2101010: |
| case HAL_PIXEL_FORMAT_RGBX_1010102: |
| case HAL_PIXEL_FORMAT_XRGB_2101010: |
| case HAL_PIXEL_FORMAT_BGRA_1010102: |
| case HAL_PIXEL_FORMAT_ABGR_2101010: |
| case HAL_PIXEL_FORMAT_BGRX_1010102: |
| case HAL_PIXEL_FORMAT_XBGR_2101010: |
| case HAL_PIXEL_FORMAT_RGBA_FP16: |
| case HAL_PIXEL_FORMAT_BGR_888: |
| return true; |
| default: |
| break; |
| } |
| |
| return false; |
| } |
| |
| bool IsCompressedRGBFormat(int format) { |
| switch (format) { |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_4x4_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_4x4_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_5x4_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_5x4_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_5x5_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_5x5_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_6x5_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_6x5_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_6x6_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_6x6_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_8x5_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_8x5_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_8x6_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_8x6_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_8x8_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_8x8_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_10x5_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_10x5_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_10x6_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_10x6_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_10x8_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_10x8_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_10x10_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_10x10_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_12x10_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_12x10_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_12x12_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_12x12_KHR: |
| return true; |
| default: |
| break; |
| } |
| |
| return false; |
| } |
| |
| bool IsCameraCustomFormat(int format) { |
| switch (format) { |
| case HAL_PIXEL_FORMAT_NV21_ZSL: |
| case HAL_PIXEL_FORMAT_NV12_UBWC_FLEX: |
| case HAL_PIXEL_FORMAT_NV12_UBWC_FLEX_2_BATCH: |
| case HAL_PIXEL_FORMAT_NV12_UBWC_FLEX_4_BATCH: |
| case HAL_PIXEL_FORMAT_NV12_UBWC_FLEX_8_BATCH: |
| case HAL_PIXEL_FORMAT_MULTIPLANAR_FLEX: |
| case HAL_PIXEL_FORMAT_RAW_OPAQUE: |
| #ifndef NO_RAW10_CUSTOM_FORMAT |
| case HAL_PIXEL_FORMAT_RAW10: |
| case HAL_PIXEL_FORMAT_RAW12: |
| #endif |
| return true; |
| default: |
| break; |
| } |
| |
| return false; |
| } |
| |
| uint32_t GetBppForUncompressedRGB(int format) { |
| uint32_t bpp = 0; |
| switch (format) { |
| case HAL_PIXEL_FORMAT_RGBA_FP16: |
| bpp = 8; |
| break; |
| case HAL_PIXEL_FORMAT_RGBA_8888: |
| case HAL_PIXEL_FORMAT_RGBX_8888: |
| case HAL_PIXEL_FORMAT_BGRA_8888: |
| case HAL_PIXEL_FORMAT_BGRX_8888: |
| case HAL_PIXEL_FORMAT_RGBA_1010102: |
| case HAL_PIXEL_FORMAT_ARGB_2101010: |
| case HAL_PIXEL_FORMAT_RGBX_1010102: |
| case HAL_PIXEL_FORMAT_XRGB_2101010: |
| case HAL_PIXEL_FORMAT_BGRA_1010102: |
| case HAL_PIXEL_FORMAT_ABGR_2101010: |
| case HAL_PIXEL_FORMAT_BGRX_1010102: |
| case HAL_PIXEL_FORMAT_XBGR_2101010: |
| bpp = 4; |
| break; |
| case HAL_PIXEL_FORMAT_RGB_888: |
| case HAL_PIXEL_FORMAT_BGR_888: |
| bpp = 3; |
| break; |
| case HAL_PIXEL_FORMAT_RGB_565: |
| case HAL_PIXEL_FORMAT_BGR_565: |
| case HAL_PIXEL_FORMAT_RGBA_5551: |
| case HAL_PIXEL_FORMAT_RGBA_4444: |
| bpp = 2; |
| break; |
| default: |
| ALOGE("Error : %s New format request = 0x%x", __FUNCTION__, format); |
| break; |
| } |
| |
| return bpp; |
| } |
| |
| bool CpuCanAccess(uint64_t usage) { |
| return CpuCanRead(usage) || CpuCanWrite(usage); |
| } |
| |
| bool CpuCanRead(uint64_t usage) { |
| if (usage & BufferUsage::CPU_READ_MASK) { |
| return true; |
| } |
| |
| return false; |
| } |
| |
| bool CpuCanWrite(uint64_t usage) { |
| if (usage & BufferUsage::CPU_WRITE_MASK) { |
| // Application intends to use CPU for rendering |
| return true; |
| } |
| |
| return false; |
| } |
| |
| uint32_t GetDataAlignment(int format, uint64_t usage) { |
| uint32_t align = UINT(getpagesize()); |
| if (format == HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED) { |
| align = SIZE_8K; |
| } |
| |
| if (usage & BufferUsage::PROTECTED) { |
| if ((usage & BufferUsage::CAMERA_OUTPUT) || (usage & GRALLOC_USAGE_PRIVATE_SECURE_DISPLAY)) { |
| // The alignment here reflects qsee mmu V7L/V8L requirement |
| align = SZ_2M; |
| } else { |
| align = SECURE_ALIGN; |
| } |
| } |
| |
| return align; |
| } |
| |
| bool IsGPUFlagSupported(uint64_t usage) { |
| bool ret = true; |
| if ((usage & BufferUsage::GPU_MIPMAP_COMPLETE)) { |
| ALOGE("GPU_MIPMAP_COMPLETE not supported"); |
| ret = false; |
| } |
| |
| if ((usage & BufferUsage::GPU_CUBE_MAP)) { |
| ALOGE("GPU_CUBE_MAP not supported"); |
| ret = false; |
| } |
| |
| return ret; |
| } |
| |
| int GetBpp(int format) { |
| if (IsUncompressedRGBFormat(format)) { |
| return GetBppForUncompressedRGB(format); |
| } |
| switch (format) { |
| case HAL_PIXEL_FORMAT_COMPRESSED_RGBA_ASTC_4x4_KHR: |
| case HAL_PIXEL_FORMAT_COMPRESSED_SRGB8_ALPHA8_ASTC_4x4_KHR: |
| case HAL_PIXEL_FORMAT_RAW8: |
| case HAL_PIXEL_FORMAT_Y8: |
| return 1; |
| case HAL_PIXEL_FORMAT_RAW16: |
| case HAL_PIXEL_FORMAT_Y16: |
| case HAL_PIXEL_FORMAT_YCbCr_422_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_422_SP: |
| case HAL_PIXEL_FORMAT_YCbCr_422_I: |
| case HAL_PIXEL_FORMAT_YCrCb_422_I: |
| case HAL_PIXEL_FORMAT_CbYCrY_422_I: |
| return 2; |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010: |
| return 3; |
| default: |
| return -1; |
| } |
| } |
| |
| // Returns the final buffer size meant to be allocated with ion |
| unsigned int GetSize(const BufferInfo &info, unsigned int alignedw, unsigned int alignedh) { |
| unsigned int size = 0; |
| int format = info.format; |
| int width = info.width; |
| int height = info.height; |
| uint64_t usage = info.usage; |
| |
| if (!IsGPUFlagSupported(usage)) { |
| ALOGE("Unsupported GPU usage flags present 0x%" PRIx64, usage); |
| return 0; |
| } |
| |
| if (IsCameraCustomFormat(format) && CameraInfo::GetInstance()) { |
| int result = CameraInfo::GetInstance()->GetBufferSize(format, width, height, &size); |
| if (result != 0) { |
| ALOGE("%s: Failed to get the buffer size through camera library. Error code: %d", |
| __FUNCTION__, result); |
| return 0; |
| } |
| } else if (IsUBwcEnabled(format, usage)) { |
| size = GetUBwcSize(width, height, format, alignedw, alignedh); |
| } else if (IsUncompressedRGBFormat(format)) { |
| uint32_t bpp = GetBppForUncompressedRGB(format); |
| size = alignedw * alignedh * bpp; |
| } else if (IsCompressedRGBFormat(format)) { |
| size = alignedw * alignedh * ASTC_BLOCK_SIZE; |
| } else { |
| // Below switch should be for only YUV/custom formats |
| switch (format) { |
| case HAL_PIXEL_FORMAT_RAW16: |
| case HAL_PIXEL_FORMAT_Y16:size = alignedw * alignedh * 2; |
| break; |
| case HAL_PIXEL_FORMAT_RAW10: |
| case HAL_PIXEL_FORMAT_RAW12:size = ALIGN(alignedw * alignedh, SIZE_4K); |
| break; |
| case HAL_PIXEL_FORMAT_RAW8: |
| case HAL_PIXEL_FORMAT_Y8:size = alignedw * alignedh * 1; |
| break; |
| // adreno formats |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: // NV21 |
| size = ALIGN(alignedw * alignedh, SIZE_4K); |
| size += (unsigned int) ALIGN(2 * ALIGN(width / 2, 32) * ALIGN(height / 2, 32), SIZE_4K); |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: // NV12 |
| // The chroma plane is subsampled, |
| // but the pitch in bytes is unchanged |
| // The GPU needs 4K alignment, but the video decoder needs 8K |
| size = ALIGN(alignedw * alignedh, SIZE_8K); |
| size += ALIGN(alignedw * (unsigned int) ALIGN(height / 2, 32), SIZE_8K); |
| break; |
| case HAL_PIXEL_FORMAT_YV12: |
| if ((format == HAL_PIXEL_FORMAT_YV12) && ((width & 1) || (height & 1))) { |
| ALOGE("w or h is odd for the YV12 format"); |
| return 0; |
| } |
| size = alignedw * alignedh + (ALIGN(alignedw / 2, 16) * (alignedh / 2)) * 2; |
| size = ALIGN(size, (unsigned int) SIZE_4K); |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP: |
| size = ALIGN((alignedw * alignedh) + (alignedw * alignedh) / 2 + 1, SIZE_4K); |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010: |
| size = ALIGN((alignedw * alignedh * 2) + (alignedw * alignedh) + 1, SIZE_4K); |
| break; |
| #ifndef QMAA |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: |
| size = VENUS_BUFFER_SIZE(COLOR_FMT_P010, |
| width, |
| height); |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_422_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_422_SP: |
| case HAL_PIXEL_FORMAT_YCbCr_422_I: |
| case HAL_PIXEL_FORMAT_YCrCb_422_I: |
| case HAL_PIXEL_FORMAT_CbYCrY_422_I: |
| if (width & 1) { |
| ALOGE("width is odd for the YUV422_SP format"); |
| return 0; |
| } |
| size = ALIGN(alignedw * alignedh * 2, SIZE_4K); |
| break; |
| case HAL_PIXEL_FORMAT_NV12_LINEAR_FLEX: |
| size = VENUS_BUFFER_SIZE(COLOR_FMT_NV12_128, width, height); |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: |
| size = VENUS_BUFFER_SIZE(COLOR_FMT_NV12, width, height); |
| break; |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: |
| size = VENUS_BUFFER_SIZE(COLOR_FMT_NV21, width, height); |
| break; |
| case HAL_PIXEL_FORMAT_BLOB: |
| if (height != 1) { |
| ALOGE("%s: Buffers with HAL_PIXEL_FORMAT_BLOB must have height 1 ", __FUNCTION__); |
| return 0; |
| } |
| size = (unsigned int) width; |
| break; |
| case HAL_PIXEL_FORMAT_NV12_HEIF: |
| size = VENUS_BUFFER_SIZE(COLOR_FMT_NV12_512, width, height); |
| break; |
| #endif |
| default:ALOGE("%s: Unrecognized pixel format: 0x%x", __FUNCTION__, format); |
| return 0; |
| } |
| } |
| auto align = GetDataAlignment(format, usage); |
| size = ALIGN(size, align) * info.layer_count; |
| return size; |
| } |
| |
| int GetBufferSizeAndDimensions(const BufferInfo &info, unsigned int *size, unsigned int *alignedw, |
| unsigned int *alignedh) { |
| GraphicsMetadata graphics_metadata = {}; |
| return GetBufferSizeAndDimensions(info, size, alignedw, alignedh, &graphics_metadata); |
| } |
| |
| int GetBufferSizeAndDimensions(const BufferInfo &info, unsigned int *size, unsigned int *alignedw, |
| unsigned int *alignedh, GraphicsMetadata *graphics_metadata) { |
| int buffer_type = GetBufferType(info.format); |
| if (CanUseAdrenoForSize(buffer_type, info.usage)) { |
| return GetGpuResourceSizeAndDimensions(info, size, alignedw, alignedh, graphics_metadata); |
| } else { |
| GetAlignedWidthAndHeight(info, alignedw, alignedh); |
| *size = GetSize(info, *alignedw, *alignedh); |
| } |
| return 0; |
| } |
| |
| void GetYuvUbwcSPPlaneInfo(uint32_t width, uint32_t height, int color_format, |
| PlaneLayoutInfo *plane_info) { |
| // UBWC buffer has these 4 planes in the following sequence: |
| // Y_Plane, UV_Plane, Y_Meta_Plane, UV_Meta_Plane |
| unsigned int y_meta_stride = 0, y_meta_height = 0, y_meta_size = 0; |
| unsigned int y_stride = 0, y_height = 0, y_size = 0; |
| unsigned int c_meta_stride = 0, c_meta_height = 0, c_meta_size = 0; |
| unsigned int alignment = 4096; |
| unsigned int c_stride = 0, c_height = 0, c_size = 0; |
| uint64_t yOffset = 0, cOffset = 0, yMetaOffset = 0, cMetaOffset = 0; |
| |
| #ifndef QMAA |
| y_meta_stride = VENUS_Y_META_STRIDE(color_format, INT(width)); |
| y_meta_height = VENUS_Y_META_SCANLINES(color_format, INT(height)); |
| y_meta_size = ALIGN((y_meta_stride * y_meta_height), alignment); |
| |
| y_stride = VENUS_Y_STRIDE(color_format, INT(width)); |
| y_height = VENUS_Y_SCANLINES(color_format, INT(height)); |
| y_size = ALIGN((y_stride * y_height), alignment); |
| |
| c_meta_stride = VENUS_UV_META_STRIDE(color_format, INT(width)); |
| c_meta_height = VENUS_UV_META_SCANLINES(color_format, INT(height)); |
| c_meta_size = ALIGN((c_meta_stride * c_meta_height), alignment); |
| |
| c_stride = VENUS_UV_STRIDE(color_format, INT(width)); |
| c_height = VENUS_UV_SCANLINES(color_format, INT(height)); |
| c_size = ALIGN((c_stride * c_height), alignment); |
| #endif |
| yMetaOffset = 0; |
| yOffset = y_meta_size; |
| cMetaOffset = y_meta_size + y_size; |
| cOffset = y_meta_size + y_size + c_meta_size; |
| |
| plane_info[0].component = (PlaneComponent)PLANE_COMPONENT_Y; |
| plane_info[0].offset = (uint32_t)yOffset; |
| plane_info[0].stride = static_cast<int32_t>(UINT(width)); |
| plane_info[0].stride_bytes = static_cast<int32_t>(y_stride); |
| plane_info[0].scanlines = static_cast<int32_t>(y_height); |
| plane_info[0].size = static_cast<uint32_t>(y_size); |
| |
| plane_info[1].component = (PlaneComponent)(PLANE_COMPONENT_Cb | PLANE_COMPONENT_Cr); |
| plane_info[1].offset = (uint32_t)cOffset; |
| plane_info[1].stride = static_cast<int32_t>(UINT(width)); |
| plane_info[1].stride_bytes = static_cast<int32_t>(c_stride); |
| plane_info[1].scanlines = static_cast<int32_t>(c_height); |
| plane_info[1].size = static_cast<uint32_t>(c_size); |
| |
| plane_info[2].component = (PlaneComponent)(PLANE_COMPONENT_META | PLANE_COMPONENT_Y); |
| plane_info[2].offset = (uint32_t)yMetaOffset; |
| plane_info[2].stride = static_cast<int32_t>(UINT(width)); |
| plane_info[2].stride_bytes = static_cast<int32_t>(y_meta_stride); |
| plane_info[2].scanlines = static_cast<int32_t>(y_meta_height); |
| plane_info[2].size = static_cast<uint32_t>(y_meta_size); |
| |
| plane_info[3].component = |
| (PlaneComponent)(PLANE_COMPONENT_META | PLANE_COMPONENT_Cb | PLANE_COMPONENT_Cr); |
| plane_info[3].offset = (uint32_t)cMetaOffset; |
| plane_info[3].stride = static_cast<int32_t>(UINT(width)); |
| plane_info[3].stride_bytes = static_cast<int32_t>(c_meta_stride); |
| plane_info[3].scanlines = static_cast<int32_t>(c_meta_height); |
| plane_info[3].size = static_cast<uint32_t>(c_meta_size); |
| } |
| |
| // This API gets information about 8 planes (Y_Plane, UV_Plane, Y_Meta_Plane, UV_Meta_Plane, |
| // Y_Plane, UV_Plane, Y_Meta_Plane, UV_Meta_Plane) and it stores the |
| // information in PlaneLayoutInfo array. |
| void GetYuvUbwcInterlacedSPPlaneInfo(uint32_t width, uint32_t height, |
| PlaneLayoutInfo plane_info[8]) { |
| // UBWC interlaced has top-bottom field layout with each field as |
| // 8-plane (including meta plane also) NV12_UBWC with width = image_width |
| // & height = image_height / 2. |
| // Client passed plane_info argument is ptr to struct PlaneLayoutInfo[8]. |
| // Plane info to be filled for each field separately. |
| height = (height + 1) >> 1; |
| |
| #ifndef QMAA |
| GetYuvUbwcSPPlaneInfo(width, height, COLOR_FMT_NV12_UBWC, &plane_info[0]); |
| |
| GetYuvUbwcSPPlaneInfo(width, height, COLOR_FMT_NV12_UBWC, &plane_info[4]); |
| #endif |
| } |
| |
| // This API gets information about 2 planes (Y_Plane & UV_Plane). |
| // Here width and height are aligned width and aligned height. |
| // bpp: bits per pixel. |
| void GetYuvSPPlaneInfo(const BufferInfo &info, int format, uint32_t width, uint32_t height, |
| uint32_t bpp, PlaneLayoutInfo *plane_info) { |
| int unaligned_width = info.width; |
| int unaligned_height = info.height; |
| unsigned int y_stride = 0, y_height = 0, y_size = 0; |
| unsigned int c_stride = 0, c_height = 0, c_size = 0; |
| uint64_t yOffset, cOffset; |
| |
| y_stride = c_stride = UINT(width) * bpp; |
| y_height = INT(height); |
| y_size = y_stride * y_height; |
| switch (format) { |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP: |
| c_size = (width * height) / 2 + 1; |
| c_height = height >> 1; |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_422_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_422_SP: |
| if (unaligned_width & 1) { |
| ALOGE("width is odd for the YUV422_SP format"); |
| return; |
| } |
| c_size = width * height; |
| c_height = height; |
| break; |
| #ifndef QMAA |
| case HAL_PIXEL_FORMAT_NV12_LINEAR_FLEX: |
| c_height = VENUS_UV_SCANLINES(COLOR_FMT_NV12_128, height); |
| c_size = c_stride * c_height; |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: |
| c_height = VENUS_UV_SCANLINES(COLOR_FMT_NV12, height); |
| c_size = c_stride * c_height; |
| break; |
| case HAL_PIXEL_FORMAT_NV12_HEIF: |
| c_height = VENUS_UV_SCANLINES(COLOR_FMT_NV12_512, height); |
| c_size = c_stride * c_height; |
| break; |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: |
| y_size = ALIGN(width * height, 4096); |
| c_size = ALIGN(2 * ALIGN(unaligned_width / 2, 32) * ALIGN(unaligned_height / 2, 32), 4096); |
| break; |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: |
| c_height = VENUS_UV_SCANLINES(COLOR_FMT_NV21, height); |
| c_size = c_stride * c_height; |
| break; |
| #endif |
| case HAL_PIXEL_FORMAT_Y16: |
| c_size = c_stride = 0; |
| c_height = 0; |
| break; |
| case HAL_PIXEL_FORMAT_Y8: |
| c_size = c_stride = 0; |
| c_height = 0; |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010: |
| c_size = (width * height) + 1; |
| c_height = height; |
| break; |
| default: |
| break; |
| } |
| |
| yOffset = 0; |
| cOffset = y_size; |
| |
| plane_info[0].component = (PlaneComponent)PLANE_COMPONENT_Y; |
| plane_info[0].offset = (uint32_t)yOffset; |
| plane_info[0].step = 1; |
| plane_info[0].stride = static_cast<int32_t>(UINT(width)); |
| plane_info[0].stride_bytes = static_cast<int32_t>(y_stride); |
| plane_info[0].scanlines = static_cast<int32_t>(y_height); |
| plane_info[0].size = static_cast<uint32_t>(y_size); |
| |
| plane_info[1].component = (PlaneComponent)(PLANE_COMPONENT_Cb | PLANE_COMPONENT_Cr); |
| plane_info[1].offset = (uint32_t)cOffset; |
| plane_info[1].step = 2 * bpp; |
| plane_info[1].stride = static_cast<int32_t>(UINT(width)); |
| plane_info[1].stride_bytes = static_cast<int32_t>(c_stride); |
| plane_info[1].scanlines = static_cast<int32_t>(c_height); |
| plane_info[1].size = static_cast<uint32_t>(c_size); |
| } |
| |
| int GetYUVPlaneInfo(const private_handle_t *hnd, struct android_ycbcr ycbcr[2]) { |
| int err = 0; |
| uint32_t width = UINT(hnd->width); |
| uint32_t height = UINT(hnd->height); |
| int format = hnd->format; |
| uint64_t usage = hnd->usage; |
| int32_t interlaced = 0; |
| int plane_count = 0; |
| int unaligned_width = INT(hnd->unaligned_width); |
| int unaligned_height = INT(hnd->unaligned_height); |
| BufferInfo info(unaligned_width, unaligned_height, format, usage); |
| |
| memset(ycbcr->reserved, 0, sizeof(ycbcr->reserved)); |
| |
| // Check if UBWC buffer has been rendered in linear format. |
| int linear_format = 0; |
| if (getMetaData(const_cast<private_handle_t *>(hnd), GET_LINEAR_FORMAT, &linear_format) == 0) { |
| format = INT(linear_format); |
| } |
| |
| // Check metadata if the geometry has been updated. |
| BufferDim_t buffer_dim; |
| if (getMetaData(const_cast<private_handle_t *>(hnd), GET_BUFFER_GEOMETRY, &buffer_dim) == 0) { |
| BufferInfo info(buffer_dim.sliceWidth, buffer_dim.sliceHeight, format, usage); |
| GetAlignedWidthAndHeight(info, &width, &height); |
| } |
| |
| // Check metadata for interlaced content. |
| int interlace_flag = 0; |
| if (getMetaData(const_cast<private_handle_t *>(hnd), GET_PP_PARAM_INTERLACED, &interlace_flag) == |
| 0) { |
| if (interlace_flag) { |
| interlaced = LAYOUT_INTERLACED_FLAG; |
| } |
| } |
| |
| PlaneLayoutInfo plane_info[8] = {}; |
| // Get the chroma offsets from the handle width/height. We take advantage |
| // of the fact the width _is_ the stride |
| err = GetYUVPlaneInfo(info, format, width, height, interlaced, &plane_count, plane_info); |
| if (err == 0) { |
| if (interlaced && format == HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC) { |
| CopyPlaneLayoutInfotoAndroidYcbcr(hnd->base, plane_count, &plane_info[0], &ycbcr[0]); |
| unsigned int uv_stride = 0, uv_height = 0, uv_size = 0; |
| unsigned int alignment = 4096; |
| uint64_t field_base; |
| height = (height + 1) >> 1; |
| #ifndef QMAA |
| uv_stride = VENUS_UV_STRIDE(COLOR_FMT_NV12_UBWC, INT(width)); |
| uv_height = VENUS_UV_SCANLINES(COLOR_FMT_NV12_UBWC, INT(height)); |
| #endif |
| uv_size = ALIGN((uv_stride * uv_height), alignment); |
| field_base = hnd->base + plane_info[1].offset + uv_size; |
| memset(ycbcr[1].reserved, 0, sizeof(ycbcr[1].reserved)); |
| CopyPlaneLayoutInfotoAndroidYcbcr(field_base, plane_count, &plane_info[4], &ycbcr[1]); |
| } else { |
| CopyPlaneLayoutInfotoAndroidYcbcr(hnd->base, plane_count, plane_info, ycbcr); |
| switch (format) { |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_422_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_NV21_ZSL: |
| std::swap(ycbcr->cb, ycbcr->cr); |
| } |
| } |
| } |
| return err; |
| } |
| |
| int GetRawPlaneInfo(int32_t format, int32_t width, int32_t height, PlaneLayoutInfo *plane_info) { |
| int32_t step = 0; |
| |
| switch (format) { |
| case HAL_PIXEL_FORMAT_RAW16: |
| step = 2; |
| break; |
| case HAL_PIXEL_FORMAT_RAW8: |
| step = 1; |
| break; |
| case HAL_PIXEL_FORMAT_RAW12: |
| case HAL_PIXEL_FORMAT_RAW10: |
| step = 0; |
| break; |
| default: |
| ALOGW("RawPlaneInfo is unsupported for format 0x%x", format); |
| return -EINVAL; |
| } |
| |
| BufferInfo info(width, height, format); |
| uint32_t alignedWidth, alignedHeight; |
| GetAlignedWidthAndHeight(info, &alignedWidth, &alignedHeight); |
| |
| uint32_t size = GetSize(info, alignedWidth, alignedHeight); |
| |
| plane_info[0].component = (PlaneComponent)PLANE_COMPONENT_RAW; |
| plane_info[0].h_subsampling = 0; |
| plane_info[0].v_subsampling = 0; |
| plane_info[0].offset = 0; |
| plane_info[0].step = step; |
| plane_info[0].stride = width; |
| plane_info[0].stride_bytes = static_cast<int32_t>(alignedWidth); |
| if (format == HAL_PIXEL_FORMAT_RAW16) { |
| plane_info[0].stride_bytes = static_cast<int32_t>(alignedWidth * GetBpp(format)); |
| } |
| plane_info[0].scanlines = height; |
| plane_info[0].size = size; |
| |
| return 0; |
| } |
| |
| // Explicitly defined UBWC formats |
| bool IsUBwcFormat(int format) { |
| switch (format) { |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: |
| case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: |
| return true; |
| default: |
| return false; |
| } |
| } |
| |
| bool IsUBwcSupported(int format) { |
| // Existing HAL formats with UBWC support |
| switch (format) { |
| case HAL_PIXEL_FORMAT_BGR_565: |
| case HAL_PIXEL_FORMAT_RGBA_8888: |
| case HAL_PIXEL_FORMAT_RGBX_8888: |
| case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_RGBA_1010102: |
| case HAL_PIXEL_FORMAT_RGBX_1010102: |
| case HAL_PIXEL_FORMAT_DEPTH_16: |
| case HAL_PIXEL_FORMAT_DEPTH_24: |
| case HAL_PIXEL_FORMAT_DEPTH_24_STENCIL_8: |
| case HAL_PIXEL_FORMAT_DEPTH_32F: |
| case HAL_PIXEL_FORMAT_STENCIL_8: |
| case HAL_PIXEL_FORMAT_RGBA_FP16: |
| return true; |
| default: |
| break; |
| } |
| |
| return false; |
| } |
| |
| bool IsUBwcPISupported(int format, uint64_t usage) { |
| // TODO(user): try and differentiate b/w mdp capability to support PI. |
| if (!(usage & GRALLOC_USAGE_PRIVATE_ALLOC_UBWC_PI)) { |
| return false; |
| } |
| |
| // As of now only two formats |
| switch (format) { |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: |
| case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: { |
| if ((usage & BufferUsage::GPU_TEXTURE) || (usage & BufferUsage::GPU_RENDER_TARGET)) { |
| if (AdrenoMemInfo::GetInstance()) { |
| return AdrenoMemInfo::GetInstance()->IsPISupportedByGPU(format, usage); |
| } |
| } else { |
| return true; |
| } |
| } |
| } |
| |
| return false; |
| } |
| |
| bool IsUBwcEnabled(int format, uint64_t usage) { |
| // Allow UBWC, if client is using an explicitly defined UBWC pixel format. |
| if (IsUBwcFormat(format)) { |
| return true; |
| } |
| |
| // Allow UBWC, if an OpenGL client sets UBWC usage flag and GPU plus MDP |
| // support the format. OR if a non-OpenGL client like Rotator, sets UBWC |
| // usage flag and MDP supports the format. |
| if (((usage & GRALLOC_USAGE_PRIVATE_ALLOC_UBWC) || |
| (usage & GRALLOC_USAGE_PRIVATE_ALLOC_UBWC_PI) || |
| (usage & BufferUsage::COMPOSER_CLIENT_TARGET)) |
| && IsUBwcSupported(format)) { |
| bool enable = true; |
| // Query GPU for UBWC only if buffer is intended to be used by GPU. |
| if ((usage & BufferUsage::GPU_TEXTURE) || (usage & BufferUsage::GPU_RENDER_TARGET)) { |
| if (AdrenoMemInfo::GetInstance()) { |
| enable = AdrenoMemInfo::GetInstance()->IsUBWCSupportedByGPU(format); |
| } |
| } |
| |
| // Allow UBWC, only if CPU usage flags are not set |
| if (enable && !(CpuCanAccess(usage))) { |
| return true; |
| } |
| } |
| |
| return false; |
| } |
| |
| void GetYuvUBwcWidthAndHeight(int width, int height, int format, unsigned int *aligned_w, |
| unsigned int *aligned_h) { |
| switch (format) { |
| #ifndef QMAA |
| case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: |
| *aligned_w = VENUS_Y_STRIDE(COLOR_FMT_NV12, width); |
| *aligned_h = VENUS_Y_SCANLINES(COLOR_FMT_NV12, height); |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: |
| *aligned_w = VENUS_Y_STRIDE(COLOR_FMT_NV12_UBWC, width); |
| *aligned_h = VENUS_Y_SCANLINES(COLOR_FMT_NV12_UBWC, height); |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: |
| // The macro returns the stride which is 4/3 times the width, hence * 3/4 |
| *aligned_w = (VENUS_Y_STRIDE(COLOR_FMT_NV12_BPP10_UBWC, width) * 3) / 4; |
| *aligned_h = VENUS_Y_SCANLINES(COLOR_FMT_NV12_BPP10_UBWC, height); |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: |
| // The macro returns the stride which is 2 times the width, hence / 2 |
| *aligned_w = (VENUS_Y_STRIDE(COLOR_FMT_P010_UBWC, width) / 2); |
| *aligned_h = VENUS_Y_SCANLINES(COLOR_FMT_P010_UBWC, height); |
| break; |
| #endif |
| default: |
| ALOGE("%s: Unsupported pixel format: 0x%x", __FUNCTION__, format); |
| *aligned_w = 0; |
| *aligned_h = 0; |
| break; |
| } |
| } |
| |
| void GetRgbUBwcBlockSize(uint32_t bpp, int *block_width, int *block_height) { |
| *block_width = 0; |
| *block_height = 0; |
| |
| switch (bpp) { |
| case 2: |
| case 4: |
| *block_width = 16; |
| *block_height = 4; |
| break; |
| case 8: |
| *block_width = 8; |
| *block_height = 4; |
| break; |
| case 16: |
| *block_width = 4; |
| *block_height = 4; |
| break; |
| default: |
| ALOGE("%s: Unsupported bpp: %d", __FUNCTION__, bpp); |
| break; |
| } |
| } |
| |
| unsigned int GetRgbUBwcMetaBufferSize(int width, int height, uint32_t bpp) { |
| unsigned int size = 0; |
| int meta_width, meta_height; |
| int block_width, block_height; |
| |
| GetRgbUBwcBlockSize(bpp, &block_width, &block_height); |
| if (!block_width || !block_height) { |
| ALOGE("%s: Unsupported bpp: %d", __FUNCTION__, bpp); |
| return size; |
| } |
| |
| // Align meta buffer height to 16 blocks |
| meta_height = ALIGN(((height + block_height - 1) / block_height), 16); |
| |
| // Align meta buffer width to 64 blocks |
| meta_width = ALIGN(((width + block_width - 1) / block_width), 64); |
| |
| // Align meta buffer size to 4K |
| size = (unsigned int)ALIGN((meta_width * meta_height), 4096); |
| |
| return size; |
| } |
| |
| unsigned int GetUBwcSize(int width, int height, int format, unsigned int alignedw, |
| unsigned int alignedh) { |
| unsigned int size = 0; |
| uint32_t bpp = 0; |
| switch (format) { |
| case HAL_PIXEL_FORMAT_BGR_565: |
| case HAL_PIXEL_FORMAT_RGBA_8888: |
| case HAL_PIXEL_FORMAT_RGBX_8888: |
| case HAL_PIXEL_FORMAT_RGBA_1010102: |
| case HAL_PIXEL_FORMAT_RGBX_1010102: |
| case HAL_PIXEL_FORMAT_RGBA_FP16: |
| bpp = GetBppForUncompressedRGB(format); |
| size = alignedw * alignedh * bpp; |
| size += GetRgbUBwcMetaBufferSize(width, height, bpp); |
| break; |
| #ifndef QMAA |
| /* |
| * 1. The CtsMediaV2TestCases#CodecEncoderSurfaceTest is a transcode use case and shares |
| * same surface between encoder and decoder. |
| * 2. Configures encoder with Opaque color format thus encoder sets ubwc usage bits and |
| * is configured with NV12_UBWC format. |
| * 3. Configures decoder as 'flexible', thus configuring decoder with NV12 format. |
| * 4. Decoder should produce output to surface that will be fed back to encoder as input. |
| * 5. Though UBWC is enabled, we need to compute the actual buffer size (including aligned |
| * width and height) based on pixel format that is set. |
| */ |
| case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: |
| size = VENUS_BUFFER_SIZE(COLOR_FMT_NV12, width, height); |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: |
| size = VENUS_BUFFER_SIZE(COLOR_FMT_NV12_UBWC, width, height); |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: |
| size = VENUS_BUFFER_SIZE(COLOR_FMT_NV12_BPP10_UBWC, width, height); |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: |
| size = VENUS_BUFFER_SIZE(COLOR_FMT_P010_UBWC, width, height); |
| break; |
| #endif |
| default: |
| ALOGE("%s: Unsupported pixel format: 0x%x", __FUNCTION__, format); |
| break; |
| } |
| |
| return size; |
| } |
| |
| unsigned int GetRgbMetaSize(int format, uint32_t width, uint32_t height, uint64_t usage) { |
| unsigned int meta_size = 0; |
| if (!IsUBwcEnabled(format, usage)) { |
| return meta_size; |
| } |
| uint32_t bpp = GetBppForUncompressedRGB(format); |
| switch (format) { |
| case HAL_PIXEL_FORMAT_BGR_565: |
| case HAL_PIXEL_FORMAT_RGBA_8888: |
| case HAL_PIXEL_FORMAT_RGBX_8888: |
| case HAL_PIXEL_FORMAT_RGBA_1010102: |
| case HAL_PIXEL_FORMAT_RGBX_1010102: |
| case HAL_PIXEL_FORMAT_RGBA_FP16: |
| meta_size = GetRgbUBwcMetaBufferSize(width, height, bpp); |
| break; |
| default: |
| ALOGE("%s:Unsupported RGB format: 0x%x", __FUNCTION__, format); |
| break; |
| } |
| return meta_size; |
| } |
| |
| int GetRgbDataAddress(private_handle_t *hnd, void **rgb_data) { |
| int err = 0; |
| |
| // This api is for RGB* formats |
| if (!IsUncompressedRGBFormat(hnd->format)) { |
| return -EINVAL; |
| } |
| |
| // linear buffer, nothing to do further |
| if (!(hnd->flags & private_handle_t::PRIV_FLAGS_UBWC_ALIGNED)) { |
| *rgb_data = reinterpret_cast<void *>(hnd->base); |
| return err; |
| } |
| unsigned int meta_size = GetRgbMetaSize(hnd->format, hnd->width, hnd->height, hnd->usage); |
| |
| *rgb_data = reinterpret_cast<void *>(hnd->base + meta_size); |
| |
| return err; |
| } |
| |
| void GetCustomDimensions(private_handle_t *hnd, int *stride, int *height) { |
| BufferDim_t buffer_dim; |
| int interlaced = 0; |
| |
| *stride = hnd->width; |
| *height = hnd->height; |
| if (getMetaData(hnd, GET_BUFFER_GEOMETRY, &buffer_dim) == 0) { |
| *stride = buffer_dim.sliceWidth; |
| *height = buffer_dim.sliceHeight; |
| } else if (getMetaData(hnd, GET_PP_PARAM_INTERLACED, &interlaced) == 0) { |
| if (interlaced && IsUBwcFormat(hnd->format)) { |
| unsigned int alignedw = 0, alignedh = 0; |
| // Get re-aligned height for single ubwc interlaced field and |
| // multiply by 2 to get frame height. |
| BufferInfo info(hnd->width, ((hnd->height + 1) >> 1), hnd->format); |
| GetAlignedWidthAndHeight(info, &alignedw, &alignedh); |
| *stride = static_cast<int>(alignedw); |
| *height = static_cast<int>(alignedh * 2); |
| } |
| } |
| } |
| |
| void GetColorSpaceFromMetadata(private_handle_t *hnd, int *color_space) { |
| ColorMetaData color_metadata; |
| if (getMetaData(hnd, GET_COLOR_METADATA, &color_metadata) == 0) { |
| switch (color_metadata.colorPrimaries) { |
| case ColorPrimaries_BT709_5: |
| *color_space = HAL_CSC_ITU_R_709; |
| break; |
| case ColorPrimaries_BT601_6_525: |
| case ColorPrimaries_BT601_6_625: |
| *color_space = ((color_metadata.range) ? HAL_CSC_ITU_R_601_FR : HAL_CSC_ITU_R_601); |
| break; |
| case ColorPrimaries_BT2020: |
| *color_space = (color_metadata.range) ? HAL_CSC_ITU_R_2020_FR : HAL_CSC_ITU_R_2020; |
| break; |
| default: |
| ALOGW("Unknown Color primary = %d", color_metadata.colorPrimaries); |
| break; |
| } |
| } else if (getMetaData(hnd, GET_COLOR_SPACE, color_space) != 0) { |
| *color_space = 0; |
| } |
| } |
| |
| void GetAlignedWidthAndHeight(const BufferInfo &info, unsigned int *alignedw, |
| unsigned int *alignedh) { |
| int width = info.width; |
| int height = info.height; |
| int format = info.format; |
| uint64_t usage = info.usage; |
| |
| // Currently surface padding is only computed for RGB* surfaces. |
| bool ubwc_enabled = IsUBwcEnabled(format, usage); |
| int tile = ubwc_enabled; |
| |
| // Use of aligned width and aligned height is to calculate the size of buffer, |
| // but in case of camera custom format size is being calculated from given width |
| // and given height. |
| if (IsCameraCustomFormat(format) && CameraInfo::GetInstance()) { |
| int aligned_w = width; |
| int aligned_h = height; |
| int result = CameraInfo::GetInstance()->GetStrideInBytes( |
| format, (PlaneComponent)PLANE_COMPONENT_Y, width, &aligned_w); |
| if (result != 0) { |
| ALOGE( |
| "%s: Failed to get the aligned width for camera custom format. width: %d, height: %d," |
| "format: %d, Error code: %d", |
| __FUNCTION__, width, height, format, result); |
| *alignedw = width; |
| *alignedh = aligned_h; |
| return; |
| } |
| |
| result = CameraInfo::GetInstance()->GetScanline(format, (PlaneComponent)PLANE_COMPONENT_Y, |
| height, &aligned_h); |
| if (result != 0) { |
| ALOGE( |
| "%s: Failed to get the aligned height for camera custom format. width: %d," |
| "height: %d, format: %d, Error code: %d", |
| __FUNCTION__, width, height, format, result); |
| *alignedw = aligned_w; |
| *alignedh = height; |
| return; |
| } |
| |
| *alignedw = aligned_w; |
| *alignedh = aligned_h; |
| return; |
| } |
| |
| if (IsUncompressedRGBFormat(format)) { |
| if (AdrenoMemInfo::GetInstance()) { |
| AdrenoMemInfo::GetInstance()->AlignUnCompressedRGB(width, height, format, tile, alignedw, |
| alignedh); |
| } |
| return; |
| } |
| |
| if (ubwc_enabled) { |
| GetYuvUBwcWidthAndHeight(width, height, format, alignedw, alignedh); |
| return; |
| } |
| |
| if (IsCompressedRGBFormat(format)) { |
| if (AdrenoMemInfo::GetInstance()) { |
| AdrenoMemInfo::GetInstance()->AlignCompressedRGB(width, height, format, alignedw, alignedh); |
| } |
| return; |
| } |
| |
| int aligned_w = width; |
| int aligned_h = height; |
| unsigned int alignment = 32; |
| |
| // Below should be only YUV family |
| switch (format) { |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP: |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP: |
| if (AdrenoMemInfo::GetInstance() == nullptr) { |
| return; |
| } |
| alignment = AdrenoMemInfo::GetInstance()->GetGpuPixelAlignment(); |
| aligned_w = ALIGN(width, alignment); |
| break; |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: |
| aligned_w = ALIGN(width, alignment); |
| break; |
| case HAL_PIXEL_FORMAT_RAW16: |
| case HAL_PIXEL_FORMAT_Y16: |
| case HAL_PIXEL_FORMAT_Y8: |
| aligned_w = ALIGN(width, 16); |
| break; |
| case HAL_PIXEL_FORMAT_RAW12: |
| aligned_w = ALIGN(width * 12 / 8, 16); |
| break; |
| case HAL_PIXEL_FORMAT_RAW10: |
| aligned_w = ALIGN(width * 10 / 8, 16); |
| break; |
| case HAL_PIXEL_FORMAT_RAW8: |
| aligned_w = ALIGN(width, 16); |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: |
| aligned_w = ALIGN(width, 128); |
| break; |
| case HAL_PIXEL_FORMAT_YV12: |
| if ((usage & BufferUsage::GPU_TEXTURE) || (usage & BufferUsage::GPU_RENDER_TARGET)) { |
| if (AdrenoMemInfo::GetInstance() == nullptr) { |
| return; |
| } |
| alignment = AdrenoMemInfo::GetInstance()->GetGpuPixelAlignment(); |
| aligned_w = ALIGN(width, alignment); |
| } else { |
| aligned_w = ALIGN(width, 16); |
| } |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_422_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_422_SP: |
| case HAL_PIXEL_FORMAT_YCbCr_422_I: |
| case HAL_PIXEL_FORMAT_YCrCb_422_I: |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010: |
| aligned_w = ALIGN(width, 16); |
| break; |
| #ifndef QMAA |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: |
| aligned_w = INT(VENUS_Y_STRIDE(COLOR_FMT_P010, width) / 2); |
| aligned_h = INT(VENUS_Y_SCANLINES(COLOR_FMT_P010, height)); |
| break; |
| case HAL_PIXEL_FORMAT_NV12_LINEAR_FLEX: |
| aligned_w = INT(VENUS_Y_STRIDE(COLOR_FMT_NV12_128, width)); |
| aligned_h = INT(VENUS_Y_SCANLINES(COLOR_FMT_NV12_128, height)); |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: |
| aligned_w = INT(VENUS_Y_STRIDE(COLOR_FMT_NV12, width)); |
| aligned_h = INT(VENUS_Y_SCANLINES(COLOR_FMT_NV12, height)); |
| break; |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: |
| aligned_w = INT(VENUS_Y_STRIDE(COLOR_FMT_NV21, width)); |
| aligned_h = INT(VENUS_Y_SCANLINES(COLOR_FMT_NV21, height)); |
| break; |
| case HAL_PIXEL_FORMAT_BLOB: |
| break; |
| case HAL_PIXEL_FORMAT_NV12_HEIF: |
| aligned_w = INT(VENUS_Y_STRIDE(COLOR_FMT_NV12_512, width)); |
| aligned_h = INT(VENUS_Y_SCANLINES(COLOR_FMT_NV12_512, height)); |
| break; |
| #endif |
| default: |
| break; |
| } |
| |
| *alignedw = (unsigned int)aligned_w; |
| *alignedh = (unsigned int)aligned_h; |
| } |
| |
| int GetBufferLayout(private_handle_t *hnd, uint32_t stride[4], uint32_t offset[4], |
| uint32_t *num_planes) { |
| if (!hnd || !stride || !offset || !num_planes) { |
| return -EINVAL; |
| } |
| |
| struct android_ycbcr yuvPlaneInfo[2] = {}; |
| *num_planes = 1; |
| |
| if (IsUncompressedRGBFormat(hnd->format)) { |
| uint32_t bpp = GetBppForUncompressedRGB(hnd->format); |
| stride[0] = static_cast<uint32_t>(hnd->width * bpp); |
| return 0; |
| } |
| |
| (*num_planes)++; |
| int ret = GetYUVPlaneInfo(hnd, yuvPlaneInfo); |
| if (ret < 0) { |
| ALOGE("%s failed", __FUNCTION__); |
| return ret; |
| } |
| |
| // We are only returning buffer layout for progressive or single field formats. |
| struct android_ycbcr yuvInfo = yuvPlaneInfo[0]; |
| stride[0] = static_cast<uint32_t>(yuvInfo.ystride); |
| offset[0] = static_cast<uint32_t>(reinterpret_cast<uint64_t>(yuvInfo.y) - hnd->base); |
| stride[1] = static_cast<uint32_t>(yuvInfo.cstride); |
| switch (hnd->format) { |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP: |
| case HAL_PIXEL_FORMAT_YCbCr_422_SP: |
| case HAL_PIXEL_FORMAT_NV12_LINEAR_FLEX: |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010: |
| case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: |
| case HAL_PIXEL_FORMAT_NV12_HEIF: |
| offset[1] = static_cast<uint32_t>(reinterpret_cast<uint64_t>(yuvInfo.cb) - hnd->base); |
| break; |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_YCrCb_422_SP: |
| case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: |
| offset[1] = static_cast<uint32_t>(reinterpret_cast<uint64_t>(yuvInfo.cr) - hnd->base); |
| break; |
| case HAL_PIXEL_FORMAT_YV12: |
| offset[1] = static_cast<uint32_t>(reinterpret_cast<uint64_t>(yuvInfo.cr) - hnd->base); |
| stride[2] = static_cast<uint32_t>(yuvInfo.cstride); |
| offset[2] = static_cast<uint32_t>(reinterpret_cast<uint64_t>(yuvInfo.cb) - hnd->base); |
| (*num_planes)++; |
| break; |
| case HAL_PIXEL_FORMAT_CbYCrY_422_I: |
| *num_planes = 1; |
| break; |
| default: |
| ALOGW("%s: Unsupported format", __FUNCTION__); |
| ret = -EINVAL; |
| } |
| |
| if (hnd->flags & private_handle_t::PRIV_FLAGS_UBWC_ALIGNED) { |
| std::fill(offset, offset + 4, 0); |
| } |
| |
| return 0; |
| } |
| |
| int GetGpuResourceSizeAndDimensions(const BufferInfo &info, unsigned int *size, |
| unsigned int *alignedw, unsigned int *alignedh, |
| GraphicsMetadata *graphics_metadata) { |
| GetAlignedWidthAndHeight(info, alignedw, alignedh); |
| AdrenoMemInfo* adreno_mem_info = AdrenoMemInfo::GetInstance(); |
| graphics_metadata->size = adreno_mem_info->AdrenoGetMetadataBlobSize(); |
| uint64_t adreno_usage = info.usage; |
| // If gralloc disables UBWC based on any of the checks, |
| // we pass modified usage flag to adreno to convey this. |
| int is_ubwc_enabled = IsUBwcEnabled(info.format, info.usage); |
| if (!is_ubwc_enabled) { |
| adreno_usage &= ~(GRALLOC_USAGE_PRIVATE_ALLOC_UBWC); |
| } else { |
| adreno_usage |= GRALLOC_USAGE_PRIVATE_ALLOC_UBWC; |
| } |
| |
| // Call adreno api for populating metadata blob |
| // Layer count is for 2D/Cubemap arrays and depth is used for 3D slice |
| // Using depth to pass layer_count here |
| int ret = adreno_mem_info->AdrenoInitMemoryLayout(graphics_metadata->data, info.width, |
| info.height, info.layer_count, /* depth */ |
| info.format, 1, is_ubwc_enabled, |
| adreno_usage, 1); |
| if (ret != 0) { |
| ALOGE("%s Graphics metadata init failed", __FUNCTION__); |
| *size = 0; |
| return -EINVAL; |
| } |
| // Call adreno api with the metadata blob to get buffer size |
| *size = adreno_mem_info->AdrenoGetAlignedGpuBufferSize(graphics_metadata->data); |
| return 0; |
| } |
| |
| bool CanUseAdrenoForSize(int buffer_type, uint64_t usage) { |
| if (buffer_type == BUFFER_TYPE_VIDEO || !GetAdrenoSizeAPIStatus()) { |
| return false; |
| } |
| |
| if ((usage & BufferUsage::PROTECTED) && ((usage & BufferUsage::CAMERA_OUTPUT) || |
| (usage & GRALLOC_USAGE_PRIVATE_SECURE_DISPLAY))) { |
| return false; |
| } |
| |
| return true; |
| } |
| |
| bool GetAdrenoSizeAPIStatus() { |
| AdrenoMemInfo* adreno_mem_info = AdrenoMemInfo::GetInstance(); |
| if (adreno_mem_info) { |
| return adreno_mem_info->AdrenoSizeAPIAvaliable(); |
| } |
| return false; |
| } |
| |
| bool UseUncached(int format, uint64_t usage) { |
| if ((usage & GRALLOC_USAGE_PRIVATE_UNCACHED) || (usage & BufferUsage::PROTECTED)) { |
| return true; |
| } |
| |
| // CPU read rarely |
| if ((usage & BufferUsage::CPU_READ_MASK) == static_cast<uint64_t>(BufferUsage::CPU_READ_RARELY)) { |
| return true; |
| } |
| |
| // CPU write rarely |
| if ((usage & BufferUsage::CPU_WRITE_MASK) == |
| static_cast<uint64_t>(BufferUsage::CPU_WRITE_RARELY)) { |
| return true; |
| } |
| |
| if ((usage & BufferUsage::SENSOR_DIRECT_DATA) || (usage & BufferUsage::GPU_DATA_BUFFER)) { |
| return true; |
| } |
| |
| if (format && IsUBwcEnabled(format, usage)) { |
| return true; |
| } |
| |
| return false; |
| } |
| |
| uint64_t GetHandleFlags(int format, uint64_t usage) { |
| uint64_t priv_flags = 0; |
| |
| if (usage & BufferUsage::VIDEO_ENCODER) { |
| priv_flags |= private_handle_t::PRIV_FLAGS_VIDEO_ENCODER; |
| } |
| |
| if (usage & BufferUsage::CAMERA_OUTPUT) { |
| priv_flags |= private_handle_t::PRIV_FLAGS_CAMERA_WRITE; |
| } |
| |
| if (usage & BufferUsage::CAMERA_INPUT) { |
| priv_flags |= private_handle_t::PRIV_FLAGS_CAMERA_READ; |
| } |
| |
| if (usage & BufferUsage::COMPOSER_OVERLAY) { |
| priv_flags |= private_handle_t::PRIV_FLAGS_DISP_CONSUMER; |
| } |
| |
| if (usage & BufferUsage::GPU_TEXTURE) { |
| priv_flags |= private_handle_t::PRIV_FLAGS_HW_TEXTURE; |
| } |
| |
| if (usage & GRALLOC_USAGE_PRIVATE_SECURE_DISPLAY) { |
| priv_flags |= private_handle_t::PRIV_FLAGS_SECURE_DISPLAY; |
| } |
| |
| if (IsUBwcEnabled(format, usage)) { |
| priv_flags |= private_handle_t::PRIV_FLAGS_UBWC_ALIGNED; |
| if (IsUBwcPISupported(format, usage)) { |
| priv_flags |= private_handle_t::PRIV_FLAGS_UBWC_ALIGNED_PI; |
| } |
| } |
| |
| if (usage & (BufferUsage::CPU_READ_MASK | BufferUsage::CPU_WRITE_MASK)) { |
| priv_flags |= private_handle_t::PRIV_FLAGS_CPU_RENDERED; |
| } |
| |
| if ((usage & (BufferUsage::VIDEO_ENCODER | BufferUsage::VIDEO_DECODER | |
| BufferUsage::CAMERA_OUTPUT | BufferUsage::GPU_RENDER_TARGET))) { |
| priv_flags |= private_handle_t::PRIV_FLAGS_NON_CPU_WRITER; |
| } |
| |
| if (!UseUncached(format, usage)) { |
| priv_flags |= private_handle_t::PRIV_FLAGS_CACHED; |
| } |
| |
| return priv_flags; |
| } |
| |
| int GetImplDefinedFormat(uint64_t usage, int format) { |
| int gr_format = format; |
| |
| // If input format is HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED then based on |
| // the usage bits, gralloc assigns a format. |
| if (format == HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED || |
| format == HAL_PIXEL_FORMAT_YCbCr_420_888) { |
| if ((usage & GRALLOC_USAGE_PRIVATE_ALLOC_UBWC || usage & GRALLOC_USAGE_PRIVATE_ALLOC_UBWC_PI) |
| && format != HAL_PIXEL_FORMAT_YCbCr_420_888) { |
| gr_format = HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC; |
| } else if (usage & BufferUsage::VIDEO_ENCODER) { |
| if (usage & GRALLOC_USAGE_PRIVATE_VIDEO_NV21_ENCODER) { |
| gr_format = HAL_PIXEL_FORMAT_NV21_ENCODEABLE; // NV21 |
| } else if (usage & GRALLOC_USAGE_PRIVATE_HEIF) { |
| gr_format = HAL_PIXEL_FORMAT_NV12_HEIF; |
| } else if (format == HAL_PIXEL_FORMAT_YCbCr_420_888) { |
| gr_format = HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS; |
| } else { |
| gr_format = HAL_PIXEL_FORMAT_NV12_ENCODEABLE; // NV12 |
| } |
| } else if (usage & BufferUsage::CAMERA_INPUT) { |
| if (usage & BufferUsage::CAMERA_OUTPUT) { |
| // Assumed ZSL if both producer and consumer camera flags set |
| gr_format = HAL_PIXEL_FORMAT_NV21_ZSL; // NV21 |
| } else { |
| gr_format = HAL_PIXEL_FORMAT_YCrCb_420_SP; // NV21 |
| } |
| } else if (usage & BufferUsage::CAMERA_OUTPUT) { |
| if (format == HAL_PIXEL_FORMAT_YCbCr_420_888) { |
| if ((usage & BufferUsage::PROTECTED) && (!CanAllocateZSLForSecureCamera())) { |
| gr_format = HAL_PIXEL_FORMAT_YCrCb_420_SP; // NV21 |
| } else { |
| gr_format = HAL_PIXEL_FORMAT_NV21_ZSL; // NV21 |
| } |
| } else { |
| #ifdef USE_YCRCB_CAMERA_PREVIEW |
| gr_format = HAL_PIXEL_FORMAT_YCrCb_420_SP; // NV21 preview |
| #elif USE_YCRCB_CAMERA_PREVIEW_VENUS |
| gr_format = HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS; // NV21 preview |
| #else |
| gr_format = HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS; // NV12 preview |
| #endif |
| } |
| } else if (usage & BufferUsage::COMPOSER_OVERLAY) { |
| // XXX: If we still haven't set a format, default to RGBA8888 |
| gr_format = HAL_PIXEL_FORMAT_RGBA_8888; |
| } else if (format == HAL_PIXEL_FORMAT_YCbCr_420_888) { |
| // If no other usage flags are detected, default the |
| // flexible YUV format to NV21_ZSL |
| gr_format = HAL_PIXEL_FORMAT_NV21_ZSL; |
| } |
| } |
| |
| return gr_format; |
| } |
| |
| int GetCustomFormatFlags(int format, uint64_t usage, |
| int *custom_format, uint64_t *priv_flags) { |
| *custom_format = GetImplDefinedFormat(usage, format); |
| *priv_flags = GetHandleFlags(*custom_format, usage); |
| |
| if (usage & GRALLOC_USAGE_PROTECTED) { |
| *priv_flags |= private_handle_t::PRIV_FLAGS_SECURE_BUFFER; |
| } |
| |
| *priv_flags |= private_handle_t::PRIV_FLAGS_USES_ION; |
| |
| return 0; |
| } |
| |
| int GetBufferType(int inputFormat) { |
| return IsYuvFormat(inputFormat) ? BUFFER_TYPE_VIDEO : BUFFER_TYPE_UI; |
| } |
| |
| // Here width and height are aligned width and aligned height. |
| int GetYUVPlaneInfo(const BufferInfo &info, int32_t format, int32_t width, int32_t height, |
| int32_t flags, int *plane_count, PlaneLayoutInfo *plane_info) { |
| int err = 0; |
| unsigned int y_stride, c_stride, y_height, c_height, y_size, c_size; |
| uint64_t yOffset, cOffset, crOffset, cbOffset; |
| int h_subsampling = 0, v_subsampling = 0; |
| if (IsCameraCustomFormat(format) && CameraInfo::GetInstance()) { |
| int result = CameraInfo::GetInstance()->GetCameraFormatPlaneInfo( |
| format, info.width, info.height, plane_count, plane_info); |
| if (result != 0) { |
| ALOGE( |
| "%s: Failed to get the plane info through camera library. width: %d, height: %d," |
| "format: %d, Error code: %d", |
| __FUNCTION__, width, height, format, result); |
| } |
| return result; |
| } |
| |
| switch (format) { |
| // Semiplanar |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP: |
| case HAL_PIXEL_FORMAT_YCbCr_422_SP: |
| case HAL_PIXEL_FORMAT_NV12_LINEAR_FLEX: |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: // Same as YCbCr_420_SP_VENUS |
| case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: |
| case HAL_PIXEL_FORMAT_NV12_HEIF: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_422_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: |
| *plane_count = 2; |
| GetYuvSPPlaneInfo(info, format, width, height, 1, plane_info); |
| GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); |
| plane_info[0].h_subsampling = 0; |
| plane_info[0].v_subsampling = 0; |
| plane_info[1].h_subsampling = h_subsampling; |
| plane_info[1].v_subsampling = v_subsampling; |
| break; |
| |
| case HAL_PIXEL_FORMAT_RAW16: |
| case HAL_PIXEL_FORMAT_RAW12: |
| case HAL_PIXEL_FORMAT_RAW10: |
| case HAL_PIXEL_FORMAT_RAW8: |
| *plane_count = 1; |
| GetRawPlaneInfo(format, info.width, info.height, plane_info); |
| break; |
| |
| case HAL_PIXEL_FORMAT_Y8: |
| *plane_count = 1; |
| GetYuvSPPlaneInfo(info, format, width, height, 1, plane_info); |
| GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); |
| plane_info[0].h_subsampling = h_subsampling; |
| plane_info[0].v_subsampling = v_subsampling; |
| break; |
| |
| case HAL_PIXEL_FORMAT_Y16: |
| *plane_count = 1; |
| GetYuvSPPlaneInfo(info, format, width, height, 2, plane_info); |
| GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); |
| plane_info[0].h_subsampling = h_subsampling; |
| plane_info[0].v_subsampling = v_subsampling; |
| break; |
| |
| #ifndef QMAA |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: |
| GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); |
| if (flags & LAYOUT_INTERLACED_FLAG) { |
| *plane_count = 8; |
| GetYuvUbwcInterlacedSPPlaneInfo(width, height, plane_info); |
| plane_info[0].step = plane_info[4].step = 1; |
| plane_info[1].step = plane_info[5].step = 2; |
| plane_info[0].h_subsampling = plane_info[4].h_subsampling = 0; |
| plane_info[0].v_subsampling = plane_info[4].v_subsampling = 0; |
| plane_info[1].h_subsampling = plane_info[5].h_subsampling = h_subsampling; |
| plane_info[1].v_subsampling = plane_info[5].v_subsampling = v_subsampling; |
| plane_info[2].h_subsampling = plane_info[3].h_subsampling = 0; |
| plane_info[2].v_subsampling = plane_info[3].v_subsampling = 0; |
| plane_info[2].step = plane_info[3].step = 0; |
| plane_info[6].h_subsampling = plane_info[7].h_subsampling = 0; |
| plane_info[6].v_subsampling = plane_info[7].v_subsampling = 0; |
| plane_info[6].step = plane_info[7].step = 0; |
| } else { |
| *plane_count = 4; |
| GetYuvUbwcSPPlaneInfo(width, height, COLOR_FMT_NV12_UBWC, plane_info); |
| plane_info[0].h_subsampling = 0; |
| plane_info[0].v_subsampling = 0; |
| plane_info[0].step = 1; |
| plane_info[1].h_subsampling = h_subsampling; |
| plane_info[1].v_subsampling = v_subsampling; |
| plane_info[1].step = 2; |
| plane_info[2].h_subsampling = plane_info[3].h_subsampling = 0; |
| plane_info[2].v_subsampling = plane_info[3].v_subsampling = 0; |
| plane_info[2].step = plane_info[3].step = 0; |
| } |
| break; |
| |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010: |
| *plane_count = 2; |
| GetYuvSPPlaneInfo(info, format, width, height, 2, plane_info); |
| GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); |
| plane_info[0].h_subsampling = 0; |
| plane_info[0].v_subsampling = 0; |
| plane_info[1].h_subsampling = h_subsampling; |
| plane_info[1].v_subsampling = v_subsampling; |
| break; |
| |
| case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: |
| *plane_count = 4; |
| GetYuvUbwcSPPlaneInfo(width, height, COLOR_FMT_NV12_BPP10_UBWC, plane_info); |
| GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); |
| plane_info[0].h_subsampling = 0; |
| plane_info[0].v_subsampling = 0; |
| plane_info[1].step = 1; |
| plane_info[1].h_subsampling = h_subsampling; |
| plane_info[1].v_subsampling = v_subsampling; |
| plane_info[1].step = 3; |
| plane_info[2].h_subsampling = plane_info[3].h_subsampling = 0; |
| plane_info[2].v_subsampling = plane_info[3].v_subsampling = 0; |
| plane_info[2].step = plane_info[3].step = 0; |
| break; |
| |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: |
| *plane_count = 4; |
| GetYuvUbwcSPPlaneInfo(width, height, COLOR_FMT_P010_UBWC, plane_info); |
| GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); |
| plane_info[0].h_subsampling = 0; |
| plane_info[0].v_subsampling = 0; |
| plane_info[1].step = 1; |
| plane_info[1].h_subsampling = h_subsampling; |
| plane_info[1].v_subsampling = v_subsampling; |
| plane_info[1].step = 4; |
| plane_info[2].h_subsampling = plane_info[3].h_subsampling = 0; |
| plane_info[2].v_subsampling = plane_info[3].v_subsampling = 0; |
| plane_info[2].step = plane_info[3].step = 0; |
| break; |
| |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: |
| *plane_count = 2; |
| y_stride = VENUS_Y_STRIDE(COLOR_FMT_P010, width); |
| c_stride = VENUS_UV_STRIDE(COLOR_FMT_P010, width); |
| y_height = VENUS_Y_SCANLINES(COLOR_FMT_P010, height); |
| y_size = y_stride * y_height; |
| yOffset = 0; |
| cOffset = y_size; |
| c_height = VENUS_UV_SCANLINES(COLOR_FMT_P010, INT(height)); |
| c_size = c_stride * c_height; |
| GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); |
| |
| plane_info[0].component = (PlaneComponent)PLANE_COMPONENT_Y; |
| plane_info[0].offset = (uint32_t)yOffset; |
| plane_info[0].stride = static_cast<int32_t>(UINT(width)); |
| plane_info[0].stride_bytes = static_cast<int32_t>(y_stride); |
| plane_info[0].scanlines = static_cast<int32_t>(y_height); |
| plane_info[0].size = static_cast<uint32_t>(y_size); |
| plane_info[0].step = 1; |
| plane_info[0].h_subsampling = 0; |
| plane_info[0].v_subsampling = 0; |
| |
| plane_info[1].component = (PlaneComponent)(PLANE_COMPONENT_Cb | PLANE_COMPONENT_Cr); |
| plane_info[1].offset = (uint32_t)cOffset; |
| plane_info[1].stride = static_cast<int32_t>(UINT(width)); |
| plane_info[1].stride_bytes = static_cast<int32_t>(c_stride); |
| plane_info[1].scanlines = static_cast<int32_t>(c_height); |
| plane_info[1].size = static_cast<uint32_t>(c_size); |
| plane_info[1].step = 4; |
| plane_info[1].h_subsampling = h_subsampling; |
| plane_info[1].v_subsampling = v_subsampling; |
| break; |
| #endif |
| // Planar |
| case HAL_PIXEL_FORMAT_YV12: |
| if ((info.width & 1) || (info.height & 1)) { |
| ALOGE("w or h is odd for the YV12 format"); |
| err = -EINVAL; |
| return err; |
| } |
| *plane_count = 3; |
| y_stride = width; |
| c_stride = ALIGN(width / 2, 16); |
| y_height = UINT(height); |
| y_size = (y_stride * y_height); |
| height = height >> 1; |
| c_height = UINT(height); |
| c_size = (c_stride * c_height); |
| yOffset = 0; |
| crOffset = y_size; |
| cbOffset = (y_size + c_size); |
| GetYuvSubSamplingFactor(format, &h_subsampling, &v_subsampling); |
| |
| plane_info[0].component = (PlaneComponent)PLANE_COMPONENT_Y; |
| plane_info[0].offset = (uint32_t)yOffset; |
| plane_info[0].stride = static_cast<int32_t>(UINT(width)); |
| plane_info[0].stride_bytes = static_cast<int32_t>(y_stride); |
| plane_info[0].scanlines = static_cast<int32_t>(y_height); |
| plane_info[0].size = static_cast<uint32_t>(y_size); |
| plane_info[0].step = 1; |
| plane_info[0].h_subsampling = 0; |
| plane_info[0].v_subsampling = 0; |
| |
| plane_info[1].component = (PlaneComponent)PLANE_COMPONENT_Cb; |
| plane_info[1].offset = (uint32_t)cbOffset; |
| plane_info[2].component = (PlaneComponent)PLANE_COMPONENT_Cr; |
| plane_info[2].offset = (uint32_t)crOffset; |
| for (int i = 1; i < 3; i++) { |
| plane_info[i].stride = static_cast<int32_t>(UINT(width)); |
| plane_info[i].stride_bytes = static_cast<int32_t>(c_stride); |
| plane_info[i].scanlines = static_cast<int32_t>(c_height); |
| plane_info[i].size = static_cast<uint32_t>(c_size); |
| plane_info[i].step = 1; |
| plane_info[i].h_subsampling = h_subsampling; |
| plane_info[i].v_subsampling = v_subsampling; |
| } |
| break; |
| case HAL_PIXEL_FORMAT_CbYCrY_422_I: |
| if (info.width & 1) { |
| ALOGE("width is odd for the YUV422_SP format"); |
| err = -EINVAL; |
| return err; |
| } |
| *plane_count = 1; |
| y_stride = width * 2; |
| y_height = UINT(height); |
| y_size = y_stride * y_height; |
| yOffset = 0; |
| plane_info[0].component = (PlaneComponent)PLANE_COMPONENT_Y; |
| plane_info[0].offset = (uint32_t)yOffset; |
| plane_info[0].stride = static_cast<int32_t>(UINT(width)); |
| plane_info[0].stride_bytes = static_cast<int32_t>(y_stride); |
| plane_info[0].scanlines = static_cast<int32_t>(y_height); |
| plane_info[0].size = static_cast<uint32_t>(y_size); |
| plane_info[0].step = 1; |
| plane_info[0].h_subsampling = 0; |
| plane_info[0].v_subsampling = 0; |
| break; |
| |
| // Unsupported formats |
| case HAL_PIXEL_FORMAT_YCbCr_422_I: |
| case HAL_PIXEL_FORMAT_YCrCb_422_I: |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: |
| default: |
| *plane_count = 0; |
| ALOGD("%s: Invalid format passed: 0x%x", __FUNCTION__, format); |
| err = -EINVAL; |
| } |
| return err; |
| } |
| |
| void GetYuvSubSamplingFactor(int32_t format, int *h_subsampling, int *v_subsampling) { |
| switch (format) { |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP: |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010: |
| case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: |
| case HAL_PIXEL_FORMAT_NV12_LINEAR_FLEX: |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_ADRENO: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP: |
| case HAL_PIXEL_FORMAT_NV12_ENCODEABLE: // Same as YCbCr_420_SP_VENUS |
| case HAL_PIXEL_FORMAT_NV21_ENCODEABLE: |
| case HAL_PIXEL_FORMAT_YV12: |
| case HAL_PIXEL_FORMAT_NV12_HEIF: |
| *h_subsampling = 1; |
| *v_subsampling = 1; |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_422_SP: |
| case HAL_PIXEL_FORMAT_YCrCb_422_SP: |
| case HAL_PIXEL_FORMAT_CbYCrY_422_I: |
| *h_subsampling = 1; |
| *v_subsampling = 0; |
| break; |
| case HAL_PIXEL_FORMAT_Y16: |
| case HAL_PIXEL_FORMAT_Y8: |
| case HAL_PIXEL_FORMAT_BLOB: |
| default: |
| *h_subsampling = 0; |
| *v_subsampling = 0; |
| break; |
| } |
| } |
| |
| void CopyPlaneLayoutInfotoAndroidYcbcr(uint64_t base, int plane_count, PlaneLayoutInfo *plane_info, |
| struct android_ycbcr *ycbcr) { |
| ycbcr->y = reinterpret_cast<void *>(base + plane_info[0].offset); |
| ycbcr->ystride = plane_info[0].stride_bytes; |
| if (plane_count == 1) { |
| ycbcr->cb = NULL; |
| ycbcr->cr = NULL; |
| ycbcr->cstride = 0; |
| ycbcr->chroma_step = 0; |
| } else if (plane_count == 2 || plane_count == 4 || plane_count == 8) { |
| /* For YUV semiplanar :- |
| * - In progressive & linear case plane count is 2 and plane_info[0] will |
| * contain info about Y plane and plane_info[1] will contain info about UV plane. |
| * - In progressive & compressed case plane count is 4 then plane_info[0] will |
| * contain info about Y plane and plane_info[1] will contain info about UV plane. |
| * Remaining two plane (plane_info[2] & plane_info[3]) contain info about the |
| * Y_Meta_Plane and UV_Meta_Plane. |
| * - In interlaced & compressed case plane count is 8 then plane_info[0], plane_info[1], |
| * plane_info[4] & plane_info[5] will contain info about Y_plane, UV_plane, Y_plane |
| * & UV_plane. Remaining plane will contain info about the meta planes. As in this case |
| * this API is called twice through GetYUVPlaneInfo() with address of plane_info[0] & |
| * plane_info[4], so this will calculate the information accordingly and will fill the |
| * ycbcr structure with interlaced plane info only. |
| */ |
| ycbcr->cb = reinterpret_cast<void *>(base + plane_info[1].offset); |
| ycbcr->cr = reinterpret_cast<void *>(base + plane_info[1].offset + 1); |
| ycbcr->cstride = plane_info[1].stride_bytes; |
| ycbcr->chroma_step = plane_info[1].step; |
| } else if (plane_count == 3) { |
| /* For YUV planar :- |
| * Plane size is 3 and plane_info[0], plane_info[1], plane_info[2] will |
| * contain info about y_plane, cb_plane and cr_plane accordingly. |
| */ |
| ycbcr->cb = reinterpret_cast<void *>(base + plane_info[1].offset); |
| ycbcr->cr = reinterpret_cast<void *>(base + plane_info[2].offset); |
| ycbcr->cstride = plane_info[1].stride_bytes; |
| ycbcr->chroma_step = plane_info[1].step; |
| } |
| } |
| |
| bool HasAlphaComponent(int32_t format) { |
| switch (format) { |
| case HAL_PIXEL_FORMAT_RGBA_8888: |
| case HAL_PIXEL_FORMAT_BGRA_8888: |
| case HAL_PIXEL_FORMAT_RGBA_5551: |
| case HAL_PIXEL_FORMAT_RGBA_4444: |
| case HAL_PIXEL_FORMAT_RGBA_1010102: |
| case HAL_PIXEL_FORMAT_ARGB_2101010: |
| case HAL_PIXEL_FORMAT_BGRA_1010102: |
| case HAL_PIXEL_FORMAT_ABGR_2101010: |
| case HAL_PIXEL_FORMAT_RGBA_FP16: |
| return true; |
| default: |
| return false; |
| } |
| } |
| |
| void GetRGBPlaneInfo(const BufferInfo &info, int32_t format, int32_t width, int32_t height, |
| int32_t /* flags */, int *plane_count, PlaneLayoutInfo *plane_info) { |
| uint64_t usage = info.usage; |
| *plane_count = 1; |
| plane_info->component = |
| (PlaneComponent)(PLANE_COMPONENT_R | PLANE_COMPONENT_G | PLANE_COMPONENT_B); |
| if (HasAlphaComponent(format)) { |
| plane_info->component = (PlaneComponent)(plane_info->component | PLANE_COMPONENT_A); |
| } |
| GetBufferSizeAndDimensions(info, &(plane_info->size), (unsigned int *) &width, |
| (unsigned int *) &height); |
| plane_info->step = GetBpp(format); |
| plane_info->offset = GetRgbMetaSize(format, width, height, usage); |
| plane_info->h_subsampling = 0; |
| plane_info->v_subsampling = 0; |
| plane_info->stride = width; |
| plane_info->stride_bytes = width * plane_info->step; |
| plane_info->scanlines = height; |
| } |
| |
| // TODO(tbalacha): tile vs ubwc -- may need to find a diff way to differentiate |
| void GetDRMFormat(uint32_t format, uint32_t flags, uint32_t *drm_format, |
| uint64_t *drm_format_modifier) { |
| bool compressed = (flags & private_handle_t::PRIV_FLAGS_UBWC_ALIGNED) ? true : false; |
| switch (format) { |
| case HAL_PIXEL_FORMAT_RGBA_8888: |
| *drm_format = DRM_FORMAT_ABGR8888; |
| break; |
| case HAL_PIXEL_FORMAT_RGBA_5551: |
| *drm_format = DRM_FORMAT_ABGR1555; |
| break; |
| case HAL_PIXEL_FORMAT_RGBA_4444: |
| *drm_format = DRM_FORMAT_ABGR4444; |
| break; |
| case HAL_PIXEL_FORMAT_BGRA_8888: |
| *drm_format = DRM_FORMAT_ARGB8888; |
| break; |
| case HAL_PIXEL_FORMAT_RGBX_8888: |
| *drm_format = DRM_FORMAT_XBGR8888; |
| if (compressed) |
| *drm_format_modifier = DRM_FORMAT_MOD_QCOM_COMPRESSED; |
| break; |
| case HAL_PIXEL_FORMAT_BGRX_8888: |
| *drm_format = DRM_FORMAT_XRGB8888; |
| break; |
| case HAL_PIXEL_FORMAT_RGB_888: |
| *drm_format = DRM_FORMAT_BGR888; |
| break; |
| case HAL_PIXEL_FORMAT_RGB_565: |
| *drm_format = DRM_FORMAT_BGR565; |
| break; |
| case HAL_PIXEL_FORMAT_BGR_565: |
| *drm_format = DRM_FORMAT_BGR565; |
| if (compressed) |
| *drm_format_modifier = DRM_FORMAT_MOD_QCOM_COMPRESSED; |
| break; |
| case HAL_PIXEL_FORMAT_RGBA_1010102: |
| *drm_format = DRM_FORMAT_ABGR2101010; |
| if (compressed) |
| *drm_format_modifier = DRM_FORMAT_MOD_QCOM_COMPRESSED; |
| break; |
| case HAL_PIXEL_FORMAT_ARGB_2101010: |
| *drm_format = DRM_FORMAT_BGRA1010102; |
| break; |
| case HAL_PIXEL_FORMAT_RGBX_1010102: |
| *drm_format = DRM_FORMAT_XBGR2101010; |
| if (compressed) |
| *drm_format_modifier = DRM_FORMAT_MOD_QCOM_COMPRESSED; |
| break; |
| case HAL_PIXEL_FORMAT_XRGB_2101010: |
| *drm_format = DRM_FORMAT_BGRX1010102; |
| break; |
| case HAL_PIXEL_FORMAT_BGRA_1010102: |
| *drm_format = DRM_FORMAT_ARGB2101010; |
| break; |
| case HAL_PIXEL_FORMAT_ABGR_2101010: |
| *drm_format = DRM_FORMAT_RGBA1010102; |
| break; |
| case HAL_PIXEL_FORMAT_BGRX_1010102: |
| *drm_format = DRM_FORMAT_XRGB2101010; |
| break; |
| case HAL_PIXEL_FORMAT_XBGR_2101010: |
| *drm_format = DRM_FORMAT_RGBX1010102; |
| break; |
| case HAL_PIXEL_FORMAT_NV12_LINEAR_FLEX: |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS: |
| *drm_format = DRM_FORMAT_NV12; |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_SP_VENUS_UBWC: |
| *drm_format = DRM_FORMAT_NV12; |
| if (compressed) { |
| *drm_format_modifier = DRM_FORMAT_MOD_QCOM_COMPRESSED; |
| } else { |
| *drm_format_modifier = DRM_FORMAT_MOD_QCOM_TILE; |
| } |
| break; |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP: |
| *drm_format = DRM_FORMAT_NV21; |
| break; |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP_VENUS: |
| *drm_format = DRM_FORMAT_NV21; |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010: |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_VENUS: |
| *drm_format = DRM_FORMAT_NV12; |
| *drm_format_modifier = DRM_FORMAT_MOD_QCOM_DX; |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_P010_UBWC: |
| *drm_format = DRM_FORMAT_NV12; |
| if (compressed) { |
| *drm_format_modifier = DRM_FORMAT_MOD_QCOM_COMPRESSED | DRM_FORMAT_MOD_QCOM_DX; |
| } else { |
| *drm_format_modifier = DRM_FORMAT_MOD_QCOM_TILE | DRM_FORMAT_MOD_QCOM_DX; |
| } |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_420_TP10_UBWC: |
| *drm_format = DRM_FORMAT_NV12; |
| if (compressed) { |
| *drm_format_modifier = |
| DRM_FORMAT_MOD_QCOM_COMPRESSED | DRM_FORMAT_MOD_QCOM_DX | DRM_FORMAT_MOD_QCOM_TIGHT; |
| } else { |
| *drm_format_modifier = |
| DRM_FORMAT_MOD_QCOM_TILE | DRM_FORMAT_MOD_QCOM_DX | DRM_FORMAT_MOD_QCOM_TIGHT; |
| } |
| break; |
| case HAL_PIXEL_FORMAT_YCbCr_422_SP: |
| *drm_format = DRM_FORMAT_NV16; |
| break; |
| /* |
| TODO: No HAL_PIXEL_FORMAT equivalent? |
| case kFormatYCrCb422H2V1SemiPlanar: |
| *drm_format = DRM_FORMAT_NV61; |
| break;*/ |
| case HAL_PIXEL_FORMAT_YV12: |
| *drm_format = DRM_FORMAT_YVU420; |
| break; |
| case HAL_PIXEL_FORMAT_RGBA_FP16: |
| ALOGW("HAL_PIXEL_FORMAT_RGBA_FP16 currently not supported"); |
| break; |
| default: |
| ALOGE("Unsupported format %d", format); |
| } |
| } |
| |
| bool CanAllocateZSLForSecureCamera() { |
| static bool inited = false; |
| static bool can_allocate = true; |
| if (inited) { |
| return can_allocate; |
| } |
| char property[PROPERTY_VALUE_MAX]; |
| property_get("vendor.gralloc.secure_preview_buffer_format", property, "0"); |
| if (!(strncmp(property, "420_sp", PROPERTY_VALUE_MAX))) { |
| can_allocate = false; |
| } |
| inited = true; |
| ALOGI("CanAllocateZSLForSecureCamera: %d", can_allocate); |
| |
| return can_allocate; |
| } |
| } // namespace gralloc |