|  | /* | 
|  | * Copyright (C) 2019 The Android Open Source Project | 
|  | * | 
|  | * Licensed under the Apache License, Version 2.0 (the "License"); | 
|  | * you may not use this file except in compliance with the License. | 
|  | * You may obtain a copy of the License at | 
|  | * | 
|  | *      http://www.apache.org/licenses/LICENSE-2.0 | 
|  | * | 
|  | * Unless required by applicable law or agreed to in writing, software | 
|  | * distributed under the License is distributed on an "AS IS" BASIS, | 
|  | * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | 
|  | * See the License for the specific language governing permissions and | 
|  | * limitations under the License. | 
|  | */ | 
|  |  | 
|  | //#define LOG_NDEBUG 0 | 
|  | #define LOG_TAG "C2SoftGav1Dec" | 
|  | #include "C2SoftGav1Dec.h" | 
|  |  | 
|  | #include <android-base/properties.h> | 
|  | #include <C2Debug.h> | 
|  | #include <C2PlatformSupport.h> | 
|  | #include <Codec2BufferUtils.h> | 
|  | #include <Codec2CommonUtils.h> | 
|  | #include <Codec2Mapper.h> | 
|  | #include <SimpleC2Interface.h> | 
|  | #include <libyuv.h> | 
|  | #include <log/log.h> | 
|  | #include <media/stagefright/foundation/AUtils.h> | 
|  | #include <media/stagefright/foundation/MediaDefs.h> | 
|  |  | 
|  | // libyuv version required for I410ToAB30Matrix and I210ToAB30Matrix. | 
|  | #if LIBYUV_VERSION >= 1780 | 
|  | #include <algorithm> | 
|  | #define HAVE_LIBYUV_I410_I210_TO_AB30 1 | 
|  | #else | 
|  | #define HAVE_LIBYUV_I410_I210_TO_AB30 0 | 
|  | #endif | 
|  |  | 
|  | namespace android { | 
|  |  | 
|  | // Property used to control the number of threads used in the gav1 decoder. | 
|  | constexpr char kNumThreadsProperty[] = "debug.c2.gav1.numthreads"; | 
|  |  | 
|  | // codecname set and passed in as a compile flag from Android.bp | 
|  | constexpr char COMPONENT_NAME[] = CODECNAME; | 
|  |  | 
|  | constexpr size_t kMinInputBufferSize = 2 * 1024 * 1024; | 
|  |  | 
|  | class C2SoftGav1Dec::IntfImpl : public SimpleInterface<void>::BaseParams { | 
|  | public: | 
|  | explicit IntfImpl(const std::shared_ptr<C2ReflectorHelper> &helper) | 
|  | : SimpleInterface<void>::BaseParams( | 
|  | helper, COMPONENT_NAME, C2Component::KIND_DECODER, | 
|  | C2Component::DOMAIN_VIDEO, MEDIA_MIMETYPE_VIDEO_AV1) { | 
|  | noPrivateBuffers();  // TODO: account for our buffers here. | 
|  | noInputReferences(); | 
|  | noOutputReferences(); | 
|  | noInputLatency(); | 
|  | noTimeStretch(); | 
|  |  | 
|  | addParameter(DefineParam(mAttrib, C2_PARAMKEY_COMPONENT_ATTRIBUTES) | 
|  | .withConstValue(new C2ComponentAttributesSetting( | 
|  | C2Component::ATTRIB_IS_TEMPORAL)) | 
|  | .build()); | 
|  |  | 
|  | addParameter( | 
|  | DefineParam(mSize, C2_PARAMKEY_PICTURE_SIZE) | 
|  | .withDefault(new C2StreamPictureSizeInfo::output(0u, 320, 240)) | 
|  | .withFields({ | 
|  | C2F(mSize, width).inRange(2, 4096), | 
|  | C2F(mSize, height).inRange(2, 4096), | 
|  | }) | 
|  | .withSetter(SizeSetter) | 
|  | .build()); | 
|  |  | 
|  | addParameter(DefineParam(mProfileLevel, C2_PARAMKEY_PROFILE_LEVEL) | 
|  | .withDefault(new C2StreamProfileLevelInfo::input( | 
|  | 0u, C2Config::PROFILE_AV1_0, C2Config::LEVEL_AV1_2_1)) | 
|  | .withFields({C2F(mProfileLevel, profile) | 
|  | .oneOf({C2Config::PROFILE_AV1_0, | 
|  | C2Config::PROFILE_AV1_1}), | 
|  | C2F(mProfileLevel, level) | 
|  | .oneOf({ | 
|  | C2Config::LEVEL_AV1_2, C2Config::LEVEL_AV1_2_1, | 
|  | C2Config::LEVEL_AV1_2_2, C2Config::LEVEL_AV1_2_3, | 
|  | C2Config::LEVEL_AV1_3, C2Config::LEVEL_AV1_3_1, | 
|  | C2Config::LEVEL_AV1_3_2, C2Config::LEVEL_AV1_3_3, | 
|  | C2Config::LEVEL_AV1_4, C2Config::LEVEL_AV1_4_1, | 
|  | C2Config::LEVEL_AV1_4_2, C2Config::LEVEL_AV1_4_3, | 
|  | C2Config::LEVEL_AV1_5, C2Config::LEVEL_AV1_5_1, | 
|  | C2Config::LEVEL_AV1_5_2, C2Config::LEVEL_AV1_5_3, | 
|  | })}) | 
|  | .withSetter(ProfileLevelSetter, mSize) | 
|  | .build()); | 
|  |  | 
|  | mHdr10PlusInfoInput = C2StreamHdr10PlusInfo::input::AllocShared(0); | 
|  | addParameter( | 
|  | DefineParam(mHdr10PlusInfoInput, C2_PARAMKEY_INPUT_HDR10_PLUS_INFO) | 
|  | .withDefault(mHdr10PlusInfoInput) | 
|  | .withFields({ | 
|  | C2F(mHdr10PlusInfoInput, m.value).any(), | 
|  | }) | 
|  | .withSetter(Hdr10PlusInfoInputSetter) | 
|  | .build()); | 
|  |  | 
|  | mHdr10PlusInfoOutput = C2StreamHdr10PlusInfo::output::AllocShared(0); | 
|  | addParameter( | 
|  | DefineParam(mHdr10PlusInfoOutput, C2_PARAMKEY_OUTPUT_HDR10_PLUS_INFO) | 
|  | .withDefault(mHdr10PlusInfoOutput) | 
|  | .withFields({ | 
|  | C2F(mHdr10PlusInfoOutput, m.value).any(), | 
|  | }) | 
|  | .withSetter(Hdr10PlusInfoOutputSetter) | 
|  | .build()); | 
|  |  | 
|  | // default static info | 
|  | C2HdrStaticMetadataStruct defaultStaticInfo{}; | 
|  | helper->addStructDescriptors<C2MasteringDisplayColorVolumeStruct, C2ColorXyStruct>(); | 
|  | addParameter( | 
|  | DefineParam(mHdrStaticInfo, C2_PARAMKEY_HDR_STATIC_INFO) | 
|  | .withDefault(new C2StreamHdrStaticInfo::output(0u, defaultStaticInfo)) | 
|  | .withFields({ | 
|  | C2F(mHdrStaticInfo, mastering.red.x).inRange(0, 1), | 
|  | C2F(mHdrStaticInfo, mastering.red.y).inRange(0, 1), | 
|  | C2F(mHdrStaticInfo, mastering.green.x).inRange(0, 1), | 
|  | C2F(mHdrStaticInfo, mastering.green.y).inRange(0, 1), | 
|  | C2F(mHdrStaticInfo, mastering.blue.x).inRange(0, 1), | 
|  | C2F(mHdrStaticInfo, mastering.blue.y).inRange(0, 1), | 
|  | C2F(mHdrStaticInfo, mastering.white.x).inRange(0, 1), | 
|  | C2F(mHdrStaticInfo, mastering.white.x).inRange(0, 1), | 
|  | C2F(mHdrStaticInfo, mastering.maxLuminance).inRange(0, 65535), | 
|  | C2F(mHdrStaticInfo, mastering.minLuminance).inRange(0, 6.5535), | 
|  | C2F(mHdrStaticInfo, maxCll).inRange(0, 0XFFFF), | 
|  | C2F(mHdrStaticInfo, maxFall).inRange(0, 0XFFFF) | 
|  | }) | 
|  | .withSetter(HdrStaticInfoSetter) | 
|  | .build()); | 
|  |  | 
|  | addParameter( | 
|  | DefineParam(mMaxSize, C2_PARAMKEY_MAX_PICTURE_SIZE) | 
|  | .withDefault(new C2StreamMaxPictureSizeTuning::output(0u, 320, 240)) | 
|  | .withFields({ | 
|  | C2F(mSize, width).inRange(2, 2048, 2), | 
|  | C2F(mSize, height).inRange(2, 2048, 2), | 
|  | }) | 
|  | .withSetter(MaxPictureSizeSetter, mSize) | 
|  | .build()); | 
|  |  | 
|  | addParameter(DefineParam(mMaxInputSize, C2_PARAMKEY_INPUT_MAX_BUFFER_SIZE) | 
|  | .withDefault(new C2StreamMaxBufferSizeInfo::input(0u, kMinInputBufferSize)) | 
|  | .withFields({ | 
|  | C2F(mMaxInputSize, value).any(), | 
|  | }) | 
|  | .calculatedAs(MaxInputSizeSetter, mMaxSize) | 
|  | .build()); | 
|  |  | 
|  | C2ChromaOffsetStruct locations[1] = {C2ChromaOffsetStruct::ITU_YUV_420_0()}; | 
|  | std::shared_ptr<C2StreamColorInfo::output> defaultColorInfo = | 
|  | C2StreamColorInfo::output::AllocShared(1u, 0u, 8u /* bitDepth */, | 
|  | C2Color::YUV_420); | 
|  | memcpy(defaultColorInfo->m.locations, locations, sizeof(locations)); | 
|  |  | 
|  | defaultColorInfo = C2StreamColorInfo::output::AllocShared( | 
|  | {C2ChromaOffsetStruct::ITU_YUV_420_0()}, 0u, 8u /* bitDepth */, | 
|  | C2Color::YUV_420); | 
|  | helper->addStructDescriptors<C2ChromaOffsetStruct>(); | 
|  |  | 
|  | addParameter(DefineParam(mColorInfo, C2_PARAMKEY_CODED_COLOR_INFO) | 
|  | .withConstValue(defaultColorInfo) | 
|  | .build()); | 
|  |  | 
|  | addParameter( | 
|  | DefineParam(mDefaultColorAspects, C2_PARAMKEY_DEFAULT_COLOR_ASPECTS) | 
|  | .withDefault(new C2StreamColorAspectsTuning::output( | 
|  | 0u, C2Color::RANGE_UNSPECIFIED, C2Color::PRIMARIES_UNSPECIFIED, | 
|  | C2Color::TRANSFER_UNSPECIFIED, C2Color::MATRIX_UNSPECIFIED)) | 
|  | .withFields( | 
|  | {C2F(mDefaultColorAspects, range) | 
|  | .inRange(C2Color::RANGE_UNSPECIFIED, C2Color::RANGE_OTHER), | 
|  | C2F(mDefaultColorAspects, primaries) | 
|  | .inRange(C2Color::PRIMARIES_UNSPECIFIED, | 
|  | C2Color::PRIMARIES_OTHER), | 
|  | C2F(mDefaultColorAspects, transfer) | 
|  | .inRange(C2Color::TRANSFER_UNSPECIFIED, | 
|  | C2Color::TRANSFER_OTHER), | 
|  | C2F(mDefaultColorAspects, matrix) | 
|  | .inRange(C2Color::MATRIX_UNSPECIFIED, | 
|  | C2Color::MATRIX_OTHER)}) | 
|  | .withSetter(DefaultColorAspectsSetter) | 
|  | .build()); | 
|  |  | 
|  | addParameter( | 
|  | DefineParam(mCodedColorAspects, C2_PARAMKEY_VUI_COLOR_ASPECTS) | 
|  | .withDefault(new C2StreamColorAspectsInfo::input( | 
|  | 0u, C2Color::RANGE_LIMITED, C2Color::PRIMARIES_UNSPECIFIED, | 
|  | C2Color::TRANSFER_UNSPECIFIED, C2Color::MATRIX_UNSPECIFIED)) | 
|  | .withFields({ | 
|  | C2F(mCodedColorAspects, range).inRange( | 
|  | C2Color::RANGE_UNSPECIFIED,     C2Color::RANGE_OTHER), | 
|  | C2F(mCodedColorAspects, primaries).inRange( | 
|  | C2Color::PRIMARIES_UNSPECIFIED, C2Color::PRIMARIES_OTHER), | 
|  | C2F(mCodedColorAspects, transfer).inRange( | 
|  | C2Color::TRANSFER_UNSPECIFIED,  C2Color::TRANSFER_OTHER), | 
|  | C2F(mCodedColorAspects, matrix).inRange( | 
|  | C2Color::MATRIX_UNSPECIFIED,    C2Color::MATRIX_OTHER) | 
|  | }) | 
|  | .withSetter(CodedColorAspectsSetter) | 
|  | .build()); | 
|  |  | 
|  | addParameter( | 
|  | DefineParam(mColorAspects, C2_PARAMKEY_COLOR_ASPECTS) | 
|  | .withDefault(new C2StreamColorAspectsInfo::output( | 
|  | 0u, C2Color::RANGE_UNSPECIFIED, C2Color::PRIMARIES_UNSPECIFIED, | 
|  | C2Color::TRANSFER_UNSPECIFIED, C2Color::MATRIX_UNSPECIFIED)) | 
|  | .withFields({ | 
|  | C2F(mColorAspects, range).inRange( | 
|  | C2Color::RANGE_UNSPECIFIED,     C2Color::RANGE_OTHER), | 
|  | C2F(mColorAspects, primaries).inRange( | 
|  | C2Color::PRIMARIES_UNSPECIFIED, C2Color::PRIMARIES_OTHER), | 
|  | C2F(mColorAspects, transfer).inRange( | 
|  | C2Color::TRANSFER_UNSPECIFIED,  C2Color::TRANSFER_OTHER), | 
|  | C2F(mColorAspects, matrix).inRange( | 
|  | C2Color::MATRIX_UNSPECIFIED,    C2Color::MATRIX_OTHER) | 
|  | }) | 
|  | .withSetter(ColorAspectsSetter, mDefaultColorAspects, mCodedColorAspects) | 
|  | .build()); | 
|  |  | 
|  | std::vector<uint32_t> pixelFormats = {HAL_PIXEL_FORMAT_YCBCR_420_888}; | 
|  | if (isHalPixelFormatSupported((AHardwareBuffer_Format)HAL_PIXEL_FORMAT_YCBCR_P010)) { | 
|  | pixelFormats.push_back(HAL_PIXEL_FORMAT_YCBCR_P010); | 
|  | } | 
|  | // If color format surface isn't added to supported formats, there is no way to know | 
|  | // when the color-format is configured to surface. This is necessary to be able to | 
|  | // choose 10-bit format while decoding 10-bit clips in surface mode. | 
|  | pixelFormats.push_back(HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED); | 
|  |  | 
|  | // TODO: support more formats? | 
|  | addParameter( | 
|  | DefineParam(mPixelFormat, C2_PARAMKEY_PIXEL_FORMAT) | 
|  | .withDefault(new C2StreamPixelFormatInfo::output( | 
|  | 0u, HAL_PIXEL_FORMAT_YCBCR_420_888)) | 
|  | .withFields({C2F(mPixelFormat, value).oneOf(pixelFormats)}) | 
|  | .withSetter((Setter<decltype(*mPixelFormat)>::StrictValueWithNoDeps)) | 
|  | .build()); | 
|  | } | 
|  |  | 
|  | static C2R SizeSetter(bool mayBlock, | 
|  | const C2P<C2StreamPictureSizeInfo::output> &oldMe, | 
|  | C2P<C2StreamPictureSizeInfo::output> &me) { | 
|  | (void)mayBlock; | 
|  | C2R res = C2R::Ok(); | 
|  | if (!me.F(me.v.width).supportsAtAll(me.v.width)) { | 
|  | res = res.plus(C2SettingResultBuilder::BadValue(me.F(me.v.width))); | 
|  | me.set().width = oldMe.v.width; | 
|  | } | 
|  | if (!me.F(me.v.height).supportsAtAll(me.v.height)) { | 
|  | res = res.plus(C2SettingResultBuilder::BadValue(me.F(me.v.height))); | 
|  | me.set().height = oldMe.v.height; | 
|  | } | 
|  | return res; | 
|  | } | 
|  |  | 
|  | static C2R MaxPictureSizeSetter( | 
|  | bool mayBlock, C2P<C2StreamMaxPictureSizeTuning::output> &me, | 
|  | const C2P<C2StreamPictureSizeInfo::output> &size) { | 
|  | (void)mayBlock; | 
|  | // TODO: get max width/height from the size's field helpers vs. | 
|  | // hardcoding | 
|  | me.set().width = c2_min(c2_max(me.v.width, size.v.width), 4096u); | 
|  | me.set().height = c2_min(c2_max(me.v.height, size.v.height), 4096u); | 
|  | return C2R::Ok(); | 
|  | } | 
|  |  | 
|  | static C2R MaxInputSizeSetter( | 
|  | bool mayBlock, C2P<C2StreamMaxBufferSizeInfo::input> &me, | 
|  | const C2P<C2StreamMaxPictureSizeTuning::output> &maxSize) { | 
|  | (void)mayBlock; | 
|  | // assume compression ratio of 2, but enforce a floor | 
|  | me.set().value = c2_max((((maxSize.v.width + 63) / 64) | 
|  | * ((maxSize.v.height + 63) / 64) * 3072), kMinInputBufferSize); | 
|  | return C2R::Ok(); | 
|  | } | 
|  |  | 
|  | static C2R DefaultColorAspectsSetter( | 
|  | bool mayBlock, C2P<C2StreamColorAspectsTuning::output> &me) { | 
|  | (void)mayBlock; | 
|  | if (me.v.range > C2Color::RANGE_OTHER) { | 
|  | me.set().range = C2Color::RANGE_OTHER; | 
|  | } | 
|  | if (me.v.primaries > C2Color::PRIMARIES_OTHER) { | 
|  | me.set().primaries = C2Color::PRIMARIES_OTHER; | 
|  | } | 
|  | if (me.v.transfer > C2Color::TRANSFER_OTHER) { | 
|  | me.set().transfer = C2Color::TRANSFER_OTHER; | 
|  | } | 
|  | if (me.v.matrix > C2Color::MATRIX_OTHER) { | 
|  | me.set().matrix = C2Color::MATRIX_OTHER; | 
|  | } | 
|  | return C2R::Ok(); | 
|  | } | 
|  |  | 
|  | static C2R CodedColorAspectsSetter(bool mayBlock, C2P<C2StreamColorAspectsInfo::input> &me) { | 
|  | (void)mayBlock; | 
|  | if (me.v.range > C2Color::RANGE_OTHER) { | 
|  | me.set().range = C2Color::RANGE_OTHER; | 
|  | } | 
|  | if (me.v.primaries > C2Color::PRIMARIES_OTHER) { | 
|  | me.set().primaries = C2Color::PRIMARIES_OTHER; | 
|  | } | 
|  | if (me.v.transfer > C2Color::TRANSFER_OTHER) { | 
|  | me.set().transfer = C2Color::TRANSFER_OTHER; | 
|  | } | 
|  | if (me.v.matrix > C2Color::MATRIX_OTHER) { | 
|  | me.set().matrix = C2Color::MATRIX_OTHER; | 
|  | } | 
|  | return C2R::Ok(); | 
|  | } | 
|  |  | 
|  | static C2R ColorAspectsSetter(bool mayBlock, C2P<C2StreamColorAspectsInfo::output> &me, | 
|  | const C2P<C2StreamColorAspectsTuning::output> &def, | 
|  | const C2P<C2StreamColorAspectsInfo::input> &coded) { | 
|  | (void)mayBlock; | 
|  | // take default values for all unspecified fields, and coded values for specified ones | 
|  | me.set().range = coded.v.range == RANGE_UNSPECIFIED ? def.v.range : coded.v.range; | 
|  | me.set().primaries = coded.v.primaries == PRIMARIES_UNSPECIFIED | 
|  | ? def.v.primaries : coded.v.primaries; | 
|  | me.set().transfer = coded.v.transfer == TRANSFER_UNSPECIFIED | 
|  | ? def.v.transfer : coded.v.transfer; | 
|  | me.set().matrix = coded.v.matrix == MATRIX_UNSPECIFIED ? def.v.matrix : coded.v.matrix; | 
|  | return C2R::Ok(); | 
|  | } | 
|  |  | 
|  | static C2R ProfileLevelSetter( | 
|  | bool mayBlock, C2P<C2StreamProfileLevelInfo::input> &me, | 
|  | const C2P<C2StreamPictureSizeInfo::output> &size) { | 
|  | (void)mayBlock; | 
|  | (void)size; | 
|  | (void)me;  // TODO: validate | 
|  | return C2R::Ok(); | 
|  | } | 
|  |  | 
|  | std::shared_ptr<C2StreamColorAspectsTuning::output> | 
|  | getDefaultColorAspects_l() { | 
|  | return mDefaultColorAspects; | 
|  | } | 
|  |  | 
|  | std::shared_ptr<C2StreamColorAspectsInfo::output> getColorAspects_l() { | 
|  | return mColorAspects; | 
|  | } | 
|  |  | 
|  | static C2R Hdr10PlusInfoInputSetter(bool mayBlock, | 
|  | C2P<C2StreamHdr10PlusInfo::input> &me) { | 
|  | (void)mayBlock; | 
|  | (void)me;  // TODO: validate | 
|  | return C2R::Ok(); | 
|  | } | 
|  |  | 
|  | static C2R Hdr10PlusInfoOutputSetter(bool mayBlock, | 
|  | C2P<C2StreamHdr10PlusInfo::output> &me) { | 
|  | (void)mayBlock; | 
|  | (void)me;  // TODO: validate | 
|  | return C2R::Ok(); | 
|  | } | 
|  |  | 
|  | // unsafe getters | 
|  | std::shared_ptr<C2StreamPixelFormatInfo::output> getPixelFormat_l() const { return mPixelFormat; } | 
|  |  | 
|  | static C2R HdrStaticInfoSetter(bool mayBlock, C2P<C2StreamHdrStaticInfo::output> &me) { | 
|  | (void)mayBlock; | 
|  | if (me.v.mastering.red.x > 1) { | 
|  | me.set().mastering.red.x = 1; | 
|  | } | 
|  | if (me.v.mastering.red.y > 1) { | 
|  | me.set().mastering.red.y = 1; | 
|  | } | 
|  | if (me.v.mastering.green.x > 1) { | 
|  | me.set().mastering.green.x = 1; | 
|  | } | 
|  | if (me.v.mastering.green.y > 1) { | 
|  | me.set().mastering.green.y = 1; | 
|  | } | 
|  | if (me.v.mastering.blue.x > 1) { | 
|  | me.set().mastering.blue.x = 1; | 
|  | } | 
|  | if (me.v.mastering.blue.y > 1) { | 
|  | me.set().mastering.blue.y = 1; | 
|  | } | 
|  | if (me.v.mastering.white.x > 1) { | 
|  | me.set().mastering.white.x = 1; | 
|  | } | 
|  | if (me.v.mastering.white.y > 1) { | 
|  | me.set().mastering.white.y = 1; | 
|  | } | 
|  | if (me.v.mastering.maxLuminance > 65535.0) { | 
|  | me.set().mastering.maxLuminance = 65535.0; | 
|  | } | 
|  | if (me.v.mastering.minLuminance > 6.5535) { | 
|  | me.set().mastering.minLuminance = 6.5535; | 
|  | } | 
|  | if (me.v.maxCll > 65535.0) { | 
|  | me.set().maxCll = 65535.0; | 
|  | } | 
|  | if (me.v.maxFall > 65535.0) { | 
|  | me.set().maxFall = 65535.0; | 
|  | } | 
|  | return C2R::Ok(); | 
|  | } | 
|  |  | 
|  | private: | 
|  | std::shared_ptr<C2StreamProfileLevelInfo::input> mProfileLevel; | 
|  | std::shared_ptr<C2StreamPictureSizeInfo::output> mSize; | 
|  | std::shared_ptr<C2StreamMaxPictureSizeTuning::output> mMaxSize; | 
|  | std::shared_ptr<C2StreamMaxBufferSizeInfo::input> mMaxInputSize; | 
|  | std::shared_ptr<C2StreamColorInfo::output> mColorInfo; | 
|  | std::shared_ptr<C2StreamPixelFormatInfo::output> mPixelFormat; | 
|  | std::shared_ptr<C2StreamColorAspectsTuning::output> mDefaultColorAspects; | 
|  | std::shared_ptr<C2StreamColorAspectsInfo::input> mCodedColorAspects; | 
|  | std::shared_ptr<C2StreamColorAspectsInfo::output> mColorAspects; | 
|  | std::shared_ptr<C2StreamHdr10PlusInfo::input> mHdr10PlusInfoInput; | 
|  | std::shared_ptr<C2StreamHdr10PlusInfo::output> mHdr10PlusInfoOutput; | 
|  | std::shared_ptr<C2StreamHdrStaticInfo::output> mHdrStaticInfo; | 
|  | }; | 
|  |  | 
|  | C2SoftGav1Dec::C2SoftGav1Dec(const char *name, c2_node_id_t id, | 
|  | const std::shared_ptr<IntfImpl> &intfImpl) | 
|  | : SimpleC2Component( | 
|  | std::make_shared<SimpleInterface<IntfImpl>>(name, id, intfImpl)), | 
|  | mIntf(intfImpl), | 
|  | mCodecCtx(nullptr) { | 
|  | mTimeStart = mTimeEnd = systemTime(); | 
|  | } | 
|  |  | 
|  | C2SoftGav1Dec::~C2SoftGav1Dec() { onRelease(); } | 
|  |  | 
|  | c2_status_t C2SoftGav1Dec::onInit() { | 
|  | return initDecoder() ? C2_OK : C2_CORRUPTED; | 
|  | } | 
|  |  | 
|  | c2_status_t C2SoftGav1Dec::onStop() { | 
|  | mSignalledError = false; | 
|  | mSignalledOutputEos = false; | 
|  | return C2_OK; | 
|  | } | 
|  |  | 
|  | void C2SoftGav1Dec::onReset() { | 
|  | (void)onStop(); | 
|  | c2_status_t err = onFlush_sm(); | 
|  | if (err != C2_OK) { | 
|  | ALOGW("Failed to flush the av1 decoder. Trying to hard reset."); | 
|  | destroyDecoder(); | 
|  | if (!initDecoder()) { | 
|  | ALOGE("Hard reset failed."); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | void C2SoftGav1Dec::onRelease() { destroyDecoder(); } | 
|  |  | 
|  | c2_status_t C2SoftGav1Dec::onFlush_sm() { | 
|  | Libgav1StatusCode status = mCodecCtx->SignalEOS(); | 
|  | if (status != kLibgav1StatusOk) { | 
|  | ALOGE("Failed to flush av1 decoder. status: %d.", status); | 
|  | return C2_CORRUPTED; | 
|  | } | 
|  |  | 
|  | // Dequeue frame (if any) that was enqueued previously. | 
|  | const libgav1::DecoderBuffer *buffer; | 
|  | status = mCodecCtx->DequeueFrame(&buffer); | 
|  | if (status != kLibgav1StatusOk && status != kLibgav1StatusNothingToDequeue) { | 
|  | ALOGE("Failed to dequeue frame after flushing the av1 decoder. status: %d", | 
|  | status); | 
|  | return C2_CORRUPTED; | 
|  | } | 
|  |  | 
|  | mSignalledError = false; | 
|  | mSignalledOutputEos = false; | 
|  |  | 
|  | return C2_OK; | 
|  | } | 
|  |  | 
|  | static int GetCPUCoreCount() { | 
|  | int cpuCoreCount = 1; | 
|  | #if defined(_SC_NPROCESSORS_ONLN) | 
|  | cpuCoreCount = sysconf(_SC_NPROCESSORS_ONLN); | 
|  | #else | 
|  | // _SC_NPROC_ONLN must be defined... | 
|  | cpuCoreCount = sysconf(_SC_NPROC_ONLN); | 
|  | #endif | 
|  | CHECK(cpuCoreCount >= 1); | 
|  | ALOGV("Number of CPU cores: %d", cpuCoreCount); | 
|  | return cpuCoreCount; | 
|  | } | 
|  |  | 
|  | bool C2SoftGav1Dec::initDecoder() { | 
|  | mSignalledError = false; | 
|  | mSignalledOutputEos = false; | 
|  | mHalPixelFormat = HAL_PIXEL_FORMAT_YV12; | 
|  | { | 
|  | IntfImpl::Lock lock = mIntf->lock(); | 
|  | mPixelFormatInfo = mIntf->getPixelFormat_l(); | 
|  | } | 
|  | mCodecCtx.reset(new libgav1::Decoder()); | 
|  |  | 
|  | if (mCodecCtx == nullptr) { | 
|  | ALOGE("mCodecCtx is null"); | 
|  | return false; | 
|  | } | 
|  |  | 
|  | libgav1::DecoderSettings settings = {}; | 
|  | settings.threads = GetCPUCoreCount(); | 
|  | int32_t numThreads = android::base::GetIntProperty(kNumThreadsProperty, 0); | 
|  | if (numThreads > 0 && numThreads < settings.threads) { | 
|  | settings.threads = numThreads; | 
|  | } | 
|  |  | 
|  | ALOGV("Using libgav1 AV1 software decoder."); | 
|  | Libgav1StatusCode status = mCodecCtx->Init(&settings); | 
|  | if (status != kLibgav1StatusOk) { | 
|  | ALOGE("av1 decoder failed to initialize. status: %d.", status); | 
|  | return false; | 
|  | } | 
|  |  | 
|  | return true; | 
|  | } | 
|  |  | 
|  | void C2SoftGav1Dec::destroyDecoder() { mCodecCtx = nullptr; } | 
|  |  | 
|  | void fillEmptyWork(const std::unique_ptr<C2Work> &work) { | 
|  | uint32_t flags = 0; | 
|  | if (work->input.flags & C2FrameData::FLAG_END_OF_STREAM) { | 
|  | flags |= C2FrameData::FLAG_END_OF_STREAM; | 
|  | ALOGV("signalling eos"); | 
|  | } | 
|  | work->worklets.front()->output.flags = (C2FrameData::flags_t)flags; | 
|  | work->worklets.front()->output.buffers.clear(); | 
|  | work->worklets.front()->output.ordinal = work->input.ordinal; | 
|  | work->workletsProcessed = 1u; | 
|  | } | 
|  |  | 
|  | void C2SoftGav1Dec::finishWork(uint64_t index, | 
|  | const std::unique_ptr<C2Work> &work, | 
|  | const std::shared_ptr<C2GraphicBlock> &block) { | 
|  | std::shared_ptr<C2Buffer> buffer = | 
|  | createGraphicBuffer(block, C2Rect(mWidth, mHeight)); | 
|  | { | 
|  | IntfImpl::Lock lock = mIntf->lock(); | 
|  | buffer->setInfo(mIntf->getColorAspects_l()); | 
|  | } | 
|  | auto fillWork = [buffer, index](const std::unique_ptr<C2Work> &work) { | 
|  | uint32_t flags = 0; | 
|  | if ((work->input.flags & C2FrameData::FLAG_END_OF_STREAM) && | 
|  | (c2_cntr64_t(index) == work->input.ordinal.frameIndex)) { | 
|  | flags |= C2FrameData::FLAG_END_OF_STREAM; | 
|  | ALOGV("signalling eos"); | 
|  | } | 
|  | work->worklets.front()->output.flags = (C2FrameData::flags_t)flags; | 
|  | work->worklets.front()->output.buffers.clear(); | 
|  | work->worklets.front()->output.buffers.push_back(buffer); | 
|  | work->worklets.front()->output.ordinal = work->input.ordinal; | 
|  | work->workletsProcessed = 1u; | 
|  | }; | 
|  | if (work && c2_cntr64_t(index) == work->input.ordinal.frameIndex) { | 
|  | fillWork(work); | 
|  | } else { | 
|  | finish(index, fillWork); | 
|  | } | 
|  | } | 
|  |  | 
|  | void C2SoftGav1Dec::process(const std::unique_ptr<C2Work> &work, | 
|  | const std::shared_ptr<C2BlockPool> &pool) { | 
|  | work->result = C2_OK; | 
|  | work->workletsProcessed = 0u; | 
|  | work->worklets.front()->output.configUpdate.clear(); | 
|  | work->worklets.front()->output.flags = work->input.flags; | 
|  | if (mSignalledError || mSignalledOutputEos) { | 
|  | work->result = C2_BAD_VALUE; | 
|  | return; | 
|  | } | 
|  |  | 
|  | size_t inOffset = 0u; | 
|  | size_t inSize = 0u; | 
|  | C2ReadView rView = mDummyReadView; | 
|  | if (!work->input.buffers.empty()) { | 
|  | rView = work->input.buffers[0]->data().linearBlocks().front().map().get(); | 
|  | inSize = rView.capacity(); | 
|  | if (inSize && rView.error()) { | 
|  | ALOGE("read view map failed %d", rView.error()); | 
|  | work->result = C2_CORRUPTED; | 
|  | return; | 
|  | } | 
|  | } | 
|  |  | 
|  | bool codecConfig = | 
|  | ((work->input.flags & C2FrameData::FLAG_CODEC_CONFIG) != 0); | 
|  | bool eos = ((work->input.flags & C2FrameData::FLAG_END_OF_STREAM) != 0); | 
|  |  | 
|  | ALOGV("in buffer attr. size %zu timestamp %d frameindex %d, flags %x", inSize, | 
|  | (int)work->input.ordinal.timestamp.peeku(), | 
|  | (int)work->input.ordinal.frameIndex.peeku(), work->input.flags); | 
|  |  | 
|  | if (codecConfig) { | 
|  | fillEmptyWork(work); | 
|  | return; | 
|  | } | 
|  |  | 
|  | int64_t frameIndex = work->input.ordinal.frameIndex.peekll(); | 
|  | if (inSize) { | 
|  | uint8_t *bitstream = const_cast<uint8_t *>(rView.data() + inOffset); | 
|  |  | 
|  | mTimeStart = systemTime(); | 
|  | nsecs_t delay = mTimeStart - mTimeEnd; | 
|  |  | 
|  | const Libgav1StatusCode status = | 
|  | mCodecCtx->EnqueueFrame(bitstream, inSize, frameIndex, | 
|  | /*buffer_private_data=*/nullptr); | 
|  |  | 
|  | mTimeEnd = systemTime(); | 
|  | nsecs_t decodeTime = mTimeEnd - mTimeStart; | 
|  | ALOGV("decodeTime=%4" PRId64 " delay=%4" PRId64 "\n", decodeTime, delay); | 
|  |  | 
|  | if (status != kLibgav1StatusOk) { | 
|  | ALOGE("av1 decoder failed to decode frame. status: %d.", status); | 
|  | work->result = C2_CORRUPTED; | 
|  | work->workletsProcessed = 1u; | 
|  | mSignalledError = true; | 
|  | return; | 
|  | } | 
|  |  | 
|  | } | 
|  |  | 
|  | (void)outputBuffer(pool, work); | 
|  |  | 
|  | if (eos) { | 
|  | drainInternal(DRAIN_COMPONENT_WITH_EOS, pool, work); | 
|  | mSignalledOutputEos = true; | 
|  | } else if (!inSize) { | 
|  | fillEmptyWork(work); | 
|  | } | 
|  | } | 
|  |  | 
|  | void C2SoftGav1Dec::getHDRStaticParams(const libgav1::DecoderBuffer *buffer, | 
|  | const std::unique_ptr<C2Work> &work) { | 
|  | C2StreamHdrStaticMetadataInfo::output hdrStaticMetadataInfo{}; | 
|  | bool infoPresent = false; | 
|  | if (buffer->has_hdr_mdcv) { | 
|  | // hdr_mdcv.primary_chromaticity_* values are in 0.16 fixed-point format. | 
|  | hdrStaticMetadataInfo.mastering.red.x = buffer->hdr_mdcv.primary_chromaticity_x[0] / 65536.0; | 
|  | hdrStaticMetadataInfo.mastering.red.y = buffer->hdr_mdcv.primary_chromaticity_y[0] / 65536.0; | 
|  |  | 
|  | hdrStaticMetadataInfo.mastering.green.x = buffer->hdr_mdcv.primary_chromaticity_x[1] / 65536.0; | 
|  | hdrStaticMetadataInfo.mastering.green.y = buffer->hdr_mdcv.primary_chromaticity_y[1] / 65536.0; | 
|  |  | 
|  | hdrStaticMetadataInfo.mastering.blue.x = buffer->hdr_mdcv.primary_chromaticity_x[2] / 65536.0; | 
|  | hdrStaticMetadataInfo.mastering.blue.y = buffer->hdr_mdcv.primary_chromaticity_y[2] / 65536.0; | 
|  |  | 
|  | // hdr_mdcv.white_point_chromaticity_* values are in 0.16 fixed-point format. | 
|  | hdrStaticMetadataInfo.mastering.white.x = buffer->hdr_mdcv.white_point_chromaticity_x / 65536.0; | 
|  | hdrStaticMetadataInfo.mastering.white.y = buffer->hdr_mdcv.white_point_chromaticity_y / 65536.0; | 
|  |  | 
|  | // hdr_mdcv.luminance_max is in 24.8 fixed-point format. | 
|  | hdrStaticMetadataInfo.mastering.maxLuminance = buffer->hdr_mdcv.luminance_max / 256.0; | 
|  | // hdr_mdcv.luminance_min is in 18.14 format. | 
|  | hdrStaticMetadataInfo.mastering.minLuminance = buffer->hdr_mdcv.luminance_min / 16384.0; | 
|  | infoPresent = true; | 
|  | } | 
|  |  | 
|  | if (buffer->has_hdr_cll) { | 
|  | hdrStaticMetadataInfo.maxCll = buffer->hdr_cll.max_cll; | 
|  | hdrStaticMetadataInfo.maxFall = buffer->hdr_cll.max_fall; | 
|  | infoPresent = true; | 
|  | } | 
|  | // config if static info has changed | 
|  | if (infoPresent && !(hdrStaticMetadataInfo == mHdrStaticMetadataInfo)) { | 
|  | mHdrStaticMetadataInfo = hdrStaticMetadataInfo; | 
|  | work->worklets.front()->output.configUpdate.push_back(C2Param::Copy(mHdrStaticMetadataInfo)); | 
|  | } | 
|  | } | 
|  |  | 
|  | void C2SoftGav1Dec::getHDR10PlusInfoData(const libgav1::DecoderBuffer *buffer, | 
|  | const std::unique_ptr<C2Work> &work) { | 
|  | if (buffer->has_itut_t35) { | 
|  | std::vector<uint8_t> payload; | 
|  | size_t payloadSize = buffer->itut_t35.payload_size; | 
|  | if (payloadSize > 0) { | 
|  | payload.push_back(buffer->itut_t35.country_code); | 
|  | if (buffer->itut_t35.country_code == 0xFF) { | 
|  | payload.push_back(buffer->itut_t35.country_code_extension_byte); | 
|  | } | 
|  | payload.insert(payload.end(), buffer->itut_t35.payload_bytes, | 
|  | buffer->itut_t35.payload_bytes + buffer->itut_t35.payload_size); | 
|  | } | 
|  |  | 
|  | std::unique_ptr<C2StreamHdr10PlusInfo::output> hdr10PlusInfo = | 
|  | C2StreamHdr10PlusInfo::output::AllocUnique(payload.size()); | 
|  | if (!hdr10PlusInfo) { | 
|  | ALOGE("Hdr10PlusInfo allocation failed"); | 
|  | mSignalledError = true; | 
|  | work->result = C2_NO_MEMORY; | 
|  | return; | 
|  | } | 
|  | memcpy(hdr10PlusInfo->m.value, payload.data(), payload.size()); | 
|  |  | 
|  | // config if hdr10Plus info has changed | 
|  | if (nullptr == mHdr10PlusInfo || !(*hdr10PlusInfo == *mHdr10PlusInfo)) { | 
|  | mHdr10PlusInfo = std::move(hdr10PlusInfo); | 
|  | work->worklets.front()->output.configUpdate.push_back(std::move(mHdr10PlusInfo)); | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | void C2SoftGav1Dec::getVuiParams(const libgav1::DecoderBuffer *buffer) { | 
|  | VuiColorAspects vuiColorAspects; | 
|  | vuiColorAspects.primaries = buffer->color_primary; | 
|  | vuiColorAspects.transfer = buffer->transfer_characteristics; | 
|  | vuiColorAspects.coeffs = buffer->matrix_coefficients; | 
|  | vuiColorAspects.fullRange = buffer->color_range; | 
|  |  | 
|  | // convert vui aspects to C2 values if changed | 
|  | if (!(vuiColorAspects == mBitstreamColorAspects)) { | 
|  | mBitstreamColorAspects = vuiColorAspects; | 
|  | ColorAspects sfAspects; | 
|  | C2StreamColorAspectsInfo::input codedAspects = { 0u }; | 
|  | ColorUtils::convertIsoColorAspectsToCodecAspects( | 
|  | vuiColorAspects.primaries, vuiColorAspects.transfer, vuiColorAspects.coeffs, | 
|  | vuiColorAspects.fullRange, sfAspects); | 
|  | if (!C2Mapper::map(sfAspects.mPrimaries, &codedAspects.primaries)) { | 
|  | codedAspects.primaries = C2Color::PRIMARIES_UNSPECIFIED; | 
|  | } | 
|  | if (!C2Mapper::map(sfAspects.mRange, &codedAspects.range)) { | 
|  | codedAspects.range = C2Color::RANGE_UNSPECIFIED; | 
|  | } | 
|  | if (!C2Mapper::map(sfAspects.mMatrixCoeffs, &codedAspects.matrix)) { | 
|  | codedAspects.matrix = C2Color::MATRIX_UNSPECIFIED; | 
|  | } | 
|  | if (!C2Mapper::map(sfAspects.mTransfer, &codedAspects.transfer)) { | 
|  | codedAspects.transfer = C2Color::TRANSFER_UNSPECIFIED; | 
|  | } | 
|  | std::vector<std::unique_ptr<C2SettingResult>> failures; | 
|  | mIntf->config({&codedAspects}, C2_MAY_BLOCK, &failures); | 
|  | } | 
|  | } | 
|  |  | 
|  | void C2SoftGav1Dec::setError(const std::unique_ptr<C2Work> &work, c2_status_t error) { | 
|  | mSignalledError = true; | 
|  | work->result = error; | 
|  | work->workletsProcessed = 1u; | 
|  | } | 
|  |  | 
|  | bool C2SoftGav1Dec::allocTmpFrameBuffer(size_t size) { | 
|  | if (size > mTmpFrameBufferSize) { | 
|  | mTmpFrameBuffer = std::make_unique<uint16_t[]>(size); | 
|  | if (mTmpFrameBuffer == nullptr) { | 
|  | mTmpFrameBufferSize = 0; | 
|  | return false; | 
|  | } | 
|  | mTmpFrameBufferSize = size; | 
|  | } | 
|  | return true; | 
|  | } | 
|  |  | 
|  | bool C2SoftGav1Dec::outputBuffer(const std::shared_ptr<C2BlockPool> &pool, | 
|  | const std::unique_ptr<C2Work> &work) { | 
|  | if (!(work && pool)) return false; | 
|  |  | 
|  | const libgav1::DecoderBuffer *buffer; | 
|  | const Libgav1StatusCode status = mCodecCtx->DequeueFrame(&buffer); | 
|  |  | 
|  | if (status != kLibgav1StatusOk && status != kLibgav1StatusNothingToDequeue) { | 
|  | ALOGE("av1 decoder DequeueFrame failed. status: %d.", status); | 
|  | return false; | 
|  | } | 
|  |  | 
|  | // |buffer| can be NULL if status was equal to kLibgav1StatusOk or | 
|  | // kLibgav1StatusNothingToDequeue. This is not an error. This could mean one | 
|  | // of two things: | 
|  | //  - The EnqueueFrame() call was either a flush (called with nullptr). | 
|  | //  - The enqueued frame did not have any displayable frames. | 
|  | if (!buffer) { | 
|  | return false; | 
|  | } | 
|  |  | 
|  | const int width = buffer->displayed_width[0]; | 
|  | const int height = buffer->displayed_height[0]; | 
|  | if (width != mWidth || height != mHeight) { | 
|  | mWidth = width; | 
|  | mHeight = height; | 
|  |  | 
|  | C2StreamPictureSizeInfo::output size(0u, mWidth, mHeight); | 
|  | std::vector<std::unique_ptr<C2SettingResult>> failures; | 
|  | c2_status_t err = mIntf->config({&size}, C2_MAY_BLOCK, &failures); | 
|  | if (err == C2_OK) { | 
|  | work->worklets.front()->output.configUpdate.push_back( | 
|  | C2Param::Copy(size)); | 
|  | } else { | 
|  | ALOGE("Config update size failed"); | 
|  | mSignalledError = true; | 
|  | work->result = C2_CORRUPTED; | 
|  | work->workletsProcessed = 1u; | 
|  | return false; | 
|  | } | 
|  | } | 
|  |  | 
|  | getVuiParams(buffer); | 
|  | getHDRStaticParams(buffer, work); | 
|  | getHDR10PlusInfoData(buffer, work); | 
|  |  | 
|  | #if LIBYUV_VERSION < 1779 | 
|  | if (buffer->bitdepth == 10 && | 
|  | !(buffer->image_format == libgav1::kImageFormatYuv420 || | 
|  | buffer->image_format == libgav1::kImageFormatMonochrome400)) { | 
|  | ALOGE("image_format %d not supported for 10bit", buffer->image_format); | 
|  | mSignalledError = true; | 
|  | work->workletsProcessed = 1u; | 
|  | work->result = C2_CORRUPTED; | 
|  | return false; | 
|  | } | 
|  | #endif | 
|  |  | 
|  | const bool isMonochrome = | 
|  | buffer->image_format == libgav1::kImageFormatMonochrome400; | 
|  |  | 
|  | std::shared_ptr<C2GraphicBlock> block; | 
|  | uint32_t format = HAL_PIXEL_FORMAT_YV12; | 
|  | std::shared_ptr<C2StreamColorAspectsInfo::output> codedColorAspects; | 
|  | if (buffer->bitdepth == 10 && mPixelFormatInfo->value != HAL_PIXEL_FORMAT_YCBCR_420_888) { | 
|  | IntfImpl::Lock lock = mIntf->lock(); | 
|  | codedColorAspects = mIntf->getColorAspects_l(); | 
|  | bool allowRGBA1010102 = false; | 
|  | if (codedColorAspects->primaries == C2Color::PRIMARIES_BT2020 && | 
|  | codedColorAspects->matrix == C2Color::MATRIX_BT2020 && | 
|  | codedColorAspects->transfer == C2Color::TRANSFER_ST2084) { | 
|  | allowRGBA1010102 = true; | 
|  | } | 
|  | format = getHalPixelFormatForBitDepth10(allowRGBA1010102); | 
|  | #if !HAVE_LIBYUV_I410_I210_TO_AB30 | 
|  | if ((format == HAL_PIXEL_FORMAT_RGBA_1010102) && | 
|  | (buffer->image_format != libgav1::kImageFormatYuv420)) { | 
|  | ALOGE("Only YUV420 output is supported when targeting RGBA_1010102"); | 
|  | mSignalledError = true; | 
|  | work->result = C2_OMITTED; | 
|  | work->workletsProcessed = 1u; | 
|  | return false; | 
|  | } | 
|  | #endif | 
|  | } | 
|  |  | 
|  | if (mHalPixelFormat != format) { | 
|  | C2StreamPixelFormatInfo::output pixelFormat(0u, format); | 
|  | std::vector<std::unique_ptr<C2SettingResult>> failures; | 
|  | c2_status_t err = mIntf->config({&pixelFormat }, C2_MAY_BLOCK, &failures); | 
|  | if (err == C2_OK) { | 
|  | work->worklets.front()->output.configUpdate.push_back( | 
|  | C2Param::Copy(pixelFormat)); | 
|  | } else { | 
|  | ALOGE("Config update pixelFormat failed"); | 
|  | mSignalledError = true; | 
|  | work->workletsProcessed = 1u; | 
|  | work->result = C2_CORRUPTED; | 
|  | return UNKNOWN_ERROR; | 
|  | } | 
|  | mHalPixelFormat = format; | 
|  | } | 
|  |  | 
|  | C2MemoryUsage usage = {C2MemoryUsage::CPU_READ, C2MemoryUsage::CPU_WRITE}; | 
|  |  | 
|  | // We always create a graphic block that is width aligned to 16 and height | 
|  | // aligned to 2. We set the correct "crop" value of the image in the call to | 
|  | // createGraphicBuffer() by setting the correct image dimensions. | 
|  | c2_status_t err = pool->fetchGraphicBlock(align(mWidth, 16), | 
|  | align(mHeight, 2), format, usage, | 
|  | &block); | 
|  |  | 
|  | if (err != C2_OK) { | 
|  | ALOGE("fetchGraphicBlock for Output failed with status %d", err); | 
|  | work->result = err; | 
|  | return false; | 
|  | } | 
|  |  | 
|  | C2GraphicView wView = block->map().get(); | 
|  |  | 
|  | if (wView.error()) { | 
|  | ALOGE("graphic view map failed %d", wView.error()); | 
|  | work->result = C2_CORRUPTED; | 
|  | return false; | 
|  | } | 
|  |  | 
|  | ALOGV("provided (%dx%d) required (%dx%d), out frameindex %d", block->width(), | 
|  | block->height(), mWidth, mHeight, (int)buffer->user_private_data); | 
|  |  | 
|  | uint8_t *dstY = const_cast<uint8_t *>(wView.data()[C2PlanarLayout::PLANE_Y]); | 
|  | uint8_t *dstU = const_cast<uint8_t *>(wView.data()[C2PlanarLayout::PLANE_U]); | 
|  | uint8_t *dstV = const_cast<uint8_t *>(wView.data()[C2PlanarLayout::PLANE_V]); | 
|  |  | 
|  | C2PlanarLayout layout = wView.layout(); | 
|  | size_t dstYStride = layout.planes[C2PlanarLayout::PLANE_Y].rowInc; | 
|  | size_t dstUStride = layout.planes[C2PlanarLayout::PLANE_U].rowInc; | 
|  | size_t dstVStride = layout.planes[C2PlanarLayout::PLANE_V].rowInc; | 
|  |  | 
|  | if (buffer->bitdepth == 10) { | 
|  | const uint16_t *srcY = (const uint16_t *)buffer->plane[0]; | 
|  | const uint16_t *srcU = (const uint16_t *)buffer->plane[1]; | 
|  | const uint16_t *srcV = (const uint16_t *)buffer->plane[2]; | 
|  | size_t srcYStride = buffer->stride[0] / 2; | 
|  | size_t srcUStride = buffer->stride[1] / 2; | 
|  | size_t srcVStride = buffer->stride[2] / 2; | 
|  |  | 
|  | if (format == HAL_PIXEL_FORMAT_RGBA_1010102) { | 
|  | bool processed = false; | 
|  | #if HAVE_LIBYUV_I410_I210_TO_AB30 | 
|  | if (buffer->image_format == libgav1::kImageFormatYuv444) { | 
|  | libyuv::I410ToAB30Matrix(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride, | 
|  | dstY, dstYStride, &libyuv::kYuvV2020Constants, | 
|  | mWidth, mHeight); | 
|  | processed = true; | 
|  | } else if (buffer->image_format == libgav1::kImageFormatYuv422) { | 
|  | libyuv::I210ToAB30Matrix(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride, | 
|  | dstY, dstYStride, &libyuv::kYuvV2020Constants, | 
|  | mWidth, mHeight); | 
|  | processed = true; | 
|  | } | 
|  | #endif  // HAVE_LIBYUV_I410_I210_TO_AB30 | 
|  | if (!processed) { | 
|  | if (isMonochrome) { | 
|  | const size_t tmpSize = mWidth; | 
|  | const bool needFill = tmpSize > mTmpFrameBufferSize; | 
|  | if (!allocTmpFrameBuffer(tmpSize)) { | 
|  | ALOGE("Error allocating temp conversion buffer (%zu bytes)", tmpSize); | 
|  | setError(work, C2_NO_MEMORY); | 
|  | return false; | 
|  | } | 
|  | srcU = srcV = mTmpFrameBuffer.get(); | 
|  | srcUStride = srcVStride = 0; | 
|  | if (needFill) { | 
|  | std::fill_n(mTmpFrameBuffer.get(), tmpSize, 512); | 
|  | } | 
|  | } | 
|  | convertYUV420Planar16ToY410OrRGBA1010102( | 
|  | (uint32_t *)dstY, srcY, srcU, srcV, srcYStride, | 
|  | srcUStride, srcVStride, | 
|  | dstYStride / sizeof(uint32_t), mWidth, mHeight, | 
|  | std::static_pointer_cast<const C2ColorAspectsStruct>(codedColorAspects)); | 
|  | } | 
|  | } else if (format == HAL_PIXEL_FORMAT_YCBCR_P010) { | 
|  | dstYStride /= 2; | 
|  | dstUStride /= 2; | 
|  | dstVStride /= 2; | 
|  | #if LIBYUV_VERSION >= 1779 | 
|  | if (buffer->image_format == libgav1::kImageFormatYuv444 || | 
|  | buffer->image_format == libgav1::kImageFormatYuv422) { | 
|  | // TODO(https://crbug.com/libyuv/952): replace this block with libyuv::I410ToP010 and | 
|  | // libyuv::I210ToP010 when they are available. | 
|  | // Note it may be safe to alias dstY in I010ToP010, but the libyuv API doesn't make any | 
|  | // guarantees. | 
|  | const size_t tmpSize = dstYStride * mHeight + dstUStride * align(mHeight, 2); | 
|  | if (!allocTmpFrameBuffer(tmpSize)) { | 
|  | ALOGE("Error allocating temp conversion buffer (%zu bytes)", tmpSize); | 
|  | setError(work, C2_NO_MEMORY); | 
|  | return false; | 
|  | } | 
|  | uint16_t *const tmpY = mTmpFrameBuffer.get(); | 
|  | uint16_t *const tmpU = tmpY + dstYStride * mHeight; | 
|  | uint16_t *const tmpV = tmpU + dstUStride * align(mHeight, 2) / 2; | 
|  | if (buffer->image_format == libgav1::kImageFormatYuv444) { | 
|  | libyuv::I410ToI010(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride, | 
|  | tmpY, dstYStride, tmpU, dstUStride, tmpV, dstUStride, | 
|  | mWidth, mHeight); | 
|  | } else { | 
|  | libyuv::I210ToI010(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride, | 
|  | tmpY, dstYStride, tmpU, dstUStride, tmpV, dstUStride, | 
|  | mWidth, mHeight); | 
|  | } | 
|  | libyuv::I010ToP010(tmpY, dstYStride, tmpU, dstUStride, tmpV, dstVStride, | 
|  | (uint16_t*)dstY, dstYStride, (uint16_t*)dstU, dstUStride, | 
|  | mWidth, mHeight); | 
|  | } else { | 
|  | convertYUV420Planar16ToP010((uint16_t *)dstY, (uint16_t *)dstU, srcY, srcU, srcV, | 
|  | srcYStride, srcUStride, srcVStride, dstYStride, | 
|  | dstUStride, mWidth, mHeight, isMonochrome); | 
|  | } | 
|  | #else  // LIBYUV_VERSION < 1779 | 
|  | convertYUV420Planar16ToP010((uint16_t *)dstY, (uint16_t *)dstU, srcY, srcU, srcV, | 
|  | srcYStride, srcUStride, srcVStride, dstYStride, | 
|  | dstUStride, mWidth, mHeight, isMonochrome); | 
|  | #endif  // LIBYUV_VERSION >= 1779 | 
|  | } else { | 
|  | #if LIBYUV_VERSION >= 1779 | 
|  | if (buffer->image_format == libgav1::kImageFormatYuv444) { | 
|  | // TODO(https://crbug.com/libyuv/950): replace this block with libyuv::I410ToI420 when | 
|  | // it's available. | 
|  | const size_t tmpSize = dstYStride * mHeight + dstUStride * align(mHeight, 2); | 
|  | if (!allocTmpFrameBuffer(tmpSize)) { | 
|  | ALOGE("Error allocating temp conversion buffer (%zu bytes)", tmpSize); | 
|  | setError(work, C2_NO_MEMORY); | 
|  | return false; | 
|  | } | 
|  | uint16_t *const tmpY = mTmpFrameBuffer.get(); | 
|  | uint16_t *const tmpU = tmpY + dstYStride * mHeight; | 
|  | uint16_t *const tmpV = tmpU + dstUStride * align(mHeight, 2) / 2; | 
|  | libyuv::I410ToI010(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride, | 
|  | tmpY, dstYStride, tmpU, dstUStride, tmpV, dstVStride, | 
|  | mWidth, mHeight); | 
|  | libyuv::I010ToI420(tmpY, dstYStride, tmpU, dstUStride, tmpV, dstUStride, | 
|  | dstY, dstYStride, dstU, dstUStride, dstV, dstVStride, | 
|  | mWidth, mHeight); | 
|  | } else if (buffer->image_format == libgav1::kImageFormatYuv422) { | 
|  | libyuv::I210ToI420(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride, | 
|  | dstY, dstYStride, dstU, dstUStride, dstV, dstVStride, | 
|  | mWidth, mHeight); | 
|  | } else { | 
|  | convertYUV420Planar16ToYV12(dstY, dstU, dstV, srcY, srcU, srcV, srcYStride, | 
|  | srcUStride, srcVStride, dstYStride, dstUStride, | 
|  | mWidth, mHeight, isMonochrome); | 
|  | } | 
|  | #else  // LIBYUV_VERSION < 1779 | 
|  | convertYUV420Planar16ToYV12(dstY, dstU, dstV, srcY, srcU, srcV, srcYStride, | 
|  | srcUStride, srcVStride, dstYStride, dstUStride, | 
|  | mWidth, mHeight, isMonochrome); | 
|  | #endif  // LIBYUV_VERSION >= 1779 | 
|  | } | 
|  | } else { | 
|  | const uint8_t *srcY = (const uint8_t *)buffer->plane[0]; | 
|  | const uint8_t *srcU = (const uint8_t *)buffer->plane[1]; | 
|  | const uint8_t *srcV = (const uint8_t *)buffer->plane[2]; | 
|  | size_t srcYStride = buffer->stride[0]; | 
|  | size_t srcUStride = buffer->stride[1]; | 
|  | size_t srcVStride = buffer->stride[2]; | 
|  |  | 
|  | if (buffer->image_format == libgav1::kImageFormatYuv444) { | 
|  | libyuv::I444ToI420(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride, | 
|  | dstY, dstYStride, dstU, dstUStride, dstV, dstVStride, | 
|  | mWidth, mHeight); | 
|  | } else if (buffer->image_format == libgav1::kImageFormatYuv422) { | 
|  | libyuv::I422ToI420(srcY, srcYStride, srcU, srcUStride, srcV, srcVStride, | 
|  | dstY, dstYStride, dstU, dstUStride, dstV, dstVStride, | 
|  | mWidth, mHeight); | 
|  | } else { | 
|  | convertYUV420Planar8ToYV12(dstY, dstU, dstV, srcY, srcU, srcV, srcYStride, srcUStride, | 
|  | srcVStride, dstYStride, dstUStride, dstVStride, mWidth, mHeight, | 
|  | isMonochrome); | 
|  | } | 
|  | } | 
|  | finishWork(buffer->user_private_data, work, std::move(block)); | 
|  | block = nullptr; | 
|  | return true; | 
|  | } | 
|  |  | 
|  | c2_status_t C2SoftGav1Dec::drainInternal( | 
|  | uint32_t drainMode, const std::shared_ptr<C2BlockPool> &pool, | 
|  | const std::unique_ptr<C2Work> &work) { | 
|  | if (drainMode == NO_DRAIN) { | 
|  | ALOGW("drain with NO_DRAIN: no-op"); | 
|  | return C2_OK; | 
|  | } | 
|  | if (drainMode == DRAIN_CHAIN) { | 
|  | ALOGW("DRAIN_CHAIN not supported"); | 
|  | return C2_OMITTED; | 
|  | } | 
|  |  | 
|  | const Libgav1StatusCode status = mCodecCtx->SignalEOS(); | 
|  | if (status != kLibgav1StatusOk) { | 
|  | ALOGE("Failed to flush av1 decoder. status: %d.", status); | 
|  | return C2_CORRUPTED; | 
|  | } | 
|  |  | 
|  | while (outputBuffer(pool, work)) { | 
|  | } | 
|  |  | 
|  | if (drainMode == DRAIN_COMPONENT_WITH_EOS && work && | 
|  | work->workletsProcessed == 0u) { | 
|  | fillEmptyWork(work); | 
|  | } | 
|  |  | 
|  | return C2_OK; | 
|  | } | 
|  |  | 
|  | c2_status_t C2SoftGav1Dec::drain(uint32_t drainMode, | 
|  | const std::shared_ptr<C2BlockPool> &pool) { | 
|  | return drainInternal(drainMode, pool, nullptr); | 
|  | } | 
|  |  | 
|  | class C2SoftGav1Factory : public C2ComponentFactory { | 
|  | public: | 
|  | C2SoftGav1Factory() | 
|  | : mHelper(std::static_pointer_cast<C2ReflectorHelper>( | 
|  | GetCodec2PlatformComponentStore()->getParamReflector())) {} | 
|  |  | 
|  | virtual c2_status_t createComponent( | 
|  | c2_node_id_t id, std::shared_ptr<C2Component> *const component, | 
|  | std::function<void(C2Component *)> deleter) override { | 
|  | *component = std::shared_ptr<C2Component>( | 
|  | new C2SoftGav1Dec(COMPONENT_NAME, id, | 
|  | std::make_shared<C2SoftGav1Dec::IntfImpl>(mHelper)), | 
|  | deleter); | 
|  | return C2_OK; | 
|  | } | 
|  |  | 
|  | virtual c2_status_t createInterface( | 
|  | c2_node_id_t id, std::shared_ptr<C2ComponentInterface> *const interface, | 
|  | std::function<void(C2ComponentInterface *)> deleter) override { | 
|  | *interface = std::shared_ptr<C2ComponentInterface>( | 
|  | new SimpleInterface<C2SoftGav1Dec::IntfImpl>( | 
|  | COMPONENT_NAME, id, | 
|  | std::make_shared<C2SoftGav1Dec::IntfImpl>(mHelper)), | 
|  | deleter); | 
|  | return C2_OK; | 
|  | } | 
|  |  | 
|  | virtual ~C2SoftGav1Factory() override = default; | 
|  |  | 
|  | private: | 
|  | std::shared_ptr<C2ReflectorHelper> mHelper; | 
|  | }; | 
|  |  | 
|  | }  // namespace android | 
|  |  | 
|  | __attribute__((cfi_canonical_jump_table)) | 
|  | extern "C" ::C2ComponentFactory *CreateCodec2Factory() { | 
|  | ALOGV("in %s", __func__); | 
|  | return new ::android::C2SoftGav1Factory(); | 
|  | } | 
|  |  | 
|  | __attribute__((cfi_canonical_jump_table)) | 
|  | extern "C" void DestroyCodec2Factory(::C2ComponentFactory *factory) { | 
|  | ALOGV("in %s", __func__); | 
|  | delete factory; | 
|  | } |