Merge "APM: Notify the client about new audio ports" into rvc-dev
diff --git a/camera/ndk/include/camera/NdkCameraMetadataTags.h b/camera/ndk/include/camera/NdkCameraMetadataTags.h
index 4f9b0d1..16457ac 100644
--- a/camera/ndk/include/camera/NdkCameraMetadataTags.h
+++ b/camera/ndk/include/camera/NdkCameraMetadataTags.h
@@ -1895,27 +1895,66 @@
* ACAMERA_SCALER_CROP_REGION can still be used to specify the horizontal or vertical
* crop to achieve aspect ratios different than the native camera sensor.</p>
* <p>By using this control, the application gains a simpler way to control zoom, which can
- * be a combination of optical and digital zoom. More specifically, for a logical
- * multi-camera with more than one focal length, using a floating point zoom ratio offers
- * more zoom precision when a telephoto lens is used, as well as allowing zoom ratio of
- * less than 1.0 to zoom out to a wide field of view.</p>
- * <p>Note that the coordinate system of cropRegion, AE/AWB/AF regions, and faces now changes
- * to the effective after-zoom field-of-view represented by rectangle of (0, 0,
- * activeArrayWidth, activeArrayHeight).</p>
- * <p>For example, if ACAMERA_SENSOR_INFO_ACTIVE_ARRAY_SIZE is 4032*3024, and the preview stream
- * is configured to the same 4:3 aspect ratio, the application can achieve 2.0x zoom in
- * one of two ways:</p>
+ * be a combination of optical and digital zoom. For example, a multi-camera system may
+ * contain more than one lens with different focal lengths, and the user can use optical
+ * zoom by switching between lenses. Using zoomRatio has benefits in the scenarios below:
+ * <em> Zooming in from a wide-angle lens to a telephoto lens: A floating-point ratio provides
+ * better precision compared to an integer value of ACAMERA_SCALER_CROP_REGION.
+ * </em> Zooming out from a wide lens to an ultrawide lens: zoomRatio supports zoom-out whereas
+ * ACAMERA_SCALER_CROP_REGION doesn't.</p>
+ * <p>To illustrate, here are several scenarios of different zoom ratios, crop regions,
+ * and output streams, for a hypothetical camera device with an active array of size
+ * <code>(2000,1500)</code>.</p>
* <ul>
- * <li>zoomRatio = 2.0, scaler.cropRegion = (0, 0, 4032, 3024)</li>
- * <li>zoomRatio = 1.0 (default), scaler.cropRegion = (1008, 756, 3024, 2268)</li>
+ * <li>Camera Configuration:<ul>
+ * <li>Active array size: <code>2000x1500</code> (3 MP, 4:3 aspect ratio)</li>
+ * <li>Output stream #1: <code>640x480</code> (VGA, 4:3 aspect ratio)</li>
+ * <li>Output stream #2: <code>1280x720</code> (720p, 16:9 aspect ratio)</li>
* </ul>
- * <p>If the application intends to set aeRegions to be top-left quarter of the preview
- * field-of-view, the ACAMERA_CONTROL_AE_REGIONS should be set to (0, 0, 2016, 1512) with
+ * </li>
+ * <li>Case #1: 4:3 crop region with 2.0x zoom ratio<ul>
+ * <li>Zoomed field of view: 1/4 of original field of view</li>
+ * <li>Crop region: <code>Rect(0, 0, 2000, 1500) // (left, top, right, bottom)</code> (post zoom)</li>
+ * </ul>
+ * </li>
+ * <li><img alt="4:3 aspect ratio crop diagram" src="../images/camera2/metadata/android.control.zoomRatio/zoom-ratio-2-crop-43.png" /><ul>
+ * <li><code>640x480</code> stream source area: <code>(0, 0, 2000, 1500)</code> (equal to crop region)</li>
+ * <li><code>1280x720</code> stream source area: <code>(0, 187, 2000, 1312)</code> (letterboxed)</li>
+ * </ul>
+ * </li>
+ * <li>Case #2: 16:9 crop region with 2.0x zoom.<ul>
+ * <li>Zoomed field of view: 1/4 of original field of view</li>
+ * <li>Crop region: <code>Rect(0, 187, 2000, 1312)</code></li>
+ * <li><img alt="16:9 aspect ratio crop diagram" src="../images/camera2/metadata/android.control.zoomRatio/zoom-ratio-2-crop-169.png" /></li>
+ * <li><code>640x480</code> stream source area: <code>(250, 187, 1750, 1312)</code> (pillarboxed)</li>
+ * <li><code>1280x720</code> stream source area: <code>(0, 187, 2000, 1312)</code> (equal to crop region)</li>
+ * </ul>
+ * </li>
+ * <li>Case #3: 1:1 crop region with 0.5x zoom out to ultrawide lens.<ul>
+ * <li>Zoomed field of view: 4x of original field of view (switched from wide lens to ultrawide lens)</li>
+ * <li>Crop region: <code>Rect(250, 0, 1750, 1500)</code></li>
+ * <li><img alt="1:1 aspect ratio crop diagram" src="../images/camera2/metadata/android.control.zoomRatio/zoom-ratio-0.5-crop-11.png" /></li>
+ * <li><code>640x480</code> stream source area: <code>(250, 187, 1750, 1312)</code> (letterboxed)</li>
+ * <li><code>1280x720</code> stream source area: <code>(250, 328, 1750, 1172)</code> (letterboxed)</li>
+ * </ul>
+ * </li>
+ * </ul>
+ * <p>As seen from the graphs above, the coordinate system of cropRegion now changes to the
+ * effective after-zoom field-of-view, and is represented by the rectangle of (0, 0,
+ * activeArrayWith, activeArrayHeight). The same applies to AE/AWB/AF regions, and faces.
+ * This coordinate system change isn't applicable to RAW capture and its related
+ * metadata such as intrinsicCalibration and lensShadingMap.</p>
+ * <p>Using the same hypothetical example above, and assuming output stream #1 (640x480) is
+ * the viewfinder stream, the application can achieve 2.0x zoom in one of two ways:</p>
+ * <ul>
+ * <li>zoomRatio = 2.0, scaler.cropRegion = (0, 0, 2000, 1500)</li>
+ * <li>zoomRatio = 1.0 (default), scaler.cropRegion = (500, 375, 1500, 1125)</li>
+ * </ul>
+ * <p>If the application intends to set aeRegions to be top-left quarter of the viewfinder
+ * field-of-view, the ACAMERA_CONTROL_AE_REGIONS should be set to (0, 0, 1000, 750) with
* zoomRatio set to 2.0. Alternatively, the application can set aeRegions to the equivalent
- * region of (1008, 756, 2016, 1512) for zoomRatio of 1.0. If the application doesn't
+ * region of (500, 375, 1000, 750) for zoomRatio of 1.0. If the application doesn't
* explicitly set ACAMERA_CONTROL_ZOOM_RATIO, its value defaults to 1.0.</p>
- * <p>This coordinate system change isn't applicable to RAW capture and its related metadata
- * such as intrinsicCalibration and lensShadingMap.</p>
* <p>One limitation of controlling zoom using zoomRatio is that the ACAMERA_SCALER_CROP_REGION
* must only be used for letterboxing or pillarboxing of the sensor active array, and no
* FREEFORM cropping can be used with ACAMERA_CONTROL_ZOOM_RATIO other than 1.0.</p>
@@ -1923,7 +1962,6 @@
* @see ACAMERA_CONTROL_AE_REGIONS
* @see ACAMERA_CONTROL_ZOOM_RATIO
* @see ACAMERA_SCALER_CROP_REGION
- * @see ACAMERA_SENSOR_INFO_ACTIVE_ARRAY_SIZE
*/
ACAMERA_CONTROL_ZOOM_RATIO = // float
ACAMERA_CONTROL_START + 47,
@@ -2395,8 +2433,11 @@
* frames before the lens can change to the requested focal length.
* While the focal length is still changing, ACAMERA_LENS_STATE will
* be set to MOVING.</p>
- * <p>Optical zoom will not be supported on most devices.</p>
+ * <p>Optical zoom via this control will not be supported on most devices. Starting from API
+ * level 30, the camera device may combine optical and digital zoom through the
+ * ACAMERA_CONTROL_ZOOM_RATIO control.</p>
*
+ * @see ACAMERA_CONTROL_ZOOM_RATIO
* @see ACAMERA_LENS_APERTURE
* @see ACAMERA_LENS_FOCUS_DISTANCE
* @see ACAMERA_LENS_STATE
diff --git a/media/codec2/components/hevc/C2SoftHevcEnc.cpp b/media/codec2/components/hevc/C2SoftHevcEnc.cpp
index 19ccbf9..c2d2540 100644
--- a/media/codec2/components/hevc/C2SoftHevcEnc.cpp
+++ b/media/codec2/components/hevc/C2SoftHevcEnc.cpp
@@ -628,6 +628,7 @@
mComplexity = mIntf->getComplexity_l();
mQuality = mIntf->getQuality_l();
mGop = mIntf->getGop_l();
+ mRequestSync = mIntf->getRequestSync_l();
}
c2_status_t status = initEncParams();
@@ -956,7 +957,7 @@
}
}
- // handle dynamic config parameters
+ // handle dynamic bitrate change
{
IntfImpl::Lock lock = mIntf->lock();
std::shared_ptr<C2StreamBitrateInfo::output> bitrate = mIntf->getBitrate_l();
@@ -983,6 +984,26 @@
work->workletsProcessed = 1u;
return;
}
+ // handle request key frame
+ {
+ IntfImpl::Lock lock = mIntf->lock();
+ std::shared_ptr<C2StreamRequestSyncFrameTuning::output> requestSync;
+ requestSync = mIntf->getRequestSync_l();
+ lock.unlock();
+ if (requestSync != mRequestSync) {
+ // we can handle IDR immediately
+ if (requestSync->value) {
+ // unset request
+ C2StreamRequestSyncFrameTuning::output clearSync(0u, C2_FALSE);
+ std::vector<std::unique_ptr<C2SettingResult>> failures;
+ mIntf->config({ &clearSync }, C2_MAY_BLOCK, &failures);
+ ALOGV("Got sync request");
+ //Force this as an IDR frame
+ s_encode_ip.i4_force_idr_flag = 1;
+ }
+ mRequestSync = requestSync;
+ }
+ }
uint64_t timeDelay = 0;
uint64_t timeTaken = 0;
diff --git a/media/codec2/components/hevc/C2SoftHevcEnc.h b/media/codec2/components/hevc/C2SoftHevcEnc.h
index 140b4a9..5ea4602 100644
--- a/media/codec2/components/hevc/C2SoftHevcEnc.h
+++ b/media/codec2/components/hevc/C2SoftHevcEnc.h
@@ -88,6 +88,7 @@
std::shared_ptr<C2StreamComplexityTuning::output> mComplexity;
std::shared_ptr<C2StreamQualityTuning::output> mQuality;
std::shared_ptr<C2StreamGopTuning::output> mGop;
+ std::shared_ptr<C2StreamRequestSyncFrameTuning::output> mRequestSync;
#ifdef FILE_DUMP_ENABLE
char mInFile[200];
char mOutFile[200];
diff --git a/media/codec2/components/vpx/C2SoftVpxDec.cpp b/media/codec2/components/vpx/C2SoftVpxDec.cpp
index c7d73f4..3eef1e3 100644
--- a/media/codec2/components/vpx/C2SoftVpxDec.cpp
+++ b/media/codec2/components/vpx/C2SoftVpxDec.cpp
@@ -784,7 +784,13 @@
}
}
- CHECK(img->fmt == VPX_IMG_FMT_I420 || img->fmt == VPX_IMG_FMT_I42016);
+ if(img->fmt != VPX_IMG_FMT_I420 && img->fmt != VPX_IMG_FMT_I42016) {
+ ALOGE("img->fmt %d not supported", img->fmt);
+ mSignalledError = true;
+ work->workletsProcessed = 1u;
+ work->result = C2_CORRUPTED;
+ return false;
+ }
std::shared_ptr<C2GraphicBlock> block;
uint32_t format = HAL_PIXEL_FORMAT_YV12;
diff --git a/media/extractors/mp4/MPEG4Extractor.cpp b/media/extractors/mp4/MPEG4Extractor.cpp
index ce82861..c493594 100755
--- a/media/extractors/mp4/MPEG4Extractor.cpp
+++ b/media/extractors/mp4/MPEG4Extractor.cpp
@@ -19,6 +19,7 @@
#include <ctype.h>
#include <inttypes.h>
+#include <algorithm>
#include <memory>
#include <stdint.h>
#include <stdlib.h>
@@ -149,9 +150,13 @@
bool mIsAudio;
sp<ItemTable> mItemTable;
- // Start offset from composition time to presentation time.
- // Support shift only for video tracks through mElstShiftStartTicks for now.
+ /* Shift start offset (move to earlier time) when media_time > 0,
+ * in media time scale.
+ */
uint64_t mElstShiftStartTicks;
+ /* Initial start offset (move to later time), empty edit list entry
+ * in media time scale.
+ */
uint64_t mElstInitialEmptyEditTicks;
size_t parseNALSize(const uint8_t *data) const;
@@ -160,8 +165,9 @@
status_t parseTrackFragmentRun(off64_t offset, off64_t size);
status_t parseSampleAuxiliaryInformationSizes(off64_t offset, off64_t size);
status_t parseSampleAuxiliaryInformationOffsets(off64_t offset, off64_t size);
- status_t parseClearEncryptedSizes(off64_t offset, bool isSubsampleEncryption, uint32_t flags);
- status_t parseSampleEncryption(off64_t offset);
+ status_t parseClearEncryptedSizes(
+ off64_t offset, bool isSubsampleEncryption, uint32_t flags, off64_t size);
+ status_t parseSampleEncryption(off64_t offset, off64_t size);
// returns -1 for invalid layer ID
int32_t parseHEVCLayerId(const uint8_t *data, size_t size);
@@ -1215,7 +1221,6 @@
off64_t entriesoffset = data_offset + 8;
uint64_t segment_duration;
int64_t media_time;
- uint64_t empty_edit_ticks = 0;
bool empty_edit_present = false;
for (int i = 0; i < entry_count; ++i) {
switch (version) {
@@ -1247,45 +1252,37 @@
}
// Empty edit entry would have to be first entry.
if (media_time == -1 && i == 0) {
- int64_t durationUs;
- if (AMediaFormat_getInt64(mFileMetaData, AMEDIAFORMAT_KEY_DURATION,
- &durationUs)) {
- empty_edit_ticks = segment_duration;
- ALOGV("initial empty edit ticks: %" PRIu64, empty_edit_ticks);
- empty_edit_present = true;
- }
- }
- // Process second entry only when the first entry was an empty edit entry.
- if (empty_edit_present && i == 1) {
- int64_t durationUs;
- if (AMediaFormat_getInt64(mLastTrack->meta, AMEDIAFORMAT_KEY_DURATION,
- &durationUs) &&
- mHeaderTimescale != 0) {
- // Support only segment_duration<=track_duration and media_time==0 case.
- uint64_t segmentDurationUs =
- segment_duration * 1000000 / mHeaderTimescale;
- if (segmentDurationUs == 0 || segmentDurationUs > durationUs ||
- media_time != 0) {
- ALOGW("for now, unsupported second entry in empty edit list");
- }
- }
+ empty_edit_present = true;
+ ALOGV("initial empty edit ticks: %" PRIu64, segment_duration);
+ /* In movie header timescale, and needs to be converted to media timescale
+ * after we get that from a track's 'mdhd' atom,
+ * which at times come after 'elst'.
+ */
+ mLastTrack->elst_initial_empty_edit_ticks = segment_duration;
+ } else if (media_time >= 0 && i == 0) {
+ ALOGV("first edit list entry");
+ mLastTrack->elst_media_time = media_time;
+ mLastTrack->elst_segment_duration = segment_duration;
+ ALOGV("segment_duration: %" PRIu64 " media_time: %" PRId64,
+ segment_duration, media_time);
+ // media_time is in media timescale as are STTS/CTTS entries.
+ mLastTrack->elst_shift_start_ticks = media_time;
+ } else if (empty_edit_present && i == 1) {
+ // Process second entry only when the first entry was an empty edit entry.
+ ALOGV("second edit list entry");
+ mLastTrack->elst_media_time = media_time;
+ mLastTrack->elst_segment_duration = segment_duration;
+ ALOGV("segment_duration: %" PRIu64 " media_time: %" PRId64,
+ segment_duration, media_time);
+ mLastTrack->elst_shift_start_ticks = media_time;
+ } else {
+ ALOGW("for now, unsupported entry in edit list %" PRIu32, entry_count);
}
}
// save these for later, because the elst atom might precede
// the atoms that actually gives us the duration and sample rate
// needed to calculate the padding and delay values
mLastTrack->elst_needs_processing = true;
- if (empty_edit_present) {
- /* In movie header timescale, and needs to be converted to media timescale once
- * we get that from a track's 'mdhd' atom, which at times come after 'elst'.
- */
- mLastTrack->elst_initial_empty_edit_ticks = empty_edit_ticks;
- } else {
- mLastTrack->elst_media_time = media_time;
- mLastTrack->elst_segment_duration = segment_duration;
- ALOGV("segment_duration: %" PRIu64 " media_time: %" PRId64, segment_duration,
- media_time);
- }
}
break;
}
@@ -4324,9 +4321,9 @@
}
}
- // media_time is in media timescale as are STTS/CTTS entries.
- track->elst_shift_start_ticks = track->elst_media_time;
ALOGV("track->elst_shift_start_ticks :%" PRIu64, track->elst_shift_start_ticks);
+
+ uint64_t elst_initial_empty_edit_ticks = 0;
if (mHeaderTimescale != 0) {
// Convert empty_edit_ticks from movie timescale to media timescale.
uint64_t elst_initial_empty_edit_ticks_mul = 0, elst_initial_empty_edit_ticks_add = 0;
@@ -4337,15 +4334,15 @@
ALOGE("track->elst_initial_empty_edit_ticks overflow");
return nullptr;
}
- track->elst_initial_empty_edit_ticks = elst_initial_empty_edit_ticks_add / mHeaderTimescale;
- ALOGV("track->elst_initial_empty_edit_ticks :%" PRIu64,
- track->elst_initial_empty_edit_ticks);
+ elst_initial_empty_edit_ticks = elst_initial_empty_edit_ticks_add / mHeaderTimescale;
}
+ ALOGV("elst_initial_empty_edit_ticks in MediaTimeScale :%" PRIu64,
+ elst_initial_empty_edit_ticks);
MPEG4Source* source =
new MPEG4Source(track->meta, mDataSource, track->timescale, track->sampleTable,
mSidxEntries, trex, mMoofOffset, itemTable,
- track->elst_shift_start_ticks, track->elst_initial_empty_edit_ticks);
+ track->elst_shift_start_ticks, elst_initial_empty_edit_ticks);
if (source->init() != OK) {
delete source;
return NULL;
@@ -5188,7 +5185,7 @@
case FOURCC("senc"): {
status_t err;
- if ((err = parseSampleEncryption(data_offset)) != OK) {
+ if ((err = parseSampleEncryption(data_offset, chunk_data_size)) != OK) {
return err;
}
*offset += chunk_size;
@@ -5380,12 +5377,13 @@
off64_t drmoffset = mCurrentSampleInfoOffsets[0]; // from moof
drmoffset += mCurrentMoofOffset;
+ size -= mCurrentMoofOffset;
- return parseClearEncryptedSizes(drmoffset, false, 0);
+ return parseClearEncryptedSizes(drmoffset, false, 0, size);
}
status_t MPEG4Source::parseClearEncryptedSizes(
- off64_t offset, bool isSubsampleEncryption, uint32_t flags) {
+ off64_t offset, bool isSubsampleEncryption, uint32_t flags, off64_t size) {
int32_t ivlength;
if (!AMediaFormat_getInt32(mFormat, AMEDIAFORMAT_KEY_CRYPTO_DEFAULT_IV_SIZE, &ivlength)) {
@@ -5400,10 +5398,14 @@
uint32_t sampleCount = mCurrentSampleInfoCount;
if (isSubsampleEncryption) {
+ if(size < 4){
+ return ERROR_MALFORMED;
+ }
if (!mDataSource->getUInt32(offset, &sampleCount)) {
return ERROR_IO;
}
offset += 4;
+ size -= 4;
}
// read CencSampleAuxiliaryDataFormats
@@ -5418,11 +5420,15 @@
}
memset(smpl->iv, 0, 16);
+ if(size < ivlength){
+ return ERROR_MALFORMED;
+ }
if (mDataSource->readAt(offset, smpl->iv, ivlength) != ivlength) {
return ERROR_IO;
}
offset += ivlength;
+ size -= ivlength;
bool readSubsamples;
if (isSubsampleEncryption) {
@@ -5437,13 +5443,20 @@
if (readSubsamples) {
uint16_t numsubsamples;
+ if(size < 2){
+ return ERROR_MALFORMED;
+ }
if (!mDataSource->getUInt16(offset, &numsubsamples)) {
return ERROR_IO;
}
offset += 2;
+ size -= 2;
for (size_t j = 0; j < numsubsamples; j++) {
uint16_t numclear;
uint32_t numencrypted;
+ if(size < 6){
+ return ERROR_MALFORMED;
+ }
if (!mDataSource->getUInt16(offset, &numclear)) {
return ERROR_IO;
}
@@ -5452,6 +5465,7 @@
return ERROR_IO;
}
offset += 4;
+ size -= 6;
smpl->clearsizes.add(numclear);
smpl->encryptedsizes.add(numencrypted);
}
@@ -5464,12 +5478,15 @@
return OK;
}
-status_t MPEG4Source::parseSampleEncryption(off64_t offset) {
+status_t MPEG4Source::parseSampleEncryption(off64_t offset, off64_t chunk_data_size) {
uint32_t flags;
+ if(chunk_data_size < 4) {
+ return ERROR_MALFORMED;
+ }
if (!mDataSource->getUInt32(offset, &flags)) { // actually version + flags
return ERROR_MALFORMED;
}
- return parseClearEncryptedSizes(offset + 4, true, flags);
+ return parseClearEncryptedSizes(offset + 4, true, flags, chunk_data_size - 4);
}
status_t MPEG4Source::parseTrackFragmentHeader(off64_t offset, off64_t size) {
@@ -5885,9 +5902,22 @@
break;
}
if( mode != ReadOptions::SEEK_FRAME_INDEX) {
- seekTimeUs += ((long double)mElstShiftStartTicks * 1000000) / mTimescale;
- ALOGV("shifted seekTimeUs :%" PRId64 ", mElstShiftStartTicks:%" PRIu64, seekTimeUs,
- mElstShiftStartTicks);
+ int64_t elstInitialEmptyEditUs = 0, elstShiftStartUs = 0;
+ if (mElstInitialEmptyEditTicks > 0) {
+ elstInitialEmptyEditUs = ((long double)mElstInitialEmptyEditTicks * 1000000) /
+ mTimescale;
+ /* Sample's composition time from ctts/stts entries are non-negative(>=0).
+ * Hence, lower bound on seekTimeUs is 0.
+ */
+ seekTimeUs = std::max(seekTimeUs - elstInitialEmptyEditUs, (int64_t)0);
+ }
+ if (mElstShiftStartTicks > 0) {
+ elstShiftStartUs = ((long double)mElstShiftStartTicks * 1000000) / mTimescale;
+ seekTimeUs += elstShiftStartUs;
+ }
+ ALOGV("shifted seekTimeUs:%" PRId64 ", elstInitialEmptyEditUs:%" PRIu64
+ ", elstShiftStartUs:%" PRIu64, seekTimeUs, elstInitialEmptyEditUs,
+ elstShiftStartUs);
}
uint32_t sampleIndex;
@@ -5933,7 +5963,12 @@
if (mode == ReadOptions::SEEK_CLOSEST
|| mode == ReadOptions::SEEK_FRAME_INDEX) {
- sampleTime -= mElstShiftStartTicks;
+ if (mElstInitialEmptyEditTicks > 0) {
+ sampleTime += mElstInitialEmptyEditTicks;
+ }
+ if (mElstShiftStartTicks > 0){
+ sampleTime -= mElstShiftStartTicks;
+ }
targetSampleTimeUs = (sampleTime * 1000000ll) / mTimescale;
}
@@ -5976,12 +6011,12 @@
if(err == OK) {
if (mElstInitialEmptyEditTicks > 0) {
cts += mElstInitialEmptyEditTicks;
- } else {
+ }
+ if (mElstShiftStartTicks > 0) {
// cts can be negative. for example, initial audio samples for gapless playback.
cts -= (int64_t)mElstShiftStartTicks;
}
}
-
} else {
err = mItemTable->getImageOffsetAndSize(
options && options->getSeekTo(&seekTimeUs, &mode) ?
@@ -6261,10 +6296,22 @@
int64_t seekTimeUs;
ReadOptions::SeekMode mode;
if (options && options->getSeekTo(&seekTimeUs, &mode)) {
-
- seekTimeUs += ((long double)mElstShiftStartTicks * 1000000) / mTimescale;
- ALOGV("shifted seekTimeUs :%" PRId64 ", mElstShiftStartTicks:%" PRIu64, seekTimeUs,
- mElstShiftStartTicks);
+ int64_t elstInitialEmptyEditUs = 0, elstShiftStartUs = 0;
+ if (mElstInitialEmptyEditTicks > 0) {
+ elstInitialEmptyEditUs = ((long double)mElstInitialEmptyEditTicks * 1000000) /
+ mTimescale;
+ /* Sample's composition time from ctts/stts entries are non-negative(>=0).
+ * Hence, lower bound on seekTimeUs is 0.
+ */
+ seekTimeUs = std::max(seekTimeUs - elstInitialEmptyEditUs, (int64_t)0);
+ }
+ if (mElstShiftStartTicks > 0){
+ elstShiftStartUs = ((long double)mElstShiftStartTicks * 1000000) / mTimescale;
+ seekTimeUs += elstShiftStartUs;
+ }
+ ALOGV("shifted seekTimeUs:%" PRId64 ", elstInitialEmptyEditUs:%" PRIu64
+ ", elstShiftStartUs:%" PRIu64, seekTimeUs, elstInitialEmptyEditUs,
+ elstShiftStartUs);
int numSidxEntries = mSegments.size();
if (numSidxEntries != 0) {
@@ -6355,7 +6402,8 @@
if (mElstInitialEmptyEditTicks > 0) {
cts += mElstInitialEmptyEditTicks;
- } else {
+ }
+ if (mElstShiftStartTicks > 0) {
// cts can be negative. for example, initial audio samples for gapless playback.
cts -= (int64_t)mElstShiftStartTicks;
}
diff --git a/media/extractors/mp4/MPEG4Extractor.h b/media/extractors/mp4/MPEG4Extractor.h
index 53ec6bc..3af432d 100644
--- a/media/extractors/mp4/MPEG4Extractor.h
+++ b/media/extractors/mp4/MPEG4Extractor.h
@@ -88,9 +88,9 @@
*/
int64_t elst_media_time;
uint64_t elst_segment_duration;
- // Shift start offset only when media_time > 0.
+ // Shift start offset (move to earlier time) when media_time > 0.
uint64_t elst_shift_start_ticks;
- // Initial start offset, empty edit list entry.
+ // Initial start offset (move to later time), from empty edit list entry.
uint64_t elst_initial_empty_edit_ticks;
bool subsample_encryption;
diff --git a/media/libaudioclient/AudioEffect.cpp b/media/libaudioclient/AudioEffect.cpp
index c183ab0..3ead6cb 100644
--- a/media/libaudioclient/AudioEffect.cpp
+++ b/media/libaudioclient/AudioEffect.cpp
@@ -151,11 +151,13 @@
// audio flinger will not be retained. initCheck() will return the creation status
// but all other APIs will return invalid operation.
if (probe || iEffect == 0 || (mStatus != NO_ERROR && mStatus != ALREADY_EXISTS)) {
- char typeBuffer[64], uuidBuffer[64];
+ char typeBuffer[64] = {}, uuidBuffer[64] = {};
guidToString(type, typeBuffer, sizeof(typeBuffer));
guidToString(uuid, uuidBuffer, sizeof(uuidBuffer));
ALOGE_IF(!probe, "set(): AudioFlinger could not create effect %s / %s, status: %d",
- typeBuffer, uuidBuffer, mStatus);
+ type != nullptr ? typeBuffer : "NULL",
+ uuid != nullptr ? uuidBuffer : "NULL",
+ mStatus);
if (!probe && iEffect == 0) {
mStatus = NO_INIT;
}
diff --git a/media/libaudioclient/AudioSystem.cpp b/media/libaudioclient/AudioSystem.cpp
index f030ab0..b961209 100644
--- a/media/libaudioclient/AudioSystem.cpp
+++ b/media/libaudioclient/AudioSystem.cpp
@@ -886,7 +886,6 @@
audio_stream_type_t *stream,
pid_t pid,
uid_t uid,
- const String16& opPackageName,
const audio_config_t *config,
audio_output_flags_t flags,
audio_port_handle_t *selectedDeviceId,
@@ -896,7 +895,7 @@
const sp<IAudioPolicyService>& aps = AudioSystem::get_audio_policy_service();
if (aps == 0) return NO_INIT;
return aps->getOutputForAttr(attr, output, session, stream, pid, uid,
- opPackageName, config,
+ config,
flags, selectedDeviceId, portId, secondaryOutputs);
}
diff --git a/media/libaudioclient/IAudioPolicyService.cpp b/media/libaudioclient/IAudioPolicyService.cpp
index f1213a3..4d2e369 100644
--- a/media/libaudioclient/IAudioPolicyService.cpp
+++ b/media/libaudioclient/IAudioPolicyService.cpp
@@ -215,7 +215,6 @@
audio_stream_type_t *stream,
pid_t pid,
uid_t uid,
- const String16& opPackageName,
const audio_config_t *config,
audio_output_flags_t flags,
audio_port_handle_t *selectedDeviceId,
@@ -254,7 +253,6 @@
}
data.writeInt32(pid);
data.writeInt32(uid);
- data.writeString16(opPackageName);
data.write(config, sizeof(audio_config_t));
data.writeInt32(static_cast <uint32_t>(flags));
data.writeInt32(*selectedDeviceId);
@@ -1656,11 +1654,6 @@
}
pid_t pid = (pid_t)data.readInt32();
uid_t uid = (uid_t)data.readInt32();
- String16 opPackageName;
- status = data.readString16(&opPackageName);
- if (status != NO_ERROR) {
- return status;
- }
audio_config_t config;
memset(&config, 0, sizeof(audio_config_t));
data.read(&config, sizeof(audio_config_t));
@@ -1672,7 +1665,7 @@
std::vector<audio_io_handle_t> secondaryOutputs;
status = getOutputForAttr(&attr,
&output, session, &stream, pid, uid,
- opPackageName, &config,
+ &config,
flags, &selectedDeviceId, &portId, &secondaryOutputs);
reply->writeInt32(status);
status = reply->write(&attr, sizeof(audio_attributes_t));
diff --git a/media/libaudioclient/include/media/AudioSystem.h b/media/libaudioclient/include/media/AudioSystem.h
index aebc875..6e395e0 100644
--- a/media/libaudioclient/include/media/AudioSystem.h
+++ b/media/libaudioclient/include/media/AudioSystem.h
@@ -241,7 +241,6 @@
audio_stream_type_t *stream,
pid_t pid,
uid_t uid,
- const String16& opPackageName,
const audio_config_t *config,
audio_output_flags_t flags,
audio_port_handle_t *selectedDeviceId,
diff --git a/media/libaudioclient/include/media/IAudioFlinger.h b/media/libaudioclient/include/media/IAudioFlinger.h
index c9d9716..612ce7a 100644
--- a/media/libaudioclient/include/media/IAudioFlinger.h
+++ b/media/libaudioclient/include/media/IAudioFlinger.h
@@ -71,7 +71,6 @@
if (clientInfo.readFromParcel(parcel) != NO_ERROR) {
return DEAD_OBJECT;
}
- opPackageName = parcel->readString16();
if (parcel->readInt32() != 0) {
// TODO: Using unsecurePointer() has some associated security
// pitfalls (see declaration for details).
@@ -101,7 +100,6 @@
(void)parcel->write(&attr, sizeof(audio_attributes_t));
(void)parcel->write(&config, sizeof(audio_config_t));
(void)clientInfo.writeToParcel(parcel);
- (void)parcel->writeString16(opPackageName);
if (sharedBuffer != 0) {
(void)parcel->writeInt32(1);
(void)parcel->writeStrongBinder(IInterface::asBinder(sharedBuffer));
@@ -125,7 +123,6 @@
audio_attributes_t attr;
audio_config_t config;
AudioClient clientInfo;
- String16 opPackageName;
sp<IMemory> sharedBuffer;
uint32_t notificationsPerBuffer;
float speed;
diff --git a/media/libaudioclient/include/media/IAudioPolicyService.h b/media/libaudioclient/include/media/IAudioPolicyService.h
index ec3461e..0ceca85 100644
--- a/media/libaudioclient/include/media/IAudioPolicyService.h
+++ b/media/libaudioclient/include/media/IAudioPolicyService.h
@@ -65,7 +65,6 @@
audio_stream_type_t *stream,
pid_t pid,
uid_t uid,
- const String16& opPackageName,
const audio_config_t *config,
audio_output_flags_t flags,
audio_port_handle_t *selectedDeviceId,
diff --git a/media/libstagefright/ACodecBufferChannel.cpp b/media/libstagefright/ACodecBufferChannel.cpp
index e5115d9..fa13f32 100644
--- a/media/libstagefright/ACodecBufferChannel.cpp
+++ b/media/libstagefright/ACodecBufferChannel.cpp
@@ -134,7 +134,12 @@
}
ssize_t result = -1;
ssize_t codecDataOffset = 0;
- if (mCrypto != NULL) {
+ if (numSubSamples == 1
+ && subSamples[0].mNumBytesOfClearData == 0
+ && subSamples[0].mNumBytesOfEncryptedData == 0) {
+ // We don't need to go through crypto or descrambler if the input is empty.
+ result = 0;
+ } else if (mCrypto != NULL) {
hardware::drm::V1_0::DestinationBuffer destination;
if (secure) {
destination.type = DrmBufferType::NATIVE_HANDLE;
diff --git a/media/libstagefright/MediaCodecSource.cpp b/media/libstagefright/MediaCodecSource.cpp
index 7fec072..1395c27 100644
--- a/media/libstagefright/MediaCodecSource.cpp
+++ b/media/libstagefright/MediaCodecSource.cpp
@@ -221,6 +221,7 @@
void MediaCodecSource::Puller::resume() {
Mutexed<Queue>::Locked queue(mQueue);
+ queue->flush();
queue->mPaused = false;
}
diff --git a/media/libstagefright/bqhelper/GraphicBufferSource.cpp b/media/libstagefright/bqhelper/GraphicBufferSource.cpp
index 986c9ac..cff14ac 100644
--- a/media/libstagefright/bqhelper/GraphicBufferSource.cpp
+++ b/media/libstagefright/bqhelper/GraphicBufferSource.cpp
@@ -1423,7 +1423,7 @@
// stall since no future events are expected.
mEndOfStream = true;
- if (mExecuting && !haveAvailableBuffers_l()) {
+ if (mStopTimeUs == -1 && mExecuting && !haveAvailableBuffers_l()) {
submitEndOfInputStream_l();
}
diff --git a/media/utils/ServiceUtilities.cpp b/media/utils/ServiceUtilities.cpp
index 7fd4d0a..87ea084 100644
--- a/media/utils/ServiceUtilities.cpp
+++ b/media/utils/ServiceUtilities.cpp
@@ -223,25 +223,6 @@
return ok;
}
-bool accessCallAudioAllowed(const String16& opPackageName, pid_t pid, uid_t uid) {
- static const String16 sAccessCallAudio("android.permission.ACCESS_CALL_AUDIO");
- PermissionController permissionController;
- const String16 resolvedOpPackageName = resolveCallingPackage(
- permissionController, opPackageName, uid);
- if (resolvedOpPackageName.size() == 0) {
- ALOGE("accessCallAudioAllowed - FAIL - package not found.");
- return false;
- }
- AppOpsManager appOps;
- const int32_t op = appOps.permissionToOpCode(sAccessCallAudio);
- const int32_t opResult = appOps.noteOp(op, uid, resolvedOpPackageName);
- if (opResult == PermissionController::MODE_DEFAULT) {
- // Only allow in case this is a system app with the proper privilege permission
- return PermissionCache::checkPermission(sAccessCallAudio, pid, uid);
- }
- return opResult == PermissionController::MODE_ALLOWED;
-}
-
// privileged behavior needed by Dialer, Settings, SetupWizard and CellBroadcastReceiver
bool bypassInterruptionPolicyAllowed(pid_t pid, uid_t uid) {
static const String16 sWriteSecureSettings("android.permission.WRITE_SECURE_SETTINGS");
@@ -278,29 +259,28 @@
return NO_ERROR;
}
-void MediaPackageManager::loadPackageManager() {
- if (mPackageManager != nullptr) {
- return;
- }
+sp<content::pm::IPackageManagerNative> MediaPackageManager::retreivePackageManager() {
const sp<IServiceManager> sm = defaultServiceManager();
if (sm == nullptr) {
ALOGW("%s: failed to retrieve defaultServiceManager", __func__);
- return;
+ return nullptr;
}
sp<IBinder> packageManager = sm->checkService(String16(nativePackageManagerName));
if (packageManager == nullptr) {
ALOGW("%s: failed to retrieve native package manager", __func__);
- return;
+ return nullptr;
}
- mPackageManager = interface_cast<content::pm::IPackageManagerNative>(packageManager);
+ return interface_cast<content::pm::IPackageManagerNative>(packageManager);
}
std::optional<bool> MediaPackageManager::doIsAllowed(uid_t uid) {
- /** Can not fetch package manager at construction it may not yet be registered. */
- loadPackageManager();
if (mPackageManager == nullptr) {
- ALOGW("%s: Playback capture is denied as package manager is not reachable", __func__);
- return std::nullopt;
+ /** Can not fetch package manager at construction it may not yet be registered. */
+ mPackageManager = retreivePackageManager();
+ if (mPackageManager == nullptr) {
+ ALOGW("%s: Playback capture is denied as package manager is not reachable", __func__);
+ return std::nullopt;
+ }
}
std::vector<std::string> packageNames;
diff --git a/media/utils/include/mediautils/ServiceUtilities.h b/media/utils/include/mediautils/ServiceUtilities.h
index 060e849..212599a 100644
--- a/media/utils/include/mediautils/ServiceUtilities.h
+++ b/media/utils/include/mediautils/ServiceUtilities.h
@@ -93,7 +93,6 @@
bool dumpAllowed();
bool modifyPhoneStateAllowed(pid_t pid, uid_t uid);
bool bypassInterruptionPolicyAllowed(pid_t pid, uid_t uid);
-bool accessCallAudioAllowed(const String16& opPackageName, pid_t pid, uid_t uid);
status_t checkIMemory(const sp<IMemory>& iMemory);
@@ -111,7 +110,7 @@
private:
static constexpr const char* nativePackageManagerName = "package_native";
std::optional<bool> doIsAllowed(uid_t uid);
- void loadPackageManager();
+ sp<content::pm::IPackageManagerNative> retreivePackageManager();
sp<content::pm::IPackageManagerNative> mPackageManager; // To check apps manifest
uint_t mPackageManagerErrors = 0;
struct Package {
diff --git a/services/audioflinger/AudioFlinger.cpp b/services/audioflinger/AudioFlinger.cpp
index ecda56b..1be2fcb 100644
--- a/services/audioflinger/AudioFlinger.cpp
+++ b/services/audioflinger/AudioFlinger.cpp
@@ -343,7 +343,7 @@
ret = AudioSystem::getOutputForAttr(&localAttr, &io,
actualSessionId,
&streamType, client.clientPid, client.clientUid,
- client.packageName, &fullConfig,
+ &fullConfig,
(audio_output_flags_t)(AUDIO_OUTPUT_FLAG_MMAP_NOIRQ |
AUDIO_OUTPUT_FLAG_DIRECT),
deviceId, &portId, &secondaryOutputs);
@@ -783,9 +783,8 @@
output.outputId = AUDIO_IO_HANDLE_NONE;
output.selectedDeviceId = input.selectedDeviceId;
lStatus = AudioSystem::getOutputForAttr(&localAttr, &output.outputId, sessionId, &streamType,
- clientPid, clientUid, input.opPackageName,
- &input.config, input.flags, &output.selectedDeviceId,
- &portId, &secondaryOutputs);
+ clientPid, clientUid, &input.config, input.flags,
+ &output.selectedDeviceId, &portId, &secondaryOutputs);
if (lStatus != NO_ERROR || output.outputId == AUDIO_IO_HANDLE_NONE) {
ALOGE("createTrack() getOutputForAttr() return error %d or invalid output handle", lStatus);
diff --git a/services/audioflinger/Threads.cpp b/services/audioflinger/Threads.cpp
index 65cf96c..d8d4d35 100644
--- a/services/audioflinger/Threads.cpp
+++ b/services/audioflinger/Threads.cpp
@@ -8774,7 +8774,6 @@
&stream,
client.clientPid,
client.clientUid,
- client.packageName,
&config,
flags,
&deviceId,
diff --git a/services/audiopolicy/service/AudioPolicyInterfaceImpl.cpp b/services/audiopolicy/service/AudioPolicyInterfaceImpl.cpp
index 0da3b9c..57b23b7 100644
--- a/services/audiopolicy/service/AudioPolicyInterfaceImpl.cpp
+++ b/services/audiopolicy/service/AudioPolicyInterfaceImpl.cpp
@@ -219,7 +219,6 @@
audio_stream_type_t *stream,
pid_t pid,
uid_t uid,
- const String16& opPackageName,
const audio_config_t *config,
audio_output_flags_t flags,
audio_port_handle_t *selectedDeviceId,
@@ -266,8 +265,7 @@
case AudioPolicyInterface::API_OUTPUT_LEGACY:
break;
case AudioPolicyInterface::API_OUTPUT_TELEPHONY_TX:
- if (!modifyPhoneStateAllowed(pid, uid) &&
- !accessCallAudioAllowed(opPackageName, pid, uid)) {
+ if (!modifyPhoneStateAllowed(pid, uid)) {
ALOGE("%s() permission denied: modify phone state not allowed for uid %d",
__func__, uid);
result = PERMISSION_DENIED;
@@ -464,22 +462,15 @@
}
bool canCaptureOutput = captureAudioOutputAllowed(pid, uid);
- bool canCaptureTelephonyOutput = canCaptureOutput
- || accessCallAudioAllowed(opPackageName, pid, uid);
-
- if ((attr->source == AUDIO_SOURCE_ECHO_REFERENCE ||
- attr->source == AUDIO_SOURCE_FM_TUNER) &&
+ if ((inputSource == AUDIO_SOURCE_VOICE_UPLINK ||
+ inputSource == AUDIO_SOURCE_VOICE_DOWNLINK ||
+ inputSource == AUDIO_SOURCE_VOICE_CALL ||
+ inputSource == AUDIO_SOURCE_ECHO_REFERENCE||
+ inputSource == AUDIO_SOURCE_FM_TUNER) &&
!canCaptureOutput) {
return PERMISSION_DENIED;
}
- if ((attr->source == AUDIO_SOURCE_VOICE_UPLINK ||
- attr->source == AUDIO_SOURCE_VOICE_DOWNLINK ||
- attr->source == AUDIO_SOURCE_VOICE_CALL) &&
- !canCaptureTelephonyOutput) {
- return PERMISSION_DENIED;
- }
-
bool canCaptureHotword = captureHotwordAllowed(opPackageName, pid, uid);
if ((inputSource == AUDIO_SOURCE_HOTWORD) && !canCaptureHotword) {
return BAD_VALUE;
@@ -511,11 +502,6 @@
break;
case AudioPolicyInterface::API_INPUT_TELEPHONY_RX:
// FIXME: use the same permission as for remote submix for now.
- if (!canCaptureTelephonyOutput) {
- ALOGE("getInputForAttr() permission denied: call capture not allowed");
- status = PERMISSION_DENIED;
- }
- break;
case AudioPolicyInterface::API_INPUT_MIX_CAPTURE:
if (!canCaptureOutput) {
ALOGE("getInputForAttr() permission denied: capture not allowed");
@@ -543,13 +529,9 @@
return status;
}
- bool allowAudioCapture = canCaptureOutput ||
- (inputType == AudioPolicyInterface::API_INPUT_TELEPHONY_RX &&
- canCaptureTelephonyOutput);
-
sp<AudioRecordClient> client = new AudioRecordClient(*attr, *input, uid, pid, session, *portId,
*selectedDeviceId, opPackageName,
- allowAudioCapture, canCaptureHotword);
+ canCaptureOutput, canCaptureHotword);
mAudioRecordClients.add(*portId, client);
}
diff --git a/services/audiopolicy/service/AudioPolicyService.cpp b/services/audiopolicy/service/AudioPolicyService.cpp
index bec6aff..dd46ffb 100644
--- a/services/audiopolicy/service/AudioPolicyService.cpp
+++ b/services/audiopolicy/service/AudioPolicyService.cpp
@@ -543,7 +543,7 @@
// else
// favor the privacy sensitive case
if (topActive != nullptr && topSensitiveActive != nullptr
- && !topActive->canCaptureCallOrOutput) {
+ && !topActive->canCaptureOutput) {
topActive = nullptr;
}
@@ -559,8 +559,8 @@
false : current->uid == topSensitiveActive->uid;
auto canCaptureIfInCallOrCommunication = [&](const auto &recordClient) {
- bool canCaptureCall = recordClient->canCaptureCallOrOutput;
- bool canCaptureCommunication = recordClient->canCaptureCallOrOutput
+ bool canCaptureCall = recordClient->canCaptureOutput;
+ bool canCaptureCommunication = recordClient->canCaptureOutput
|| recordClient->uid == mPhoneStateOwnerUid
|| isServiceUid(mPhoneStateOwnerUid);
return !(isInCall && !canCaptureCall)
@@ -575,7 +575,7 @@
bool allowCapture = !isAssistantOnTop
&& (isTopOrLatestActive || isTopOrLatestSensitive)
&& !(isSensitiveActive
- && !(isTopOrLatestSensitive || current->canCaptureCallOrOutput))
+ && !(isTopOrLatestSensitive || current->canCaptureOutput))
&& canCaptureIfInCallOrCommunication(current);
if (isVirtualSource(source)) {
@@ -596,7 +596,7 @@
} else {
if (((isAssistantOnTop && source == AUDIO_SOURCE_VOICE_RECOGNITION) ||
source == AUDIO_SOURCE_HOTWORD)
- && !(isSensitiveActive && !current->canCaptureCallOrOutput)
+ && !(isSensitiveActive && !current->canCaptureOutput)
&& canCaptureIfInCallOrCommunication(current)) {
allowCapture = true;
}
@@ -609,7 +609,7 @@
// OR
// Is on TOP AND the source is VOICE_RECOGNITION or HOTWORD
if (!isAssistantOnTop
- && !(isSensitiveActive && !current->canCaptureCallOrOutput)
+ && !(isSensitiveActive && !current->canCaptureOutput)
&& canCaptureIfInCallOrCommunication(current)) {
allowCapture = true;
}
@@ -945,7 +945,7 @@
}
}
ActivityManager am;
- bool active = am.isUidActive(uid, String16("audioserver"));
+ bool active = am.isUidActiveOrForeground(uid, String16("audioserver"));
{
Mutex::Autolock _l(mLock);
mCachedUids.insert(std::pair<uid_t,
@@ -990,7 +990,7 @@
}
}
ActivityManager am;
- bool active = am.isUidActive(uid, String16("audioserver"));
+ bool active = am.isUidActiveOrForeground(uid, String16("audioserver"));
int state = ActivityManager::PROCESS_STATE_UNKNOWN;
if (active) {
state = am.getUidProcessState(uid, String16("audioserver"));
diff --git a/services/audiopolicy/service/AudioPolicyService.h b/services/audiopolicy/service/AudioPolicyService.h
index ff99124..9a619dd 100644
--- a/services/audiopolicy/service/AudioPolicyService.h
+++ b/services/audiopolicy/service/AudioPolicyService.h
@@ -83,7 +83,6 @@
audio_stream_type_t *stream,
pid_t pid,
uid_t uid,
- const String16& opPackageName,
const audio_config_t *config,
audio_output_flags_t flags,
audio_port_handle_t *selectedDeviceId,
@@ -812,16 +811,15 @@
const audio_io_handle_t io, uid_t uid, pid_t pid,
const audio_session_t session, audio_port_handle_t portId,
const audio_port_handle_t deviceId, const String16& opPackageName,
- bool canCaptureCallOrOutput, bool canCaptureHotword) :
+ bool canCaptureOutput, bool canCaptureHotword) :
AudioClient(attributes, io, uid, pid, session, portId, deviceId),
opPackageName(opPackageName), startTimeNs(0),
- canCaptureCallOrOutput(canCaptureCallOrOutput),
- canCaptureHotword(canCaptureHotword) {}
+ canCaptureOutput(canCaptureOutput), canCaptureHotword(canCaptureHotword) {}
~AudioRecordClient() override = default;
const String16 opPackageName; // client package name
nsecs_t startTimeNs;
- const bool canCaptureCallOrOutput;
+ const bool canCaptureOutput;
const bool canCaptureHotword;
};
diff --git a/services/camera/libcameraservice/CameraService.cpp b/services/camera/libcameraservice/CameraService.cpp
index 9bc79e0..9af1c36 100644
--- a/services/camera/libcameraservice/CameraService.cpp
+++ b/services/camera/libcameraservice/CameraService.cpp
@@ -3188,7 +3188,9 @@
// some polling which should happen pretty rarely anyway as the race is hard
// to hit.
active = mActiveUids.find(uid) != mActiveUids.end();
- if (!active) active = am.isUidActive(uid, callingPackage);
+ if (!active) {
+ active = am.isUidActiveOrForeground(uid, callingPackage);
+ }
if (active) {
break;
}
diff --git a/services/mediacodec/registrant/CodecServiceRegistrant.cpp b/services/mediacodec/registrant/CodecServiceRegistrant.cpp
index 83d233e..184251a 100644
--- a/services/mediacodec/registrant/CodecServiceRegistrant.cpp
+++ b/services/mediacodec/registrant/CodecServiceRegistrant.cpp
@@ -17,11 +17,13 @@
//#define LOG_NDEBUG 0
#define LOG_TAG "CodecServiceRegistrant"
+#include <android-base/properties.h>
#include <android-base/logging.h>
#include <android-base/properties.h>
#include <C2Component.h>
#include <C2PlatformSupport.h>
+#include <codec2/hidl/1.0/ComponentStore.h>
#include <codec2/hidl/1.1/ComponentStore.h>
#include <codec2/hidl/1.1/Configurable.h>
#include <codec2/hidl/1.1/types.h>
@@ -43,6 +45,10 @@
// Converter from IComponentStore to C2ComponentStore.
class H2C2ComponentStore : public C2ComponentStore {
protected:
+ using IComponentStore =
+ ::android::hardware::media::c2::V1_0::IComponentStore;
+ using IConfigurable =
+ ::android::hardware::media::c2::V1_0::IConfigurable;
sp<IComponentStore> mStore;
sp<IConfigurable> mConfigurable;
public:
@@ -399,36 +405,67 @@
} // unnamed namespace
extern "C" void RegisterCodecServices() {
- using namespace ::android::hardware::media::c2::V1_1;
LOG(INFO) << "Creating software Codec2 service...";
- sp<ComponentStore> store =
- new ComponentStore(::android::GetCodec2PlatformComponentStore());
- if (store == nullptr) {
- LOG(ERROR) <<
- "Cannot create software Codec2 service.";
- } else {
- if (!ionPropertiesDefined()) {
- std::string preferredStoreName = "default";
- sp<IComponentStore> preferredStore =
- IComponentStore::getService(preferredStoreName.c_str());
- if (preferredStore) {
- ::android::SetPreferredCodec2ComponentStore(
- std::make_shared<H2C2ComponentStore>(preferredStore));
- LOG(INFO) <<
- "Preferred Codec2 store is set to \"" <<
- preferredStoreName << "\".";
- } else {
- LOG(INFO) <<
- "Preferred Codec2 store is defaulted to \"software\".";
+ std::shared_ptr<C2ComponentStore> store =
+ android::GetCodec2PlatformComponentStore();
+ if (!store) {
+ LOG(ERROR) << "Failed to create Codec2 service.";
+ return;
+ }
+
+ using namespace ::android::hardware::media::c2;
+
+ int platformVersion =
+ android::base::GetIntProperty("ro.build.version.sdk", int32_t(29));
+ // STOPSHIP: Remove code name checking once platform version bumps up to 30.
+ std::string codeName =
+ android::base::GetProperty("ro.build.version.codename", "");
+ if (codeName == "R") {
+ platformVersion = 30;
+ }
+
+ switch (platformVersion) {
+ case 30: {
+ android::sp<V1_1::IComponentStore> storeV1_1 =
+ new V1_1::utils::ComponentStore(store);
+ if (storeV1_1->registerAsService("software") != android::OK) {
+ LOG(ERROR) << "Cannot register software Codec2 v1.1 service.";
+ return;
}
+ break;
}
- if (store->registerAsService("software") != android::OK) {
- LOG(ERROR) <<
- "Cannot register software Codec2 service.";
- } else {
- LOG(INFO) <<
- "Software Codec2 service created.";
+ case 29: {
+ android::sp<V1_0::IComponentStore> storeV1_0 =
+ new V1_0::utils::ComponentStore(store);
+ if (storeV1_0->registerAsService("software") != android::OK) {
+ LOG(ERROR) << "Cannot register software Codec2 v1.0 service.";
+ return;
+ }
+ break;
+ }
+ default: {
+ LOG(ERROR) << "The platform version " << platformVersion <<
+ " is not supported.";
+ return;
}
}
+ if (!ionPropertiesDefined()) {
+ using IComponentStore =
+ ::android::hardware::media::c2::V1_0::IComponentStore;
+ std::string const preferredStoreName = "default";
+ sp<IComponentStore> preferredStore =
+ IComponentStore::getService(preferredStoreName.c_str());
+ if (preferredStore) {
+ ::android::SetPreferredCodec2ComponentStore(
+ std::make_shared<H2C2ComponentStore>(preferredStore));
+ LOG(INFO) <<
+ "Preferred Codec2 store is set to \"" <<
+ preferredStoreName << "\".";
+ } else {
+ LOG(INFO) <<
+ "Preferred Codec2 store is defaulted to \"software\".";
+ }
+ }
+ LOG(INFO) << "Software Codec2 service created and registered.";
}