| /* |
| * Copyright (C) 2010 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| //#define LOG_NDEBUG 0 |
| #define LOG_TAG "NuPlayer" |
| |
| #include <inttypes.h> |
| |
| #include <utils/Log.h> |
| |
| #include "NuPlayer.h" |
| |
| #include "HTTPLiveSource.h" |
| #include "NuPlayerCCDecoder.h" |
| #include "NuPlayerDecoder.h" |
| #include "NuPlayerDecoderBase.h" |
| #include "NuPlayerDecoderPassThrough.h" |
| #include "NuPlayerDriver.h" |
| #include "NuPlayerRenderer.h" |
| #include "NuPlayerSource.h" |
| #include "RTPSource.h" |
| #include "RTSPSource.h" |
| #include "StreamingSource.h" |
| #include "GenericSource.h" |
| #include <timedtext/TextDescriptions.h> |
| |
| #include <cutils/properties.h> |
| |
| #include <media/AudioResamplerPublic.h> |
| #include <media/AVSyncSettings.h> |
| #include <media/MediaCodecBuffer.h> |
| |
| #include <media/stagefright/foundation/hexdump.h> |
| #include <media/stagefright/foundation/ABuffer.h> |
| #include <media/stagefright/foundation/ADebug.h> |
| #include <media/stagefright/foundation/AMessage.h> |
| #include <media/stagefright/foundation/avc_utils.h> |
| #include <media/stagefright/MediaBuffer.h> |
| #include <media/stagefright/MediaClock.h> |
| #include <media/stagefright/MediaDefs.h> |
| #include <media/stagefright/MediaErrors.h> |
| #include <media/stagefright/MetaData.h> |
| |
| #include <mpeg2ts/ATSParser.h> |
| |
| #include <gui/IGraphicBufferProducer.h> |
| #include <gui/Surface.h> |
| |
| |
| #include <media/esds/ESDS.h> |
| #include <media/stagefright/Utils.h> |
| |
| namespace android { |
| |
| struct NuPlayer::Action : public RefBase { |
| Action() {} |
| |
| virtual void execute(NuPlayer *player) = 0; |
| |
| private: |
| DISALLOW_EVIL_CONSTRUCTORS(Action); |
| }; |
| |
| struct NuPlayer::SeekAction : public Action { |
| explicit SeekAction(int64_t seekTimeUs, MediaPlayerSeekMode mode) |
| : mSeekTimeUs(seekTimeUs), |
| mMode(mode) { |
| } |
| |
| virtual void execute(NuPlayer *player) { |
| player->performSeek(mSeekTimeUs, mMode); |
| } |
| |
| private: |
| int64_t mSeekTimeUs; |
| MediaPlayerSeekMode mMode; |
| |
| DISALLOW_EVIL_CONSTRUCTORS(SeekAction); |
| }; |
| |
| struct NuPlayer::ResumeDecoderAction : public Action { |
| explicit ResumeDecoderAction(bool needNotify) |
| : mNeedNotify(needNotify) { |
| } |
| |
| virtual void execute(NuPlayer *player) { |
| player->performResumeDecoders(mNeedNotify); |
| } |
| |
| private: |
| bool mNeedNotify; |
| |
| DISALLOW_EVIL_CONSTRUCTORS(ResumeDecoderAction); |
| }; |
| |
| struct NuPlayer::SetSurfaceAction : public Action { |
| explicit SetSurfaceAction(const sp<Surface> &surface) |
| : mSurface(surface) { |
| } |
| |
| virtual void execute(NuPlayer *player) { |
| player->performSetSurface(mSurface); |
| } |
| |
| private: |
| sp<Surface> mSurface; |
| |
| DISALLOW_EVIL_CONSTRUCTORS(SetSurfaceAction); |
| }; |
| |
| struct NuPlayer::FlushDecoderAction : public Action { |
| FlushDecoderAction(FlushCommand audio, FlushCommand video) |
| : mAudio(audio), |
| mVideo(video) { |
| } |
| |
| virtual void execute(NuPlayer *player) { |
| player->performDecoderFlush(mAudio, mVideo); |
| } |
| |
| private: |
| FlushCommand mAudio; |
| FlushCommand mVideo; |
| |
| DISALLOW_EVIL_CONSTRUCTORS(FlushDecoderAction); |
| }; |
| |
| struct NuPlayer::PostMessageAction : public Action { |
| explicit PostMessageAction(const sp<AMessage> &msg) |
| : mMessage(msg) { |
| } |
| |
| virtual void execute(NuPlayer *) { |
| mMessage->post(); |
| } |
| |
| private: |
| sp<AMessage> mMessage; |
| |
| DISALLOW_EVIL_CONSTRUCTORS(PostMessageAction); |
| }; |
| |
| // Use this if there's no state necessary to save in order to execute |
| // the action. |
| struct NuPlayer::SimpleAction : public Action { |
| typedef void (NuPlayer::*ActionFunc)(); |
| |
| explicit SimpleAction(ActionFunc func) |
| : mFunc(func) { |
| } |
| |
| virtual void execute(NuPlayer *player) { |
| (player->*mFunc)(); |
| } |
| |
| private: |
| ActionFunc mFunc; |
| |
| DISALLOW_EVIL_CONSTRUCTORS(SimpleAction); |
| }; |
| |
| //////////////////////////////////////////////////////////////////////////////// |
| |
| NuPlayer::NuPlayer(pid_t pid, const sp<MediaClock> &mediaClock) |
| : mUIDValid(false), |
| mPID(pid), |
| mMediaClock(mediaClock), |
| mSourceFlags(0), |
| mOffloadAudio(false), |
| mAudioDecoderGeneration(0), |
| mVideoDecoderGeneration(0), |
| mRendererGeneration(0), |
| mLastStartedPlayingTimeNs(0), |
| mLastStartedRebufferingTimeNs(0), |
| mPreviousSeekTimeUs(0), |
| mAudioEOS(false), |
| mVideoEOS(false), |
| mScanSourcesPending(false), |
| mScanSourcesGeneration(0), |
| mPollDurationGeneration(0), |
| mTimedTextGeneration(0), |
| mFlushingAudio(NONE), |
| mFlushingVideo(NONE), |
| mResumePending(false), |
| mVideoScalingMode(NATIVE_WINDOW_SCALING_MODE_SCALE_TO_WINDOW), |
| mPlaybackSettings(AUDIO_PLAYBACK_RATE_DEFAULT), |
| mVideoFpsHint(-1.f), |
| mStarted(false), |
| mPrepared(false), |
| mResetting(false), |
| mSourceStarted(false), |
| mAudioDecoderError(false), |
| mVideoDecoderError(false), |
| mPaused(false), |
| mPausedByClient(true), |
| mPausedForBuffering(false), |
| mIsDrmProtected(false), |
| mDataSourceType(DATA_SOURCE_TYPE_NONE) { |
| CHECK(mediaClock != NULL); |
| clearFlushComplete(); |
| } |
| |
| NuPlayer::~NuPlayer() { |
| } |
| |
| void NuPlayer::setUID(uid_t uid) { |
| mUIDValid = true; |
| mUID = uid; |
| } |
| |
| void NuPlayer::init(const wp<NuPlayerDriver> &driver) { |
| mDriver = driver; |
| |
| sp<AMessage> notify = new AMessage(kWhatMediaClockNotify, this); |
| mMediaClock->setNotificationMessage(notify); |
| } |
| |
| void NuPlayer::setDataSourceAsync(const sp<IStreamSource> &source) { |
| sp<AMessage> msg = new AMessage(kWhatSetDataSource, this); |
| |
| sp<AMessage> notify = new AMessage(kWhatSourceNotify, this); |
| |
| msg->setObject("source", new StreamingSource(notify, source)); |
| msg->post(); |
| mDataSourceType = DATA_SOURCE_TYPE_STREAM; |
| } |
| |
| static bool IsHTTPLiveURL(const char *url) { |
| if (!strncasecmp("http://", url, 7) |
| || !strncasecmp("https://", url, 8) |
| || !strncasecmp("file://", url, 7)) { |
| size_t len = strlen(url); |
| if (len >= 5 && !strcasecmp(".m3u8", &url[len - 5])) { |
| return true; |
| } |
| |
| if (strstr(url,"m3u8")) { |
| return true; |
| } |
| } |
| |
| return false; |
| } |
| |
| void NuPlayer::setDataSourceAsync( |
| const sp<IMediaHTTPService> &httpService, |
| const char *url, |
| const KeyedVector<String8, String8> *headers) { |
| |
| sp<AMessage> msg = new AMessage(kWhatSetDataSource, this); |
| size_t len = strlen(url); |
| |
| sp<AMessage> notify = new AMessage(kWhatSourceNotify, this); |
| |
| sp<Source> source; |
| if (IsHTTPLiveURL(url)) { |
| source = new HTTPLiveSource(notify, httpService, url, headers); |
| ALOGV("setDataSourceAsync HTTPLiveSource %s", url); |
| mDataSourceType = DATA_SOURCE_TYPE_HTTP_LIVE; |
| } else if (!strncasecmp(url, "rtsp://", 7)) { |
| source = new RTSPSource( |
| notify, httpService, url, headers, mUIDValid, mUID); |
| ALOGV("setDataSourceAsync RTSPSource %s", url); |
| mDataSourceType = DATA_SOURCE_TYPE_RTSP; |
| } else if ((!strncasecmp(url, "http://", 7) |
| || !strncasecmp(url, "https://", 8)) |
| && ((len >= 4 && !strcasecmp(".sdp", &url[len - 4])) |
| || strstr(url, ".sdp?"))) { |
| source = new RTSPSource( |
| notify, httpService, url, headers, mUIDValid, mUID, true); |
| ALOGV("setDataSourceAsync RTSPSource http/https/.sdp %s", url); |
| mDataSourceType = DATA_SOURCE_TYPE_RTSP; |
| } else { |
| ALOGV("setDataSourceAsync GenericSource %s", url); |
| |
| sp<GenericSource> genericSource = |
| new GenericSource(notify, mUIDValid, mUID, mMediaClock); |
| |
| status_t err = genericSource->setDataSource(httpService, url, headers); |
| |
| if (err == OK) { |
| source = genericSource; |
| } else { |
| ALOGE("Failed to set data source!"); |
| } |
| |
| // regardless of success/failure |
| mDataSourceType = DATA_SOURCE_TYPE_GENERIC_URL; |
| } |
| msg->setObject("source", source); |
| msg->post(); |
| } |
| |
| void NuPlayer::setDataSourceAsync(int fd, int64_t offset, int64_t length) { |
| sp<AMessage> msg = new AMessage(kWhatSetDataSource, this); |
| |
| sp<AMessage> notify = new AMessage(kWhatSourceNotify, this); |
| |
| sp<GenericSource> source = |
| new GenericSource(notify, mUIDValid, mUID, mMediaClock); |
| |
| ALOGV("setDataSourceAsync fd %d/%lld/%lld source: %p", |
| fd, (long long)offset, (long long)length, source.get()); |
| |
| status_t err = source->setDataSource(fd, offset, length); |
| |
| if (err != OK) { |
| ALOGE("Failed to set data source!"); |
| source = NULL; |
| } |
| |
| msg->setObject("source", source); |
| msg->post(); |
| mDataSourceType = DATA_SOURCE_TYPE_GENERIC_FD; |
| } |
| |
| void NuPlayer::setDataSourceAsync(const sp<DataSource> &dataSource) { |
| sp<AMessage> msg = new AMessage(kWhatSetDataSource, this); |
| sp<AMessage> notify = new AMessage(kWhatSourceNotify, this); |
| |
| sp<GenericSource> source = new GenericSource(notify, mUIDValid, mUID, mMediaClock); |
| status_t err = source->setDataSource(dataSource); |
| |
| if (err != OK) { |
| ALOGE("Failed to set data source!"); |
| source = NULL; |
| } |
| |
| msg->setObject("source", source); |
| msg->post(); |
| mDataSourceType = DATA_SOURCE_TYPE_MEDIA; |
| } |
| |
| status_t NuPlayer::getBufferingSettings( |
| BufferingSettings *buffering /* nonnull */) { |
| sp<AMessage> msg = new AMessage(kWhatGetBufferingSettings, this); |
| sp<AMessage> response; |
| status_t err = msg->postAndAwaitResponse(&response); |
| if (err == OK && response != NULL) { |
| CHECK(response->findInt32("err", &err)); |
| if (err == OK) { |
| readFromAMessage(response, buffering); |
| } |
| } |
| return err; |
| } |
| |
| status_t NuPlayer::setBufferingSettings(const BufferingSettings& buffering) { |
| sp<AMessage> msg = new AMessage(kWhatSetBufferingSettings, this); |
| writeToAMessage(msg, buffering); |
| sp<AMessage> response; |
| status_t err = msg->postAndAwaitResponse(&response); |
| if (err == OK && response != NULL) { |
| CHECK(response->findInt32("err", &err)); |
| } |
| return err; |
| } |
| |
| void NuPlayer::setDataSourceAsync(const String8& rtpParams) { |
| ALOGD("setDataSourceAsync for RTP = %s", rtpParams.c_str()); |
| sp<AMessage> msg = new AMessage(kWhatSetDataSource, this); |
| |
| sp<AMessage> notify = new AMessage(kWhatSourceNotify, this); |
| sp<Source> source = new RTPSource(notify, rtpParams); |
| |
| msg->setObject("source", source); |
| msg->post(); |
| mDataSourceType = DATA_SOURCE_TYPE_RTP; |
| } |
| |
| void NuPlayer::prepareAsync() { |
| ALOGV("prepareAsync"); |
| |
| (new AMessage(kWhatPrepare, this))->post(); |
| } |
| |
| void NuPlayer::setVideoSurfaceTextureAsync( |
| const sp<IGraphicBufferProducer> &bufferProducer) { |
| sp<AMessage> msg = new AMessage(kWhatSetVideoSurface, this); |
| |
| if (bufferProducer == NULL) { |
| msg->setObject("surface", NULL); |
| } else { |
| msg->setObject("surface", new Surface(bufferProducer, true /* controlledByApp */)); |
| } |
| |
| msg->post(); |
| } |
| |
| void NuPlayer::setAudioSink(const sp<MediaPlayerBase::AudioSink> &sink) { |
| sp<AMessage> msg = new AMessage(kWhatSetAudioSink, this); |
| msg->setObject("sink", sink); |
| msg->post(); |
| } |
| |
| void NuPlayer::start() { |
| (new AMessage(kWhatStart, this))->post(); |
| } |
| |
| status_t NuPlayer::setPlaybackSettings(const AudioPlaybackRate &rate) { |
| // do some cursory validation of the settings here. audio modes are |
| // only validated when set on the audiosink. |
| if ((rate.mSpeed != 0.f && rate.mSpeed < AUDIO_TIMESTRETCH_SPEED_MIN) |
| || rate.mSpeed > AUDIO_TIMESTRETCH_SPEED_MAX |
| || rate.mPitch < AUDIO_TIMESTRETCH_SPEED_MIN |
| || rate.mPitch > AUDIO_TIMESTRETCH_SPEED_MAX) { |
| return BAD_VALUE; |
| } |
| sp<AMessage> msg = new AMessage(kWhatConfigPlayback, this); |
| writeToAMessage(msg, rate); |
| sp<AMessage> response; |
| status_t err = msg->postAndAwaitResponse(&response); |
| if (err == OK && response != NULL) { |
| CHECK(response->findInt32("err", &err)); |
| } |
| return err; |
| } |
| |
| status_t NuPlayer::getPlaybackSettings(AudioPlaybackRate *rate /* nonnull */) { |
| sp<AMessage> msg = new AMessage(kWhatGetPlaybackSettings, this); |
| sp<AMessage> response; |
| status_t err = msg->postAndAwaitResponse(&response); |
| if (err == OK && response != NULL) { |
| CHECK(response->findInt32("err", &err)); |
| if (err == OK) { |
| readFromAMessage(response, rate); |
| } |
| } |
| return err; |
| } |
| |
| status_t NuPlayer::setSyncSettings(const AVSyncSettings &sync, float videoFpsHint) { |
| sp<AMessage> msg = new AMessage(kWhatConfigSync, this); |
| writeToAMessage(msg, sync, videoFpsHint); |
| sp<AMessage> response; |
| status_t err = msg->postAndAwaitResponse(&response); |
| if (err == OK && response != NULL) { |
| CHECK(response->findInt32("err", &err)); |
| } |
| return err; |
| } |
| |
| status_t NuPlayer::getSyncSettings( |
| AVSyncSettings *sync /* nonnull */, float *videoFps /* nonnull */) { |
| sp<AMessage> msg = new AMessage(kWhatGetSyncSettings, this); |
| sp<AMessage> response; |
| status_t err = msg->postAndAwaitResponse(&response); |
| if (err == OK && response != NULL) { |
| CHECK(response->findInt32("err", &err)); |
| if (err == OK) { |
| readFromAMessage(response, sync, videoFps); |
| } |
| } |
| return err; |
| } |
| |
| void NuPlayer::pause() { |
| (new AMessage(kWhatPause, this))->post(); |
| } |
| |
| void NuPlayer::resetAsync() { |
| sp<Source> source; |
| { |
| Mutex::Autolock autoLock(mSourceLock); |
| source = mSource; |
| } |
| |
| if (source != NULL) { |
| // During a reset, the data source might be unresponsive already, we need to |
| // disconnect explicitly so that reads exit promptly. |
| // We can't queue the disconnect request to the looper, as it might be |
| // queued behind a stuck read and never gets processed. |
| // Doing a disconnect outside the looper to allows the pending reads to exit |
| // (either successfully or with error). |
| source->disconnect(); |
| } |
| |
| (new AMessage(kWhatReset, this))->post(); |
| } |
| |
| status_t NuPlayer::notifyAt(int64_t mediaTimeUs) { |
| sp<AMessage> notify = new AMessage(kWhatNotifyTime, this); |
| notify->setInt64("timerUs", mediaTimeUs); |
| mMediaClock->addTimer(notify, mediaTimeUs); |
| return OK; |
| } |
| |
| void NuPlayer::seekToAsync(int64_t seekTimeUs, MediaPlayerSeekMode mode, bool needNotify) { |
| sp<AMessage> msg = new AMessage(kWhatSeek, this); |
| msg->setInt64("seekTimeUs", seekTimeUs); |
| msg->setInt32("mode", mode); |
| msg->setInt32("needNotify", needNotify); |
| msg->post(); |
| } |
| |
| |
| void NuPlayer::writeTrackInfo( |
| Parcel* reply, const sp<AMessage>& format) const { |
| if (format == NULL) { |
| ALOGE("NULL format"); |
| return; |
| } |
| int32_t trackType; |
| if (!format->findInt32("type", &trackType)) { |
| ALOGE("no track type"); |
| return; |
| } |
| |
| AString mime; |
| if (!format->findString("mime", &mime)) { |
| // Java MediaPlayer only uses mimetype for subtitle and timedtext tracks. |
| // If we can't find the mimetype here it means that we wouldn't be needing |
| // the mimetype on the Java end. We still write a placeholder mime to keep the |
| // (de)serialization logic simple. |
| if (trackType == MEDIA_TRACK_TYPE_AUDIO) { |
| mime = "audio/"; |
| } else if (trackType == MEDIA_TRACK_TYPE_VIDEO) { |
| mime = "video/"; |
| } else { |
| ALOGE("unknown track type: %d", trackType); |
| return; |
| } |
| } |
| |
| AString lang; |
| if (!format->findString("language", &lang)) { |
| ALOGE("no language"); |
| return; |
| } |
| |
| reply->writeInt32(2); // write something non-zero |
| reply->writeInt32(trackType); |
| reply->writeString16(String16(mime.c_str())); |
| reply->writeString16(String16(lang.c_str())); |
| |
| if (trackType == MEDIA_TRACK_TYPE_SUBTITLE) { |
| int32_t isAuto, isDefault, isForced; |
| CHECK(format->findInt32("auto", &isAuto)); |
| CHECK(format->findInt32("default", &isDefault)); |
| CHECK(format->findInt32("forced", &isForced)); |
| |
| reply->writeInt32(isAuto); |
| reply->writeInt32(isDefault); |
| reply->writeInt32(isForced); |
| } else if (trackType == MEDIA_TRACK_TYPE_AUDIO) { |
| int32_t hapticChannelCount; |
| bool hasHapticChannels = format->findInt32("haptic-channel-count", &hapticChannelCount); |
| reply->writeInt32(hasHapticChannels); |
| if (hasHapticChannels) { |
| reply->writeInt32(hapticChannelCount); |
| } |
| } |
| } |
| |
| void NuPlayer::onMessageReceived(const sp<AMessage> &msg) { |
| switch (msg->what()) { |
| case kWhatSetDataSource: |
| { |
| ALOGV("kWhatSetDataSource"); |
| |
| CHECK(mSource == NULL); |
| |
| status_t err = OK; |
| sp<RefBase> obj; |
| CHECK(msg->findObject("source", &obj)); |
| if (obj != NULL) { |
| Mutex::Autolock autoLock(mSourceLock); |
| mSource = static_cast<Source *>(obj.get()); |
| } else { |
| err = UNKNOWN_ERROR; |
| } |
| |
| CHECK(mDriver != NULL); |
| sp<NuPlayerDriver> driver = mDriver.promote(); |
| if (driver != NULL) { |
| driver->notifySetDataSourceCompleted(err); |
| } |
| break; |
| } |
| |
| case kWhatGetBufferingSettings: |
| { |
| sp<AReplyToken> replyID; |
| CHECK(msg->senderAwaitsResponse(&replyID)); |
| |
| ALOGV("kWhatGetBufferingSettings"); |
| BufferingSettings buffering; |
| status_t err = OK; |
| if (mSource != NULL) { |
| err = mSource->getBufferingSettings(&buffering); |
| } else { |
| err = INVALID_OPERATION; |
| } |
| sp<AMessage> response = new AMessage; |
| if (err == OK) { |
| writeToAMessage(response, buffering); |
| } |
| response->setInt32("err", err); |
| response->postReply(replyID); |
| break; |
| } |
| |
| case kWhatSetBufferingSettings: |
| { |
| sp<AReplyToken> replyID; |
| CHECK(msg->senderAwaitsResponse(&replyID)); |
| |
| ALOGV("kWhatSetBufferingSettings"); |
| BufferingSettings buffering; |
| readFromAMessage(msg, &buffering); |
| status_t err = OK; |
| if (mSource != NULL) { |
| err = mSource->setBufferingSettings(buffering); |
| } else { |
| err = INVALID_OPERATION; |
| } |
| sp<AMessage> response = new AMessage; |
| response->setInt32("err", err); |
| response->postReply(replyID); |
| break; |
| } |
| |
| case kWhatPrepare: |
| { |
| ALOGV("onMessageReceived kWhatPrepare"); |
| |
| mSource->prepareAsync(); |
| break; |
| } |
| |
| case kWhatGetTrackInfo: |
| { |
| sp<AReplyToken> replyID; |
| CHECK(msg->senderAwaitsResponse(&replyID)); |
| |
| Parcel* reply; |
| CHECK(msg->findPointer("reply", (void**)&reply)); |
| |
| size_t inbandTracks = 0; |
| if (mSource != NULL) { |
| inbandTracks = mSource->getTrackCount(); |
| } |
| |
| size_t ccTracks = 0; |
| if (mCCDecoder != NULL) { |
| ccTracks = mCCDecoder->getTrackCount(); |
| } |
| |
| // total track count |
| reply->writeInt32(inbandTracks + ccTracks); |
| |
| // write inband tracks |
| for (size_t i = 0; i < inbandTracks; ++i) { |
| writeTrackInfo(reply, mSource->getTrackInfo(i)); |
| } |
| |
| // write CC track |
| for (size_t i = 0; i < ccTracks; ++i) { |
| writeTrackInfo(reply, mCCDecoder->getTrackInfo(i)); |
| } |
| |
| sp<AMessage> response = new AMessage; |
| response->postReply(replyID); |
| break; |
| } |
| |
| case kWhatGetSelectedTrack: |
| { |
| int32_t type32; |
| CHECK(msg->findInt32("type", (int32_t*)&type32)); |
| media_track_type type = (media_track_type)type32; |
| |
| size_t inbandTracks = 0; |
| status_t err = INVALID_OPERATION; |
| ssize_t selectedTrack = -1; |
| if (mSource != NULL) { |
| err = OK; |
| inbandTracks = mSource->getTrackCount(); |
| selectedTrack = mSource->getSelectedTrack(type); |
| } |
| |
| if (selectedTrack == -1 && mCCDecoder != NULL) { |
| err = OK; |
| selectedTrack = mCCDecoder->getSelectedTrack(type); |
| if (selectedTrack != -1) { |
| selectedTrack += inbandTracks; |
| } |
| } |
| |
| Parcel* reply; |
| CHECK(msg->findPointer("reply", (void**)&reply)); |
| reply->writeInt32(selectedTrack); |
| |
| sp<AMessage> response = new AMessage; |
| response->setInt32("err", err); |
| |
| sp<AReplyToken> replyID; |
| CHECK(msg->senderAwaitsResponse(&replyID)); |
| response->postReply(replyID); |
| break; |
| } |
| |
| case kWhatSelectTrack: |
| { |
| sp<AReplyToken> replyID; |
| CHECK(msg->senderAwaitsResponse(&replyID)); |
| |
| size_t trackIndex; |
| int32_t select; |
| int64_t timeUs; |
| CHECK(msg->findSize("trackIndex", &trackIndex)); |
| CHECK(msg->findInt32("select", &select)); |
| CHECK(msg->findInt64("timeUs", &timeUs)); |
| |
| status_t err = INVALID_OPERATION; |
| |
| size_t inbandTracks = 0; |
| if (mSource != NULL) { |
| inbandTracks = mSource->getTrackCount(); |
| } |
| size_t ccTracks = 0; |
| if (mCCDecoder != NULL) { |
| ccTracks = mCCDecoder->getTrackCount(); |
| } |
| |
| if (trackIndex < inbandTracks) { |
| err = mSource->selectTrack(trackIndex, select, timeUs); |
| |
| if (!select && err == OK) { |
| int32_t type; |
| sp<AMessage> info = mSource->getTrackInfo(trackIndex); |
| if (info != NULL |
| && info->findInt32("type", &type) |
| && type == MEDIA_TRACK_TYPE_TIMEDTEXT) { |
| ++mTimedTextGeneration; |
| } |
| } |
| } else { |
| trackIndex -= inbandTracks; |
| |
| if (trackIndex < ccTracks) { |
| err = mCCDecoder->selectTrack(trackIndex, select); |
| } |
| } |
| |
| sp<AMessage> response = new AMessage; |
| response->setInt32("err", err); |
| |
| response->postReply(replyID); |
| break; |
| } |
| |
| case kWhatPollDuration: |
| { |
| int32_t generation; |
| CHECK(msg->findInt32("generation", &generation)); |
| |
| if (generation != mPollDurationGeneration) { |
| // stale |
| break; |
| } |
| |
| int64_t durationUs; |
| if (mDriver != NULL && mSource->getDuration(&durationUs) == OK) { |
| sp<NuPlayerDriver> driver = mDriver.promote(); |
| if (driver != NULL) { |
| driver->notifyDuration(durationUs); |
| } |
| } |
| |
| msg->post(1000000LL); // poll again in a second. |
| break; |
| } |
| |
| case kWhatSetVideoSurface: |
| { |
| |
| sp<RefBase> obj; |
| CHECK(msg->findObject("surface", &obj)); |
| sp<Surface> surface = static_cast<Surface *>(obj.get()); |
| |
| ALOGD("onSetVideoSurface(%p, %s video decoder)", |
| surface.get(), |
| (mSource != NULL && mStarted && mSource->getFormat(false /* audio */) != NULL |
| && mVideoDecoder != NULL) ? "have" : "no"); |
| |
| // Need to check mStarted before calling mSource->getFormat because NuPlayer might |
| // be in preparing state and it could take long time. |
| // When mStarted is true, mSource must have been set. |
| if (mSource == NULL || !mStarted || mSource->getFormat(false /* audio */) == NULL |
| // NOTE: mVideoDecoder's mSurface is always non-null |
| || (mVideoDecoder != NULL && mVideoDecoder->setVideoSurface(surface) == OK)) { |
| performSetSurface(surface); |
| break; |
| } |
| |
| mDeferredActions.push_back( |
| new FlushDecoderAction( |
| (obj != NULL ? FLUSH_CMD_FLUSH : FLUSH_CMD_NONE) /* audio */, |
| FLUSH_CMD_SHUTDOWN /* video */)); |
| |
| mDeferredActions.push_back(new SetSurfaceAction(surface)); |
| |
| if (obj != NULL) { |
| if (mStarted) { |
| // Issue a seek to refresh the video screen only if started otherwise |
| // the extractor may not yet be started and will assert. |
| // If the video decoder is not set (perhaps audio only in this case) |
| // do not perform a seek as it is not needed. |
| int64_t currentPositionUs = 0; |
| if (getCurrentPosition(¤tPositionUs) == OK) { |
| mDeferredActions.push_back( |
| new SeekAction(currentPositionUs, |
| MediaPlayerSeekMode::SEEK_PREVIOUS_SYNC /* mode */)); |
| } |
| } |
| |
| // If there is a new surface texture, instantiate decoders |
| // again if possible. |
| mDeferredActions.push_back( |
| new SimpleAction(&NuPlayer::performScanSources)); |
| |
| // After a flush without shutdown, decoder is paused. |
| // Don't resume it until source seek is done, otherwise it could |
| // start pulling stale data too soon. |
| mDeferredActions.push_back( |
| new ResumeDecoderAction(false /* needNotify */)); |
| } |
| |
| processDeferredActions(); |
| break; |
| } |
| |
| case kWhatSetAudioSink: |
| { |
| ALOGV("kWhatSetAudioSink"); |
| |
| sp<RefBase> obj; |
| CHECK(msg->findObject("sink", &obj)); |
| |
| mAudioSink = static_cast<MediaPlayerBase::AudioSink *>(obj.get()); |
| break; |
| } |
| |
| case kWhatStart: |
| { |
| ALOGV("kWhatStart"); |
| if (mStarted) { |
| // do not resume yet if the source is still buffering |
| if (!mPausedForBuffering) { |
| onResume(); |
| } |
| } else { |
| onStart(); |
| } |
| mPausedByClient = false; |
| break; |
| } |
| |
| case kWhatConfigPlayback: |
| { |
| sp<AReplyToken> replyID; |
| CHECK(msg->senderAwaitsResponse(&replyID)); |
| AudioPlaybackRate rate /* sanitized */; |
| readFromAMessage(msg, &rate); |
| status_t err = OK; |
| if (mRenderer != NULL) { |
| // AudioSink allows only 1.f and 0.f for offload and direct modes. |
| // For other speeds, restart audio to fallback to supported paths |
| bool audioDirectOutput = (mAudioSink->getFlags() & AUDIO_OUTPUT_FLAG_DIRECT) != 0; |
| if ((mOffloadAudio || audioDirectOutput) && |
| ((rate.mSpeed != 0.f && rate.mSpeed != 1.f) || rate.mPitch != 1.f)) { |
| |
| int64_t currentPositionUs; |
| if (getCurrentPosition(¤tPositionUs) != OK) { |
| currentPositionUs = mPreviousSeekTimeUs; |
| } |
| |
| // Set mPlaybackSettings so that the new audio decoder can |
| // be created correctly. |
| mPlaybackSettings = rate; |
| if (!mPaused) { |
| mRenderer->pause(); |
| } |
| restartAudio( |
| currentPositionUs, true /* forceNonOffload */, |
| true /* needsToCreateAudioDecoder */); |
| if (!mPaused) { |
| mRenderer->resume(); |
| } |
| } |
| |
| err = mRenderer->setPlaybackSettings(rate); |
| } |
| if (err == OK) { |
| if (rate.mSpeed == 0.f) { |
| onPause(); |
| mPausedByClient = true; |
| // save all other settings (using non-paused speed) |
| // so we can restore them on start |
| AudioPlaybackRate newRate = rate; |
| newRate.mSpeed = mPlaybackSettings.mSpeed; |
| mPlaybackSettings = newRate; |
| } else { /* rate.mSpeed != 0.f */ |
| mPlaybackSettings = rate; |
| if (mStarted) { |
| // do not resume yet if the source is still buffering |
| if (!mPausedForBuffering) { |
| onResume(); |
| } |
| } else if (mPrepared) { |
| onStart(); |
| } |
| |
| mPausedByClient = false; |
| } |
| } |
| |
| if (mVideoDecoder != NULL) { |
| sp<AMessage> params = new AMessage(); |
| params->setFloat("playback-speed", mPlaybackSettings.mSpeed); |
| mVideoDecoder->setParameters(params); |
| } |
| |
| sp<AMessage> response = new AMessage; |
| response->setInt32("err", err); |
| response->postReply(replyID); |
| break; |
| } |
| |
| case kWhatGetPlaybackSettings: |
| { |
| sp<AReplyToken> replyID; |
| CHECK(msg->senderAwaitsResponse(&replyID)); |
| AudioPlaybackRate rate = mPlaybackSettings; |
| status_t err = OK; |
| if (mRenderer != NULL) { |
| err = mRenderer->getPlaybackSettings(&rate); |
| } |
| if (err == OK) { |
| // get playback settings used by renderer, as it may be |
| // slightly off due to audiosink not taking small changes. |
| mPlaybackSettings = rate; |
| if (mPaused) { |
| rate.mSpeed = 0.f; |
| } |
| } |
| sp<AMessage> response = new AMessage; |
| if (err == OK) { |
| writeToAMessage(response, rate); |
| } |
| response->setInt32("err", err); |
| response->postReply(replyID); |
| break; |
| } |
| |
| case kWhatConfigSync: |
| { |
| sp<AReplyToken> replyID; |
| CHECK(msg->senderAwaitsResponse(&replyID)); |
| |
| ALOGV("kWhatConfigSync"); |
| AVSyncSettings sync; |
| float videoFpsHint; |
| readFromAMessage(msg, &sync, &videoFpsHint); |
| status_t err = OK; |
| if (mRenderer != NULL) { |
| err = mRenderer->setSyncSettings(sync, videoFpsHint); |
| } |
| if (err == OK) { |
| mSyncSettings = sync; |
| mVideoFpsHint = videoFpsHint; |
| } |
| sp<AMessage> response = new AMessage; |
| response->setInt32("err", err); |
| response->postReply(replyID); |
| break; |
| } |
| |
| case kWhatGetSyncSettings: |
| { |
| sp<AReplyToken> replyID; |
| CHECK(msg->senderAwaitsResponse(&replyID)); |
| AVSyncSettings sync = mSyncSettings; |
| float videoFps = mVideoFpsHint; |
| status_t err = OK; |
| if (mRenderer != NULL) { |
| err = mRenderer->getSyncSettings(&sync, &videoFps); |
| if (err == OK) { |
| mSyncSettings = sync; |
| mVideoFpsHint = videoFps; |
| } |
| } |
| sp<AMessage> response = new AMessage; |
| if (err == OK) { |
| writeToAMessage(response, sync, videoFps); |
| } |
| response->setInt32("err", err); |
| response->postReply(replyID); |
| break; |
| } |
| |
| case kWhatScanSources: |
| { |
| int32_t generation; |
| CHECK(msg->findInt32("generation", &generation)); |
| if (generation != mScanSourcesGeneration) { |
| // Drop obsolete msg. |
| break; |
| } |
| |
| mScanSourcesPending = false; |
| |
| ALOGV("scanning sources haveAudio=%d, haveVideo=%d", |
| mAudioDecoder != NULL, mVideoDecoder != NULL); |
| |
| bool mHadAnySourcesBefore = |
| (mAudioDecoder != NULL) || (mVideoDecoder != NULL); |
| bool rescan = false; |
| |
| // initialize video before audio because successful initialization of |
| // video may change deep buffer mode of audio. |
| if (mSurface != NULL) { |
| if (instantiateDecoder(false, &mVideoDecoder) == -EWOULDBLOCK) { |
| rescan = true; |
| } |
| } |
| |
| // Don't try to re-open audio sink if there's an existing decoder. |
| if (mAudioSink != NULL && mAudioDecoder == NULL) { |
| if (instantiateDecoder(true, &mAudioDecoder) == -EWOULDBLOCK) { |
| rescan = true; |
| } |
| } |
| |
| if (!mHadAnySourcesBefore |
| && (mAudioDecoder != NULL || mVideoDecoder != NULL)) { |
| // This is the first time we've found anything playable. |
| |
| if (mSourceFlags & Source::FLAG_DYNAMIC_DURATION) { |
| schedulePollDuration(); |
| } |
| } |
| |
| status_t err; |
| if ((err = mSource->feedMoreTSData()) != OK) { |
| if (mAudioDecoder == NULL && mVideoDecoder == NULL) { |
| // We're not currently decoding anything (no audio or |
| // video tracks found) and we just ran out of input data. |
| |
| if (err == ERROR_END_OF_STREAM) { |
| notifyListener(MEDIA_PLAYBACK_COMPLETE, 0, 0); |
| } else { |
| notifyListener(MEDIA_ERROR, MEDIA_ERROR_UNKNOWN, err); |
| } |
| } |
| break; |
| } |
| |
| if (rescan) { |
| msg->post(100000LL); |
| mScanSourcesPending = true; |
| } |
| break; |
| } |
| |
| case kWhatVideoNotify: |
| case kWhatAudioNotify: |
| { |
| bool audio = msg->what() == kWhatAudioNotify; |
| |
| int32_t currentDecoderGeneration = |
| (audio? mAudioDecoderGeneration : mVideoDecoderGeneration); |
| int32_t requesterGeneration = currentDecoderGeneration - 1; |
| CHECK(msg->findInt32("generation", &requesterGeneration)); |
| |
| if (requesterGeneration != currentDecoderGeneration) { |
| ALOGV("got message from old %s decoder, generation(%d:%d)", |
| audio ? "audio" : "video", requesterGeneration, |
| currentDecoderGeneration); |
| sp<AMessage> reply; |
| if (!(msg->findMessage("reply", &reply))) { |
| return; |
| } |
| |
| reply->setInt32("err", INFO_DISCONTINUITY); |
| reply->post(); |
| return; |
| } |
| |
| int32_t what; |
| CHECK(msg->findInt32("what", &what)); |
| |
| if (what == DecoderBase::kWhatInputDiscontinuity) { |
| int32_t formatChange; |
| CHECK(msg->findInt32("formatChange", &formatChange)); |
| |
| ALOGV("%s discontinuity: formatChange %d", |
| audio ? "audio" : "video", formatChange); |
| |
| if (formatChange) { |
| mDeferredActions.push_back( |
| new FlushDecoderAction( |
| audio ? FLUSH_CMD_SHUTDOWN : FLUSH_CMD_NONE, |
| audio ? FLUSH_CMD_NONE : FLUSH_CMD_SHUTDOWN)); |
| } |
| |
| mDeferredActions.push_back( |
| new SimpleAction( |
| &NuPlayer::performScanSources)); |
| |
| processDeferredActions(); |
| } else if (what == DecoderBase::kWhatEOS) { |
| int32_t err; |
| CHECK(msg->findInt32("err", &err)); |
| |
| if (err == ERROR_END_OF_STREAM) { |
| ALOGV("got %s decoder EOS", audio ? "audio" : "video"); |
| } else { |
| ALOGV("got %s decoder EOS w/ error %d", |
| audio ? "audio" : "video", |
| err); |
| } |
| |
| mRenderer->queueEOS(audio, err); |
| } else if (what == DecoderBase::kWhatFlushCompleted) { |
| ALOGV("decoder %s flush completed", audio ? "audio" : "video"); |
| |
| handleFlushComplete(audio, true /* isDecoder */); |
| finishFlushIfPossible(); |
| } else if (what == DecoderBase::kWhatVideoSizeChanged) { |
| sp<AMessage> format; |
| CHECK(msg->findMessage("format", &format)); |
| |
| sp<AMessage> inputFormat = |
| mSource->getFormat(false /* audio */); |
| |
| setVideoScalingMode(mVideoScalingMode); |
| updateVideoSize(inputFormat, format); |
| } else if (what == DecoderBase::kWhatShutdownCompleted) { |
| ALOGV("%s shutdown completed", audio ? "audio" : "video"); |
| if (audio) { |
| Mutex::Autolock autoLock(mDecoderLock); |
| mAudioDecoder.clear(); |
| mAudioDecoderError = false; |
| ++mAudioDecoderGeneration; |
| |
| CHECK_EQ((int)mFlushingAudio, (int)SHUTTING_DOWN_DECODER); |
| mFlushingAudio = SHUT_DOWN; |
| } else { |
| Mutex::Autolock autoLock(mDecoderLock); |
| mVideoDecoder.clear(); |
| mVideoDecoderError = false; |
| ++mVideoDecoderGeneration; |
| |
| CHECK_EQ((int)mFlushingVideo, (int)SHUTTING_DOWN_DECODER); |
| mFlushingVideo = SHUT_DOWN; |
| } |
| |
| finishFlushIfPossible(); |
| } else if (what == DecoderBase::kWhatResumeCompleted) { |
| finishResume(); |
| } else if (what == DecoderBase::kWhatError) { |
| status_t err; |
| if (!msg->findInt32("err", &err) || err == OK) { |
| err = UNKNOWN_ERROR; |
| } |
| |
| // Decoder errors can be due to Source (e.g. from streaming), |
| // or from decoding corrupted bitstreams, or from other decoder |
| // MediaCodec operations (e.g. from an ongoing reset or seek). |
| // They may also be due to openAudioSink failure at |
| // decoder start or after a format change. |
| // |
| // We try to gracefully shut down the affected decoder if possible, |
| // rather than trying to force the shutdown with something |
| // similar to performReset(). This method can lead to a hang |
| // if MediaCodec functions block after an error, but they should |
| // typically return INVALID_OPERATION instead of blocking. |
| |
| FlushStatus *flushing = audio ? &mFlushingAudio : &mFlushingVideo; |
| ALOGE("received error(%#x) from %s decoder, flushing(%d), now shutting down", |
| err, audio ? "audio" : "video", *flushing); |
| |
| switch (*flushing) { |
| case NONE: |
| mDeferredActions.push_back( |
| new FlushDecoderAction( |
| audio ? FLUSH_CMD_SHUTDOWN : FLUSH_CMD_NONE, |
| audio ? FLUSH_CMD_NONE : FLUSH_CMD_SHUTDOWN)); |
| processDeferredActions(); |
| break; |
| case FLUSHING_DECODER: |
| *flushing = FLUSHING_DECODER_SHUTDOWN; // initiate shutdown after flush. |
| break; // Wait for flush to complete. |
| case FLUSHING_DECODER_SHUTDOWN: |
| break; // Wait for flush to complete. |
| case SHUTTING_DOWN_DECODER: |
| break; // Wait for shutdown to complete. |
| case FLUSHED: |
| getDecoder(audio)->initiateShutdown(); // In the middle of a seek. |
| *flushing = SHUTTING_DOWN_DECODER; // Shut down. |
| break; |
| case SHUT_DOWN: |
| finishFlushIfPossible(); // Should not occur. |
| break; // Finish anyways. |
| } |
| if (mSource != nullptr) { |
| if (audio) { |
| if (mVideoDecoderError || mSource->getFormat(false /* audio */) == NULL |
| || mSurface == NULL || mVideoDecoder == NULL) { |
| // When both audio and video have error, or this stream has only audio |
| // which has error, notify client of error. |
| notifyListener(MEDIA_ERROR, MEDIA_ERROR_UNKNOWN, err); |
| } else { |
| // Only audio track has error. Video track could be still good to play. |
| if (mVideoEOS) { |
| notifyListener(MEDIA_PLAYBACK_COMPLETE, 0, 0); |
| } else { |
| notifyListener(MEDIA_INFO, MEDIA_INFO_PLAY_AUDIO_ERROR, err); |
| } |
| } |
| mAudioDecoderError = true; |
| } else { |
| if (mAudioDecoderError || mSource->getFormat(true /* audio */) == NULL |
| || mAudioSink == NULL || mAudioDecoder == NULL) { |
| // When both audio and video have error, or this stream has only video |
| // which has error, notify client of error. |
| notifyListener(MEDIA_ERROR, MEDIA_ERROR_UNKNOWN, err); |
| } else { |
| // Only video track has error. Audio track could be still good to play. |
| if (mAudioEOS) { |
| notifyListener(MEDIA_PLAYBACK_COMPLETE, 0, 0); |
| } else { |
| notifyListener(MEDIA_INFO, MEDIA_INFO_PLAY_VIDEO_ERROR, err); |
| } |
| } |
| mVideoDecoderError = true; |
| } |
| } |
| } else { |
| ALOGV("Unhandled decoder notification %d '%c%c%c%c'.", |
| what, |
| what >> 24, |
| (what >> 16) & 0xff, |
| (what >> 8) & 0xff, |
| what & 0xff); |
| } |
| |
| break; |
| } |
| |
| case kWhatRendererNotify: |
| { |
| int32_t requesterGeneration = mRendererGeneration - 1; |
| CHECK(msg->findInt32("generation", &requesterGeneration)); |
| if (requesterGeneration != mRendererGeneration) { |
| ALOGV("got message from old renderer, generation(%d:%d)", |
| requesterGeneration, mRendererGeneration); |
| return; |
| } |
| |
| int32_t what; |
| CHECK(msg->findInt32("what", &what)); |
| |
| if (what == Renderer::kWhatEOS) { |
| int32_t audio; |
| CHECK(msg->findInt32("audio", &audio)); |
| |
| int32_t finalResult; |
| CHECK(msg->findInt32("finalResult", &finalResult)); |
| |
| if (audio) { |
| mAudioEOS = true; |
| } else { |
| mVideoEOS = true; |
| } |
| |
| if (finalResult == ERROR_END_OF_STREAM) { |
| ALOGV("reached %s EOS", audio ? "audio" : "video"); |
| } else { |
| ALOGE("%s track encountered an error (%d)", |
| audio ? "audio" : "video", finalResult); |
| |
| notifyListener( |
| MEDIA_ERROR, MEDIA_ERROR_UNKNOWN, finalResult); |
| } |
| |
| if ((mAudioEOS || mAudioDecoder == NULL) |
| && (mVideoEOS || mVideoDecoder == NULL)) { |
| notifyListener(MEDIA_PLAYBACK_COMPLETE, 0, 0); |
| } |
| } else if (what == Renderer::kWhatFlushComplete) { |
| int32_t audio; |
| CHECK(msg->findInt32("audio", &audio)); |
| |
| if (audio) { |
| mAudioEOS = false; |
| } else { |
| mVideoEOS = false; |
| } |
| |
| ALOGV("renderer %s flush completed.", audio ? "audio" : "video"); |
| if (audio && (mFlushingAudio == NONE || mFlushingAudio == FLUSHED |
| || mFlushingAudio == SHUT_DOWN)) { |
| // Flush has been handled by tear down. |
| break; |
| } |
| handleFlushComplete(audio, false /* isDecoder */); |
| finishFlushIfPossible(); |
| } else if (what == Renderer::kWhatVideoRenderingStart) { |
| notifyListener(MEDIA_INFO, MEDIA_INFO_RENDERING_START, 0); |
| } else if (what == Renderer::kWhatMediaRenderingStart) { |
| ALOGV("media rendering started"); |
| notifyListener(MEDIA_STARTED, 0, 0); |
| } else if (what == Renderer::kWhatAudioTearDown) { |
| int32_t reason; |
| CHECK(msg->findInt32("reason", &reason)); |
| ALOGV("Tear down audio with reason %d.", reason); |
| if (reason == Renderer::kDueToTimeout && !(mPaused && mOffloadAudio)) { |
| // TimeoutWhenPaused is only for offload mode. |
| ALOGW("Received a stale message for teardown, mPaused(%d), mOffloadAudio(%d)", |
| mPaused, mOffloadAudio); |
| break; |
| } |
| int64_t positionUs; |
| if (!msg->findInt64("positionUs", &positionUs)) { |
| positionUs = mPreviousSeekTimeUs; |
| } |
| |
| restartAudio( |
| positionUs, reason == Renderer::kForceNonOffload /* forceNonOffload */, |
| reason != Renderer::kDueToTimeout /* needsToCreateAudioDecoder */); |
| } |
| break; |
| } |
| |
| case kWhatMoreDataQueued: |
| { |
| break; |
| } |
| |
| case kWhatReset: |
| { |
| ALOGV("kWhatReset"); |
| |
| mResetting = true; |
| updatePlaybackTimer(true /* stopping */, "kWhatReset"); |
| updateRebufferingTimer(true /* stopping */, true /* exiting */); |
| |
| mDeferredActions.push_back( |
| new FlushDecoderAction( |
| FLUSH_CMD_SHUTDOWN /* audio */, |
| FLUSH_CMD_SHUTDOWN /* video */)); |
| |
| mDeferredActions.push_back( |
| new SimpleAction(&NuPlayer::performReset)); |
| |
| processDeferredActions(); |
| break; |
| } |
| |
| case kWhatNotifyTime: |
| { |
| ALOGV("kWhatNotifyTime"); |
| int64_t timerUs; |
| CHECK(msg->findInt64("timerUs", &timerUs)); |
| |
| notifyListener(MEDIA_NOTIFY_TIME, timerUs, 0); |
| break; |
| } |
| |
| case kWhatSeek: |
| { |
| int64_t seekTimeUs; |
| int32_t mode; |
| int32_t needNotify; |
| CHECK(msg->findInt64("seekTimeUs", &seekTimeUs)); |
| CHECK(msg->findInt32("mode", &mode)); |
| CHECK(msg->findInt32("needNotify", &needNotify)); |
| |
| ALOGV("kWhatSeek seekTimeUs=%lld us, mode=%d, needNotify=%d", |
| (long long)seekTimeUs, mode, needNotify); |
| |
| if (!mStarted) { |
| // Seek before the player is started. In order to preview video, |
| // need to start the player and pause it. This branch is called |
| // only once if needed. After the player is started, any seek |
| // operation will go through normal path. |
| // Audio-only cases are handled separately. |
| onStart(seekTimeUs, (MediaPlayerSeekMode)mode); |
| if (mStarted) { |
| onPause(); |
| mPausedByClient = true; |
| } |
| if (needNotify) { |
| notifyDriverSeekComplete(); |
| } |
| break; |
| } |
| |
| mDeferredActions.push_back( |
| new FlushDecoderAction(FLUSH_CMD_FLUSH /* audio */, |
| FLUSH_CMD_FLUSH /* video */)); |
| |
| mDeferredActions.push_back( |
| new SeekAction(seekTimeUs, (MediaPlayerSeekMode)mode)); |
| |
| // After a flush without shutdown, decoder is paused. |
| // Don't resume it until source seek is done, otherwise it could |
| // start pulling stale data too soon. |
| mDeferredActions.push_back( |
| new ResumeDecoderAction(needNotify)); |
| |
| processDeferredActions(); |
| break; |
| } |
| |
| case kWhatPause: |
| { |
| onPause(); |
| mPausedByClient = true; |
| break; |
| } |
| |
| case kWhatSourceNotify: |
| { |
| onSourceNotify(msg); |
| break; |
| } |
| |
| case kWhatClosedCaptionNotify: |
| { |
| onClosedCaptionNotify(msg); |
| break; |
| } |
| |
| case kWhatPrepareDrm: |
| { |
| status_t status = onPrepareDrm(msg); |
| |
| sp<AMessage> response = new AMessage; |
| response->setInt32("status", status); |
| sp<AReplyToken> replyID; |
| CHECK(msg->senderAwaitsResponse(&replyID)); |
| response->postReply(replyID); |
| break; |
| } |
| |
| case kWhatReleaseDrm: |
| { |
| status_t status = onReleaseDrm(); |
| |
| sp<AMessage> response = new AMessage; |
| response->setInt32("status", status); |
| sp<AReplyToken> replyID; |
| CHECK(msg->senderAwaitsResponse(&replyID)); |
| response->postReply(replyID); |
| break; |
| } |
| |
| case kWhatMediaClockNotify: |
| { |
| ALOGV("kWhatMediaClockNotify"); |
| int64_t anchorMediaUs, anchorRealUs; |
| float playbackRate; |
| CHECK(msg->findInt64("anchor-media-us", &anchorMediaUs)); |
| CHECK(msg->findInt64("anchor-real-us", &anchorRealUs)); |
| CHECK(msg->findFloat("playback-rate", &playbackRate)); |
| |
| Parcel in; |
| in.writeInt64(anchorMediaUs); |
| in.writeInt64(anchorRealUs); |
| in.writeFloat(playbackRate); |
| |
| notifyListener(MEDIA_TIME_DISCONTINUITY, 0, 0, &in); |
| break; |
| } |
| |
| default: |
| TRESPASS(); |
| break; |
| } |
| } |
| |
| void NuPlayer::onResume() { |
| if (!mPaused || mResetting) { |
| ALOGD_IF(mResetting, "resetting, onResume discarded"); |
| return; |
| } |
| mPaused = false; |
| if (mSource != NULL) { |
| mSource->resume(); |
| } else { |
| ALOGW("resume called when source is gone or not set"); |
| } |
| // |mAudioDecoder| may have been released due to the pause timeout, so re-create it if |
| // needed. |
| if (audioDecoderStillNeeded() && mAudioDecoder == NULL) { |
| instantiateDecoder(true /* audio */, &mAudioDecoder); |
| } |
| if (mRenderer != NULL) { |
| mRenderer->resume(); |
| } else { |
| ALOGW("resume called when renderer is gone or not set"); |
| } |
| |
| startPlaybackTimer("onresume"); |
| } |
| |
| status_t NuPlayer::onInstantiateSecureDecoders() { |
| status_t err; |
| if (!(mSourceFlags & Source::FLAG_SECURE)) { |
| return BAD_TYPE; |
| } |
| |
| if (mRenderer != NULL) { |
| ALOGE("renderer should not be set when instantiating secure decoders"); |
| return UNKNOWN_ERROR; |
| } |
| |
| // TRICKY: We rely on mRenderer being null, so that decoder does not start requesting |
| // data on instantiation. |
| if (mSurface != NULL) { |
| err = instantiateDecoder(false, &mVideoDecoder); |
| if (err != OK) { |
| return err; |
| } |
| } |
| |
| if (mAudioSink != NULL) { |
| err = instantiateDecoder(true, &mAudioDecoder); |
| if (err != OK) { |
| return err; |
| } |
| } |
| return OK; |
| } |
| |
| void NuPlayer::onStart(int64_t startPositionUs, MediaPlayerSeekMode mode) { |
| ALOGV("onStart: mCrypto: %p (%d)", mCrypto.get(), |
| (mCrypto != NULL ? mCrypto->getStrongCount() : 0)); |
| |
| if (!mSourceStarted) { |
| mSourceStarted = true; |
| mSource->start(); |
| } |
| if (startPositionUs > 0) { |
| performSeek(startPositionUs, mode); |
| if (mSource->getFormat(false /* audio */) == NULL) { |
| return; |
| } |
| } |
| |
| mOffloadAudio = false; |
| mAudioEOS = false; |
| mVideoEOS = false; |
| mStarted = true; |
| mPaused = false; |
| |
| uint32_t flags = 0; |
| |
| if (mSource->isRealTime()) { |
| flags |= Renderer::FLAG_REAL_TIME; |
| } |
| |
| bool hasAudio = (mSource->getFormat(true /* audio */) != NULL); |
| bool hasVideo = (mSource->getFormat(false /* audio */) != NULL); |
| if (!hasAudio && !hasVideo) { |
| ALOGE("no metadata for either audio or video source"); |
| mSource->stop(); |
| mSourceStarted = false; |
| notifyListener(MEDIA_ERROR, MEDIA_ERROR_UNKNOWN, ERROR_MALFORMED); |
| return; |
| } |
| ALOGV_IF(!hasAudio, "no metadata for audio source"); // video only stream |
| |
| sp<MetaData> audioMeta = mSource->getFormatMeta(true /* audio */); |
| |
| audio_stream_type_t streamType = AUDIO_STREAM_MUSIC; |
| if (mAudioSink != NULL) { |
| streamType = mAudioSink->getAudioStreamType(); |
| } |
| |
| mOffloadAudio = |
| canOffloadStream(audioMeta, hasVideo, mSource->isStreaming(), streamType) |
| && (mPlaybackSettings.mSpeed == 1.f && mPlaybackSettings.mPitch == 1.f); |
| |
| // Modular DRM: Disabling audio offload if the source is protected |
| if (mOffloadAudio && mIsDrmProtected) { |
| mOffloadAudio = false; |
| ALOGV("onStart: Disabling mOffloadAudio now that the source is protected."); |
| } |
| |
| if (mOffloadAudio) { |
| flags |= Renderer::FLAG_OFFLOAD_AUDIO; |
| } |
| |
| sp<AMessage> notify = new AMessage(kWhatRendererNotify, this); |
| ++mRendererGeneration; |
| notify->setInt32("generation", mRendererGeneration); |
| mRenderer = new Renderer(mAudioSink, mMediaClock, notify, flags); |
| mRendererLooper = new ALooper; |
| mRendererLooper->setName("NuPlayerRenderer"); |
| mRendererLooper->start(false, false, ANDROID_PRIORITY_AUDIO); |
| mRendererLooper->registerHandler(mRenderer); |
| |
| status_t err = mRenderer->setPlaybackSettings(mPlaybackSettings); |
| if (err != OK) { |
| mSource->stop(); |
| mSourceStarted = false; |
| notifyListener(MEDIA_ERROR, MEDIA_ERROR_UNKNOWN, err); |
| return; |
| } |
| |
| float rate = getFrameRate(); |
| if (rate > 0) { |
| mRenderer->setVideoFrameRate(rate); |
| } |
| |
| if (mVideoDecoder != NULL) { |
| mVideoDecoder->setRenderer(mRenderer); |
| } |
| if (mAudioDecoder != NULL) { |
| mAudioDecoder->setRenderer(mRenderer); |
| } |
| |
| startPlaybackTimer("onstart"); |
| |
| postScanSources(); |
| } |
| |
| void NuPlayer::startPlaybackTimer(const char *where) { |
| Mutex::Autolock autoLock(mPlayingTimeLock); |
| if (mLastStartedPlayingTimeNs == 0) { |
| mLastStartedPlayingTimeNs = systemTime(); |
| ALOGV("startPlaybackTimer() time %20" PRId64 " (%s)", mLastStartedPlayingTimeNs, where); |
| } |
| } |
| |
| void NuPlayer::updatePlaybackTimer(bool stopping, const char *where) { |
| Mutex::Autolock autoLock(mPlayingTimeLock); |
| |
| ALOGV("updatePlaybackTimer(%s) time %20" PRId64 " (%s)", |
| stopping ? "stop" : "snap", mLastStartedPlayingTimeNs, where); |
| |
| if (mLastStartedPlayingTimeNs != 0) { |
| sp<NuPlayerDriver> driver = mDriver.promote(); |
| int64_t now = systemTime(); |
| if (driver != NULL) { |
| int64_t played = now - mLastStartedPlayingTimeNs; |
| ALOGV("updatePlaybackTimer() log %20" PRId64 "", played); |
| |
| if (played > 0) { |
| driver->notifyMorePlayingTimeUs((played+500)/1000); |
| } |
| } |
| if (stopping) { |
| mLastStartedPlayingTimeNs = 0; |
| } else { |
| mLastStartedPlayingTimeNs = now; |
| } |
| } |
| } |
| |
| void NuPlayer::startRebufferingTimer() { |
| Mutex::Autolock autoLock(mPlayingTimeLock); |
| if (mLastStartedRebufferingTimeNs == 0) { |
| mLastStartedRebufferingTimeNs = systemTime(); |
| ALOGV("startRebufferingTimer() time %20" PRId64 "", mLastStartedRebufferingTimeNs); |
| } |
| } |
| |
| void NuPlayer::updateRebufferingTimer(bool stopping, bool exitingPlayback) { |
| Mutex::Autolock autoLock(mPlayingTimeLock); |
| |
| ALOGV("updateRebufferingTimer(%s) time %20" PRId64 " (exiting %d)", |
| stopping ? "stop" : "snap", mLastStartedRebufferingTimeNs, exitingPlayback); |
| |
| if (mLastStartedRebufferingTimeNs != 0) { |
| sp<NuPlayerDriver> driver = mDriver.promote(); |
| int64_t now = systemTime(); |
| if (driver != NULL) { |
| int64_t rebuffered = now - mLastStartedRebufferingTimeNs; |
| ALOGV("updateRebufferingTimer() log %20" PRId64 "", rebuffered); |
| |
| if (rebuffered > 0) { |
| driver->notifyMoreRebufferingTimeUs((rebuffered+500)/1000); |
| if (exitingPlayback) { |
| driver->notifyRebufferingWhenExit(true); |
| } |
| } |
| } |
| if (stopping) { |
| mLastStartedRebufferingTimeNs = 0; |
| } else { |
| mLastStartedRebufferingTimeNs = now; |
| } |
| } |
| } |
| |
| void NuPlayer::updateInternalTimers() { |
| // update values, but ticking clocks keep ticking |
| ALOGV("updateInternalTimers()"); |
| updatePlaybackTimer(false /* stopping */, "updateInternalTimers"); |
| updateRebufferingTimer(false /* stopping */, false /* exiting */); |
| } |
| |
| void NuPlayer::setTargetBitrate(int bitrate) { |
| if (mSource != NULL) { |
| mSource->setTargetBitrate(bitrate); |
| } |
| } |
| |
| void NuPlayer::onPause() { |
| |
| updatePlaybackTimer(true /* stopping */, "onPause"); |
| |
| if (mPaused) { |
| return; |
| } |
| mPaused = true; |
| if (mSource != NULL) { |
| mSource->pause(); |
| } else { |
| ALOGW("pause called when source is gone or not set"); |
| } |
| if (mRenderer != NULL) { |
| mRenderer->pause(); |
| } else { |
| ALOGW("pause called when renderer is gone or not set"); |
| } |
| |
| } |
| |
| bool NuPlayer::audioDecoderStillNeeded() { |
| // Audio decoder is no longer needed if it's in shut/shutting down status. |
| return ((mFlushingAudio != SHUT_DOWN) && (mFlushingAudio != SHUTTING_DOWN_DECODER)); |
| } |
| |
| void NuPlayer::handleFlushComplete(bool audio, bool isDecoder) { |
| // We wait for both the decoder flush and the renderer flush to complete |
| // before entering either the FLUSHED or the SHUTTING_DOWN_DECODER state. |
| |
| mFlushComplete[audio][isDecoder] = true; |
| if (!mFlushComplete[audio][!isDecoder]) { |
| return; |
| } |
| |
| FlushStatus *state = audio ? &mFlushingAudio : &mFlushingVideo; |
| switch (*state) { |
| case FLUSHING_DECODER: |
| { |
| *state = FLUSHED; |
| break; |
| } |
| |
| case FLUSHING_DECODER_SHUTDOWN: |
| { |
| *state = SHUTTING_DOWN_DECODER; |
| |
| ALOGV("initiating %s decoder shutdown", audio ? "audio" : "video"); |
| getDecoder(audio)->initiateShutdown(); |
| break; |
| } |
| |
| default: |
| // decoder flush completes only occur in a flushing state. |
| LOG_ALWAYS_FATAL_IF(isDecoder, "decoder flush in invalid state %d", *state); |
| break; |
| } |
| } |
| |
| void NuPlayer::finishFlushIfPossible() { |
| if (mFlushingAudio != NONE && mFlushingAudio != FLUSHED |
| && mFlushingAudio != SHUT_DOWN) { |
| return; |
| } |
| |
| if (mFlushingVideo != NONE && mFlushingVideo != FLUSHED |
| && mFlushingVideo != SHUT_DOWN) { |
| return; |
| } |
| |
| ALOGV("both audio and video are flushed now."); |
| |
| mFlushingAudio = NONE; |
| mFlushingVideo = NONE; |
| |
| clearFlushComplete(); |
| |
| processDeferredActions(); |
| } |
| |
| void NuPlayer::postScanSources() { |
| if (mScanSourcesPending) { |
| return; |
| } |
| |
| sp<AMessage> msg = new AMessage(kWhatScanSources, this); |
| msg->setInt32("generation", mScanSourcesGeneration); |
| msg->post(); |
| |
| mScanSourcesPending = true; |
| } |
| |
| void NuPlayer::tryOpenAudioSinkForOffload( |
| const sp<AMessage> &format, const sp<MetaData> &audioMeta, bool hasVideo) { |
| // Note: This is called early in NuPlayer to determine whether offloading |
| // is possible; otherwise the decoders call the renderer openAudioSink directly. |
| |
| status_t err = mRenderer->openAudioSink( |
| format, true /* offloadOnly */, hasVideo, |
| AUDIO_OUTPUT_FLAG_NONE, &mOffloadAudio, mSource->isStreaming()); |
| if (err != OK) { |
| // Any failure we turn off mOffloadAudio. |
| mOffloadAudio = false; |
| } else if (mOffloadAudio) { |
| sendMetaDataToHal(mAudioSink, audioMeta); |
| } |
| } |
| |
| void NuPlayer::closeAudioSink() { |
| if (mRenderer != NULL) { |
| mRenderer->closeAudioSink(); |
| } |
| } |
| |
| void NuPlayer::restartAudio( |
| int64_t currentPositionUs, bool forceNonOffload, bool needsToCreateAudioDecoder) { |
| ALOGD("restartAudio timeUs(%lld), dontOffload(%d), createDecoder(%d)", |
| (long long)currentPositionUs, forceNonOffload, needsToCreateAudioDecoder); |
| if (mAudioDecoder != NULL) { |
| mAudioDecoder->pause(); |
| Mutex::Autolock autoLock(mDecoderLock); |
| mAudioDecoder.clear(); |
| mAudioDecoderError = false; |
| ++mAudioDecoderGeneration; |
| } |
| if (mFlushingAudio == FLUSHING_DECODER) { |
| mFlushComplete[1 /* audio */][1 /* isDecoder */] = true; |
| mFlushingAudio = FLUSHED; |
| finishFlushIfPossible(); |
| } else if (mFlushingAudio == FLUSHING_DECODER_SHUTDOWN |
| || mFlushingAudio == SHUTTING_DOWN_DECODER) { |
| mFlushComplete[1 /* audio */][1 /* isDecoder */] = true; |
| mFlushingAudio = SHUT_DOWN; |
| finishFlushIfPossible(); |
| needsToCreateAudioDecoder = false; |
| } |
| if (mRenderer == NULL) { |
| return; |
| } |
| closeAudioSink(); |
| mRenderer->flush(true /* audio */, false /* notifyComplete */); |
| if (mVideoDecoder != NULL) { |
| mDeferredActions.push_back( |
| new FlushDecoderAction(FLUSH_CMD_NONE /* audio */, |
| FLUSH_CMD_FLUSH /* video */)); |
| mDeferredActions.push_back( |
| new SeekAction(currentPositionUs, |
| MediaPlayerSeekMode::SEEK_PREVIOUS_SYNC /* mode */)); |
| // After a flush without shutdown, decoder is paused. |
| // Don't resume it until source seek is done, otherwise it could |
| // start pulling stale data too soon. |
| mDeferredActions.push_back(new ResumeDecoderAction(false)); |
| processDeferredActions(); |
| } else { |
| performSeek(currentPositionUs, MediaPlayerSeekMode::SEEK_PREVIOUS_SYNC /* mode */); |
| } |
| |
| if (forceNonOffload) { |
| mRenderer->signalDisableOffloadAudio(); |
| mOffloadAudio = false; |
| } |
| if (needsToCreateAudioDecoder) { |
| instantiateDecoder(true /* audio */, &mAudioDecoder, !forceNonOffload); |
| } |
| } |
| |
| void NuPlayer::determineAudioModeChange(const sp<AMessage> &audioFormat) { |
| if (mSource == NULL || mAudioSink == NULL) { |
| return; |
| } |
| |
| if (mRenderer == NULL) { |
| ALOGW("No renderer can be used to determine audio mode. Use non-offload for safety."); |
| mOffloadAudio = false; |
| return; |
| } |
| |
| sp<MetaData> audioMeta = mSource->getFormatMeta(true /* audio */); |
| sp<AMessage> videoFormat = mSource->getFormat(false /* audio */); |
| audio_stream_type_t streamType = mAudioSink->getAudioStreamType(); |
| const bool hasVideo = (videoFormat != NULL); |
| bool canOffload = canOffloadStream( |
| audioMeta, hasVideo, mSource->isStreaming(), streamType) |
| && (mPlaybackSettings.mSpeed == 1.f && mPlaybackSettings.mPitch == 1.f); |
| |
| // Modular DRM: Disabling audio offload if the source is protected |
| if (canOffload && mIsDrmProtected) { |
| canOffload = false; |
| ALOGV("determineAudioModeChange: Disabling mOffloadAudio b/c the source is protected."); |
| } |
| |
| if (canOffload) { |
| if (!mOffloadAudio) { |
| mRenderer->signalEnableOffloadAudio(); |
| } |
| // open audio sink early under offload mode. |
| tryOpenAudioSinkForOffload(audioFormat, audioMeta, hasVideo); |
| } else { |
| if (mOffloadAudio) { |
| mRenderer->signalDisableOffloadAudio(); |
| mOffloadAudio = false; |
| } |
| } |
| } |
| |
| status_t NuPlayer::instantiateDecoder( |
| bool audio, sp<DecoderBase> *decoder, bool checkAudioModeChange) { |
| // The audio decoder could be cleared by tear down. If still in shut down |
| // process, no need to create a new audio decoder. |
| if (*decoder != NULL || (audio && mFlushingAudio == SHUT_DOWN)) { |
| return OK; |
| } |
| |
| sp<AMessage> format = mSource->getFormat(audio); |
| |
| if (format == NULL) { |
| return UNKNOWN_ERROR; |
| } else { |
| status_t err; |
| if (format->findInt32("err", &err) && err) { |
| return err; |
| } |
| } |
| |
| format->setInt32("priority", 0 /* realtime */); |
| |
| if (mDataSourceType == DATA_SOURCE_TYPE_RTP) { |
| ALOGV("instantiateDecoder: set decoder error free on stream corrupt."); |
| format->setInt32("corrupt-free", true); |
| } |
| |
| if (!audio) { |
| AString mime; |
| CHECK(format->findString("mime", &mime)); |
| |
| sp<AMessage> ccNotify = new AMessage(kWhatClosedCaptionNotify, this); |
| if (mCCDecoder == NULL) { |
| mCCDecoder = new CCDecoder(ccNotify); |
| } |
| |
| if (mSourceFlags & Source::FLAG_SECURE) { |
| format->setInt32("secure", true); |
| } |
| |
| if (mSourceFlags & Source::FLAG_PROTECTED) { |
| format->setInt32("protected", true); |
| } |
| |
| float rate = getFrameRate(); |
| if (rate > 0) { |
| format->setFloat("operating-rate", rate * mPlaybackSettings.mSpeed); |
| } |
| |
| format->setInt32("android._video-scaling", mVideoScalingMode); |
| } |
| |
| Mutex::Autolock autoLock(mDecoderLock); |
| |
| if (audio) { |
| sp<AMessage> notify = new AMessage(kWhatAudioNotify, this); |
| ++mAudioDecoderGeneration; |
| notify->setInt32("generation", mAudioDecoderGeneration); |
| |
| if (checkAudioModeChange) { |
| determineAudioModeChange(format); |
| } |
| if (mOffloadAudio) { |
| mSource->setOffloadAudio(true /* offload */); |
| |
| const bool hasVideo = (mSource->getFormat(false /*audio */) != NULL); |
| format->setInt32("has-video", hasVideo); |
| *decoder = new DecoderPassThrough(notify, mSource, mRenderer); |
| ALOGV("instantiateDecoder audio DecoderPassThrough hasVideo: %d", hasVideo); |
| } else { |
| mSource->setOffloadAudio(false /* offload */); |
| |
| *decoder = new Decoder(notify, mSource, mPID, mUID, mRenderer); |
| ALOGV("instantiateDecoder audio Decoder"); |
| } |
| mAudioDecoderError = false; |
| } else { |
| sp<AMessage> notify = new AMessage(kWhatVideoNotify, this); |
| ++mVideoDecoderGeneration; |
| notify->setInt32("generation", mVideoDecoderGeneration); |
| |
| *decoder = new Decoder( |
| notify, mSource, mPID, mUID, mRenderer, mSurface, mCCDecoder); |
| mVideoDecoderError = false; |
| |
| // enable FRC if high-quality AV sync is requested, even if not |
| // directly queuing to display, as this will even improve textureview |
| // playback. |
| { |
| if (property_get_bool("persist.sys.media.avsync", false)) { |
| format->setInt32("auto-frc", 1); |
| } |
| } |
| } |
| (*decoder)->init(); |
| |
| // Modular DRM |
| if (mIsDrmProtected) { |
| format->setPointer("crypto", mCrypto.get()); |
| ALOGV("instantiateDecoder: mCrypto: %p (%d) isSecure: %d", mCrypto.get(), |
| (mCrypto != NULL ? mCrypto->getStrongCount() : 0), |
| (mSourceFlags & Source::FLAG_SECURE) != 0); |
| } |
| |
| (*decoder)->configure(format); |
| |
| if (!audio) { |
| sp<AMessage> params = new AMessage(); |
| float rate = getFrameRate(); |
| if (rate > 0) { |
| params->setFloat("frame-rate-total", rate); |
| } |
| |
| sp<MetaData> fileMeta = getFileMeta(); |
| if (fileMeta != NULL) { |
| int32_t videoTemporalLayerCount; |
| if (fileMeta->findInt32(kKeyTemporalLayerCount, &videoTemporalLayerCount) |
| && videoTemporalLayerCount > 0) { |
| params->setInt32("temporal-layer-count", videoTemporalLayerCount); |
| } |
| } |
| |
| if (params->countEntries() > 0) { |
| (*decoder)->setParameters(params); |
| } |
| } |
| return OK; |
| } |
| |
| void NuPlayer::updateVideoSize( |
| const sp<AMessage> &inputFormat, |
| const sp<AMessage> &outputFormat) { |
| if (inputFormat == NULL) { |
| ALOGW("Unknown video size, reporting 0x0!"); |
| notifyListener(MEDIA_SET_VIDEO_SIZE, 0, 0); |
| return; |
| } |
| int32_t err = OK; |
| inputFormat->findInt32("err", &err); |
| if (err == -EWOULDBLOCK) { |
| ALOGW("Video meta is not available yet!"); |
| return; |
| } |
| if (err != OK) { |
| ALOGW("Something is wrong with video meta!"); |
| return; |
| } |
| |
| int32_t displayWidth, displayHeight; |
| if (outputFormat != NULL) { |
| int32_t width, height; |
| CHECK(outputFormat->findInt32("width", &width)); |
| CHECK(outputFormat->findInt32("height", &height)); |
| |
| int32_t cropLeft, cropTop, cropRight, cropBottom; |
| CHECK(outputFormat->findRect( |
| "crop", |
| &cropLeft, &cropTop, &cropRight, &cropBottom)); |
| |
| displayWidth = cropRight - cropLeft + 1; |
| displayHeight = cropBottom - cropTop + 1; |
| |
| ALOGV("Video output format changed to %d x %d " |
| "(crop: %d x %d @ (%d, %d))", |
| width, height, |
| displayWidth, |
| displayHeight, |
| cropLeft, cropTop); |
| } else { |
| CHECK(inputFormat->findInt32("width", &displayWidth)); |
| CHECK(inputFormat->findInt32("height", &displayHeight)); |
| |
| ALOGV("Video input format %d x %d", displayWidth, displayHeight); |
| } |
| |
| // Take into account sample aspect ratio if necessary: |
| int32_t sarWidth, sarHeight; |
| if (inputFormat->findInt32("sar-width", &sarWidth) |
| && inputFormat->findInt32("sar-height", &sarHeight) |
| && sarWidth > 0 && sarHeight > 0) { |
| ALOGV("Sample aspect ratio %d : %d", sarWidth, sarHeight); |
| |
| displayWidth = (displayWidth * sarWidth) / sarHeight; |
| |
| ALOGV("display dimensions %d x %d", displayWidth, displayHeight); |
| } else { |
| int32_t width, height; |
| if (inputFormat->findInt32("display-width", &width) |
| && inputFormat->findInt32("display-height", &height) |
| && width > 0 && height > 0 |
| && displayWidth > 0 && displayHeight > 0) { |
| if (displayHeight * (int64_t)width / height > (int64_t)displayWidth) { |
| displayHeight = (int32_t)(displayWidth * (int64_t)height / width); |
| } else { |
| displayWidth = (int32_t)(displayHeight * (int64_t)width / height); |
| } |
| ALOGV("Video display width and height are overridden to %d x %d", |
| displayWidth, displayHeight); |
| } |
| } |
| |
| int32_t rotationDegrees; |
| if (!inputFormat->findInt32("rotation-degrees", &rotationDegrees)) { |
| rotationDegrees = 0; |
| } |
| |
| if (rotationDegrees == 90 || rotationDegrees == 270) { |
| int32_t tmp = displayWidth; |
| displayWidth = displayHeight; |
| displayHeight = tmp; |
| } |
| |
| notifyListener( |
| MEDIA_SET_VIDEO_SIZE, |
| displayWidth, |
| displayHeight); |
| } |
| |
| void NuPlayer::notifyListener(int msg, int ext1, int ext2, const Parcel *in) { |
| if (mDriver == NULL) { |
| return; |
| } |
| |
| sp<NuPlayerDriver> driver = mDriver.promote(); |
| |
| if (driver == NULL) { |
| return; |
| } |
| |
| driver->notifyListener(msg, ext1, ext2, in); |
| } |
| |
| void NuPlayer::flushDecoder(bool audio, bool needShutdown) { |
| ALOGV("[%s] flushDecoder needShutdown=%d", |
| audio ? "audio" : "video", needShutdown); |
| |
| const sp<DecoderBase> &decoder = getDecoder(audio); |
| if (decoder == NULL) { |
| ALOGI("flushDecoder %s without decoder present", |
| audio ? "audio" : "video"); |
| return; |
| } |
| |
| // Make sure we don't continue to scan sources until we finish flushing. |
| ++mScanSourcesGeneration; |
| if (mScanSourcesPending) { |
| if (!needShutdown) { |
| mDeferredActions.push_back( |
| new SimpleAction(&NuPlayer::performScanSources)); |
| } |
| mScanSourcesPending = false; |
| } |
| |
| decoder->signalFlush(); |
| |
| FlushStatus newStatus = |
| needShutdown ? FLUSHING_DECODER_SHUTDOWN : FLUSHING_DECODER; |
| |
| mFlushComplete[audio][false /* isDecoder */] = (mRenderer == NULL); |
| mFlushComplete[audio][true /* isDecoder */] = false; |
| if (audio) { |
| ALOGE_IF(mFlushingAudio != NONE, |
| "audio flushDecoder() is called in state %d", mFlushingAudio); |
| mFlushingAudio = newStatus; |
| } else { |
| ALOGE_IF(mFlushingVideo != NONE, |
| "video flushDecoder() is called in state %d", mFlushingVideo); |
| mFlushingVideo = newStatus; |
| } |
| } |
| |
| void NuPlayer::queueDecoderShutdown( |
| bool audio, bool video, const sp<AMessage> &reply) { |
| ALOGI("queueDecoderShutdown audio=%d, video=%d", audio, video); |
| |
| mDeferredActions.push_back( |
| new FlushDecoderAction( |
| audio ? FLUSH_CMD_SHUTDOWN : FLUSH_CMD_NONE, |
| video ? FLUSH_CMD_SHUTDOWN : FLUSH_CMD_NONE)); |
| |
| mDeferredActions.push_back( |
| new SimpleAction(&NuPlayer::performScanSources)); |
| |
| mDeferredActions.push_back(new PostMessageAction(reply)); |
| |
| processDeferredActions(); |
| } |
| |
| status_t NuPlayer::setVideoScalingMode(int32_t mode) { |
| mVideoScalingMode = mode; |
| if (mSurface != NULL) { |
| status_t ret = native_window_set_scaling_mode(mSurface.get(), mVideoScalingMode); |
| if (ret != OK) { |
| ALOGE("Failed to set scaling mode (%d): %s", |
| -ret, strerror(-ret)); |
| return ret; |
| } |
| if (mVideoDecoder != NULL) { |
| sp<AMessage> params = new AMessage(); |
| params->setInt32("android._video-scaling", mode); |
| mVideoDecoder->setParameters(params); |
| } |
| } |
| return OK; |
| } |
| |
| status_t NuPlayer::getTrackInfo(Parcel* reply) const { |
| sp<AMessage> msg = new AMessage(kWhatGetTrackInfo, this); |
| msg->setPointer("reply", reply); |
| |
| sp<AMessage> response; |
| status_t err = msg->postAndAwaitResponse(&response); |
| return err; |
| } |
| |
| status_t NuPlayer::getSelectedTrack(int32_t type, Parcel* reply) const { |
| sp<AMessage> msg = new AMessage(kWhatGetSelectedTrack, this); |
| msg->setPointer("reply", reply); |
| msg->setInt32("type", type); |
| |
| sp<AMessage> response; |
| status_t err = msg->postAndAwaitResponse(&response); |
| if (err == OK && response != NULL) { |
| CHECK(response->findInt32("err", &err)); |
| } |
| return err; |
| } |
| |
| status_t NuPlayer::selectTrack(size_t trackIndex, bool select, int64_t timeUs) { |
| sp<AMessage> msg = new AMessage(kWhatSelectTrack, this); |
| msg->setSize("trackIndex", trackIndex); |
| msg->setInt32("select", select); |
| msg->setInt64("timeUs", timeUs); |
| |
| sp<AMessage> response; |
| status_t err = msg->postAndAwaitResponse(&response); |
| |
| if (err != OK) { |
| return err; |
| } |
| |
| if (!response->findInt32("err", &err)) { |
| err = OK; |
| } |
| |
| return err; |
| } |
| |
| status_t NuPlayer::getCurrentPosition(int64_t *mediaUs) { |
| sp<Renderer> renderer = mRenderer; |
| if (renderer == NULL) { |
| return NO_INIT; |
| } |
| |
| return renderer->getCurrentPosition(mediaUs); |
| } |
| |
| void NuPlayer::getStats(Vector<sp<AMessage> > *trackStats) { |
| CHECK(trackStats != NULL); |
| |
| trackStats->clear(); |
| |
| Mutex::Autolock autoLock(mDecoderLock); |
| if (mVideoDecoder != NULL) { |
| trackStats->push_back(mVideoDecoder->getStats()); |
| } |
| if (mAudioDecoder != NULL) { |
| trackStats->push_back(mAudioDecoder->getStats()); |
| } |
| } |
| |
| sp<MetaData> NuPlayer::getFileMeta() { |
| return mSource->getFileFormatMeta(); |
| } |
| |
| float NuPlayer::getFrameRate() { |
| sp<MetaData> meta = mSource->getFormatMeta(false /* audio */); |
| if (meta == NULL) { |
| return 0; |
| } |
| int32_t rate; |
| if (!meta->findInt32(kKeyFrameRate, &rate)) { |
| // fall back to try file meta |
| sp<MetaData> fileMeta = getFileMeta(); |
| if (fileMeta == NULL) { |
| ALOGW("source has video meta but not file meta"); |
| return -1; |
| } |
| int32_t fileMetaRate; |
| if (!fileMeta->findInt32(kKeyFrameRate, &fileMetaRate)) { |
| return -1; |
| } |
| return fileMetaRate; |
| } |
| return rate; |
| } |
| |
| void NuPlayer::schedulePollDuration() { |
| sp<AMessage> msg = new AMessage(kWhatPollDuration, this); |
| msg->setInt32("generation", mPollDurationGeneration); |
| msg->post(); |
| } |
| |
| void NuPlayer::cancelPollDuration() { |
| ++mPollDurationGeneration; |
| } |
| |
| void NuPlayer::processDeferredActions() { |
| while (!mDeferredActions.empty()) { |
| // We won't execute any deferred actions until we're no longer in |
| // an intermediate state, i.e. one more more decoders are currently |
| // flushing or shutting down. |
| |
| if (mFlushingAudio != NONE || mFlushingVideo != NONE) { |
| // We're currently flushing, postpone the reset until that's |
| // completed. |
| |
| ALOGV("postponing action mFlushingAudio=%d, mFlushingVideo=%d", |
| mFlushingAudio, mFlushingVideo); |
| |
| break; |
| } |
| |
| sp<Action> action = *mDeferredActions.begin(); |
| mDeferredActions.erase(mDeferredActions.begin()); |
| |
| action->execute(this); |
| } |
| } |
| |
| void NuPlayer::performSeek(int64_t seekTimeUs, MediaPlayerSeekMode mode) { |
| ALOGV("performSeek seekTimeUs=%lld us (%.2f secs), mode=%d", |
| (long long)seekTimeUs, seekTimeUs / 1E6, mode); |
| |
| if (mSource == NULL) { |
| // This happens when reset occurs right before the loop mode |
| // asynchronously seeks to the start of the stream. |
| LOG_ALWAYS_FATAL_IF(mAudioDecoder != NULL || mVideoDecoder != NULL, |
| "mSource is NULL and decoders not NULL audio(%p) video(%p)", |
| mAudioDecoder.get(), mVideoDecoder.get()); |
| return; |
| } |
| mPreviousSeekTimeUs = seekTimeUs; |
| mSource->seekTo(seekTimeUs, mode); |
| ++mTimedTextGeneration; |
| |
| // everything's flushed, continue playback. |
| } |
| |
| void NuPlayer::performDecoderFlush(FlushCommand audio, FlushCommand video) { |
| ALOGV("performDecoderFlush audio=%d, video=%d", audio, video); |
| |
| if ((audio == FLUSH_CMD_NONE || mAudioDecoder == NULL) |
| && (video == FLUSH_CMD_NONE || mVideoDecoder == NULL)) { |
| return; |
| } |
| |
| if (audio != FLUSH_CMD_NONE && mAudioDecoder != NULL) { |
| flushDecoder(true /* audio */, (audio == FLUSH_CMD_SHUTDOWN)); |
| } |
| |
| if (video != FLUSH_CMD_NONE && mVideoDecoder != NULL) { |
| flushDecoder(false /* audio */, (video == FLUSH_CMD_SHUTDOWN)); |
| } |
| } |
| |
| void NuPlayer::performReset() { |
| ALOGV("performReset"); |
| |
| CHECK(mAudioDecoder == NULL); |
| CHECK(mVideoDecoder == NULL); |
| |
| updatePlaybackTimer(true /* stopping */, "performReset"); |
| updateRebufferingTimer(true /* stopping */, true /* exiting */); |
| |
| cancelPollDuration(); |
| |
| ++mScanSourcesGeneration; |
| mScanSourcesPending = false; |
| |
| if (mRendererLooper != NULL) { |
| if (mRenderer != NULL) { |
| mRendererLooper->unregisterHandler(mRenderer->id()); |
| } |
| mRendererLooper->stop(); |
| mRendererLooper.clear(); |
| } |
| mRenderer.clear(); |
| ++mRendererGeneration; |
| |
| if (mSource != NULL) { |
| mSource->stop(); |
| |
| Mutex::Autolock autoLock(mSourceLock); |
| mSource.clear(); |
| } |
| |
| if (mDriver != NULL) { |
| sp<NuPlayerDriver> driver = mDriver.promote(); |
| if (driver != NULL) { |
| driver->notifyResetComplete(); |
| } |
| } |
| |
| mStarted = false; |
| mPrepared = false; |
| mResetting = false; |
| mSourceStarted = false; |
| |
| // Modular DRM |
| if (mCrypto != NULL) { |
| // decoders will be flushed before this so their mCrypto would go away on their own |
| // TODO change to ALOGV |
| ALOGD("performReset mCrypto: %p (%d)", mCrypto.get(), |
| (mCrypto != NULL ? mCrypto->getStrongCount() : 0)); |
| mCrypto.clear(); |
| } |
| mIsDrmProtected = false; |
| } |
| |
| void NuPlayer::performScanSources() { |
| ALOGV("performScanSources"); |
| |
| if (!mStarted) { |
| return; |
| } |
| |
| if (mAudioDecoder == NULL || mVideoDecoder == NULL) { |
| postScanSources(); |
| } |
| } |
| |
| void NuPlayer::performSetSurface(const sp<Surface> &surface) { |
| ALOGV("performSetSurface"); |
| |
| mSurface = surface; |
| |
| // XXX - ignore error from setVideoScalingMode for now |
| setVideoScalingMode(mVideoScalingMode); |
| |
| if (mDriver != NULL) { |
| sp<NuPlayerDriver> driver = mDriver.promote(); |
| if (driver != NULL) { |
| driver->notifySetSurfaceComplete(); |
| } |
| } |
| } |
| |
| void NuPlayer::performResumeDecoders(bool needNotify) { |
| if (needNotify) { |
| mResumePending = true; |
| if (mVideoDecoder == NULL) { |
| // if audio-only, we can notify seek complete now, |
| // as the resume operation will be relatively fast. |
| finishResume(); |
| } |
| } |
| |
| if (mVideoDecoder != NULL) { |
| // When there is continuous seek, MediaPlayer will cache the seek |
| // position, and send down new seek request when previous seek is |
| // complete. Let's wait for at least one video output frame before |
| // notifying seek complete, so that the video thumbnail gets updated |
| // when seekbar is dragged. |
| mVideoDecoder->signalResume(needNotify); |
| } |
| |
| if (mAudioDecoder != NULL) { |
| mAudioDecoder->signalResume(false /* needNotify */); |
| } |
| } |
| |
| void NuPlayer::finishResume() { |
| if (mResumePending) { |
| mResumePending = false; |
| notifyDriverSeekComplete(); |
| } |
| } |
| |
| void NuPlayer::notifyDriverSeekComplete() { |
| if (mDriver != NULL) { |
| sp<NuPlayerDriver> driver = mDriver.promote(); |
| if (driver != NULL) { |
| driver->notifySeekComplete(); |
| } |
| } |
| } |
| |
| void NuPlayer::onSourceNotify(const sp<AMessage> &msg) { |
| int32_t what; |
| CHECK(msg->findInt32("what", &what)); |
| |
| switch (what) { |
| case Source::kWhatInstantiateSecureDecoders: |
| { |
| if (mSource == NULL) { |
| // This is a stale notification from a source that was |
| // asynchronously preparing when the client called reset(). |
| // We handled the reset, the source is gone. |
| break; |
| } |
| |
| sp<AMessage> reply; |
| CHECK(msg->findMessage("reply", &reply)); |
| status_t err = onInstantiateSecureDecoders(); |
| reply->setInt32("err", err); |
| reply->post(); |
| break; |
| } |
| |
| case Source::kWhatPrepared: |
| { |
| ALOGV("NuPlayer::onSourceNotify Source::kWhatPrepared source: %p", mSource.get()); |
| if (mSource == NULL) { |
| // This is a stale notification from a source that was |
| // asynchronously preparing when the client called reset(). |
| // We handled the reset, the source is gone. |
| break; |
| } |
| |
| int32_t err; |
| CHECK(msg->findInt32("err", &err)); |
| |
| if (err != OK) { |
| // shut down potential secure codecs in case client never calls reset |
| mDeferredActions.push_back( |
| new FlushDecoderAction(FLUSH_CMD_SHUTDOWN /* audio */, |
| FLUSH_CMD_SHUTDOWN /* video */)); |
| processDeferredActions(); |
| } else { |
| mPrepared = true; |
| } |
| |
| sp<NuPlayerDriver> driver = mDriver.promote(); |
| if (driver != NULL) { |
| // notify duration first, so that it's definitely set when |
| // the app received the "prepare complete" callback. |
| int64_t durationUs; |
| if (mSource->getDuration(&durationUs) == OK) { |
| driver->notifyDuration(durationUs); |
| } |
| driver->notifyPrepareCompleted(err); |
| } |
| |
| break; |
| } |
| |
| // Modular DRM |
| case Source::kWhatDrmInfo: |
| { |
| Parcel parcel; |
| sp<ABuffer> drmInfo; |
| CHECK(msg->findBuffer("drmInfo", &drmInfo)); |
| parcel.setData(drmInfo->data(), drmInfo->size()); |
| |
| ALOGV("onSourceNotify() kWhatDrmInfo MEDIA_DRM_INFO drmInfo: %p parcel size: %zu", |
| drmInfo.get(), parcel.dataSize()); |
| |
| notifyListener(MEDIA_DRM_INFO, 0 /* ext1 */, 0 /* ext2 */, &parcel); |
| |
| break; |
| } |
| |
| case Source::kWhatFlagsChanged: |
| { |
| uint32_t flags; |
| CHECK(msg->findInt32("flags", (int32_t *)&flags)); |
| |
| sp<NuPlayerDriver> driver = mDriver.promote(); |
| if (driver != NULL) { |
| |
| ALOGV("onSourceNotify() kWhatFlagsChanged FLAG_CAN_PAUSE: %d " |
| "FLAG_CAN_SEEK_BACKWARD: %d \n\t\t\t\t FLAG_CAN_SEEK_FORWARD: %d " |
| "FLAG_CAN_SEEK: %d FLAG_DYNAMIC_DURATION: %d \n" |
| "\t\t\t\t FLAG_SECURE: %d FLAG_PROTECTED: %d", |
| (flags & Source::FLAG_CAN_PAUSE) != 0, |
| (flags & Source::FLAG_CAN_SEEK_BACKWARD) != 0, |
| (flags & Source::FLAG_CAN_SEEK_FORWARD) != 0, |
| (flags & Source::FLAG_CAN_SEEK) != 0, |
| (flags & Source::FLAG_DYNAMIC_DURATION) != 0, |
| (flags & Source::FLAG_SECURE) != 0, |
| (flags & Source::FLAG_PROTECTED) != 0); |
| |
| if ((flags & NuPlayer::Source::FLAG_CAN_SEEK) == 0) { |
| driver->notifyListener( |
| MEDIA_INFO, MEDIA_INFO_NOT_SEEKABLE, 0); |
| } |
| driver->notifyFlagsChanged(flags); |
| } |
| |
| if ((mSourceFlags & Source::FLAG_DYNAMIC_DURATION) |
| && (!(flags & Source::FLAG_DYNAMIC_DURATION))) { |
| cancelPollDuration(); |
| } else if (!(mSourceFlags & Source::FLAG_DYNAMIC_DURATION) |
| && (flags & Source::FLAG_DYNAMIC_DURATION) |
| && (mAudioDecoder != NULL || mVideoDecoder != NULL)) { |
| schedulePollDuration(); |
| } |
| |
| mSourceFlags = flags; |
| break; |
| } |
| |
| case Source::kWhatVideoSizeChanged: |
| { |
| sp<AMessage> format; |
| CHECK(msg->findMessage("format", &format)); |
| |
| updateVideoSize(format); |
| break; |
| } |
| |
| case Source::kWhatBufferingUpdate: |
| { |
| int32_t percentage; |
| CHECK(msg->findInt32("percentage", &percentage)); |
| |
| notifyListener(MEDIA_BUFFERING_UPDATE, percentage, 0); |
| break; |
| } |
| |
| case Source::kWhatPauseOnBufferingStart: |
| { |
| // ignore if not playing |
| if (mStarted) { |
| ALOGI("buffer low, pausing..."); |
| |
| startRebufferingTimer(); |
| mPausedForBuffering = true; |
| onPause(); |
| } |
| notifyListener(MEDIA_INFO, MEDIA_INFO_BUFFERING_START, 0); |
| break; |
| } |
| |
| case Source::kWhatResumeOnBufferingEnd: |
| { |
| // ignore if not playing |
| if (mStarted) { |
| ALOGI("buffer ready, resuming..."); |
| |
| updateRebufferingTimer(true /* stopping */, false /* exiting */); |
| mPausedForBuffering = false; |
| |
| // do not resume yet if client didn't unpause |
| if (!mPausedByClient) { |
| onResume(); |
| } |
| } |
| notifyListener(MEDIA_INFO, MEDIA_INFO_BUFFERING_END, 0); |
| break; |
| } |
| |
| case Source::kWhatCacheStats: |
| { |
| int32_t kbps; |
| CHECK(msg->findInt32("bandwidth", &kbps)); |
| |
| notifyListener(MEDIA_INFO, MEDIA_INFO_NETWORK_BANDWIDTH, kbps); |
| break; |
| } |
| |
| case Source::kWhatSubtitleData: |
| { |
| sp<ABuffer> buffer; |
| CHECK(msg->findBuffer("buffer", &buffer)); |
| |
| sendSubtitleData(buffer, 0 /* baseIndex */); |
| break; |
| } |
| |
| case Source::kWhatTimedMetaData: |
| { |
| sp<ABuffer> buffer; |
| if (!msg->findBuffer("buffer", &buffer)) { |
| notifyListener(MEDIA_INFO, MEDIA_INFO_METADATA_UPDATE, 0); |
| } else { |
| sendTimedMetaData(buffer); |
| } |
| break; |
| } |
| |
| case Source::kWhatTimedTextData: |
| { |
| int32_t generation; |
| if (msg->findInt32("generation", &generation) |
| && generation != mTimedTextGeneration) { |
| break; |
| } |
| |
| sp<ABuffer> buffer; |
| CHECK(msg->findBuffer("buffer", &buffer)); |
| |
| sp<NuPlayerDriver> driver = mDriver.promote(); |
| if (driver == NULL) { |
| break; |
| } |
| |
| int posMs; |
| int64_t timeUs, posUs; |
| driver->getCurrentPosition(&posMs); |
| posUs = (int64_t) posMs * 1000LL; |
| CHECK(buffer->meta()->findInt64("timeUs", &timeUs)); |
| |
| if (posUs < timeUs) { |
| if (!msg->findInt32("generation", &generation)) { |
| msg->setInt32("generation", mTimedTextGeneration); |
| } |
| msg->post(timeUs - posUs); |
| } else { |
| sendTimedTextData(buffer); |
| } |
| break; |
| } |
| |
| case Source::kWhatQueueDecoderShutdown: |
| { |
| int32_t audio, video; |
| CHECK(msg->findInt32("audio", &audio)); |
| CHECK(msg->findInt32("video", &video)); |
| |
| sp<AMessage> reply; |
| CHECK(msg->findMessage("reply", &reply)); |
| |
| queueDecoderShutdown(audio, video, reply); |
| break; |
| } |
| |
| case Source::kWhatDrmNoLicense: |
| { |
| notifyListener(MEDIA_ERROR, MEDIA_ERROR_UNKNOWN, ERROR_DRM_NO_LICENSE); |
| break; |
| } |
| |
| case Source::kWhatIMSRxNotice: |
| { |
| sp<AMessage> IMSRxNotice; |
| CHECK(msg->findMessage("message", &IMSRxNotice)); |
| sendIMSRxNotice(IMSRxNotice); |
| break; |
| } |
| |
| default: |
| TRESPASS(); |
| } |
| } |
| |
| void NuPlayer::onClosedCaptionNotify(const sp<AMessage> &msg) { |
| int32_t what; |
| CHECK(msg->findInt32("what", &what)); |
| |
| switch (what) { |
| case NuPlayer::CCDecoder::kWhatClosedCaptionData: |
| { |
| sp<ABuffer> buffer; |
| CHECK(msg->findBuffer("buffer", &buffer)); |
| |
| size_t inbandTracks = 0; |
| if (mSource != NULL) { |
| inbandTracks = mSource->getTrackCount(); |
| } |
| |
| sendSubtitleData(buffer, inbandTracks); |
| break; |
| } |
| |
| case NuPlayer::CCDecoder::kWhatTrackAdded: |
| { |
| notifyListener(MEDIA_INFO, MEDIA_INFO_METADATA_UPDATE, 0); |
| |
| break; |
| } |
| |
| default: |
| TRESPASS(); |
| } |
| |
| |
| } |
| |
| void NuPlayer::sendSubtitleData(const sp<ABuffer> &buffer, int32_t baseIndex) { |
| int32_t trackIndex; |
| int64_t timeUs, durationUs; |
| CHECK(buffer->meta()->findInt32("track-index", &trackIndex)); |
| CHECK(buffer->meta()->findInt64("timeUs", &timeUs)); |
| CHECK(buffer->meta()->findInt64("durationUs", &durationUs)); |
| |
| Parcel in; |
| in.writeInt32(trackIndex + baseIndex); |
| in.writeInt64(timeUs); |
| in.writeInt64(durationUs); |
| in.writeInt32(buffer->size()); |
| in.writeInt32(buffer->size()); |
| in.write(buffer->data(), buffer->size()); |
| |
| notifyListener(MEDIA_SUBTITLE_DATA, 0, 0, &in); |
| } |
| |
| void NuPlayer::sendTimedMetaData(const sp<ABuffer> &buffer) { |
| int64_t timeUs; |
| CHECK(buffer->meta()->findInt64("timeUs", &timeUs)); |
| |
| Parcel in; |
| in.writeInt64(timeUs); |
| in.writeInt32(buffer->size()); |
| in.writeInt32(buffer->size()); |
| in.write(buffer->data(), buffer->size()); |
| |
| notifyListener(MEDIA_META_DATA, 0, 0, &in); |
| } |
| |
| void NuPlayer::sendTimedTextData(const sp<ABuffer> &buffer) { |
| const void *data; |
| size_t size = 0; |
| int64_t timeUs; |
| int32_t flag = TextDescriptions::IN_BAND_TEXT_3GPP; |
| |
| AString mime; |
| CHECK(buffer->meta()->findString("mime", &mime)); |
| CHECK(strcasecmp(mime.c_str(), MEDIA_MIMETYPE_TEXT_3GPP) == 0); |
| |
| data = buffer->data(); |
| size = buffer->size(); |
| |
| Parcel parcel; |
| if (size > 0) { |
| CHECK(buffer->meta()->findInt64("timeUs", &timeUs)); |
| int32_t global = 0; |
| if (buffer->meta()->findInt32("global", &global) && global) { |
| flag |= TextDescriptions::GLOBAL_DESCRIPTIONS; |
| } else { |
| flag |= TextDescriptions::LOCAL_DESCRIPTIONS; |
| } |
| TextDescriptions::getParcelOfDescriptions( |
| (const uint8_t *)data, size, flag, timeUs / 1000, &parcel); |
| } |
| |
| if ((parcel.dataSize() > 0)) { |
| notifyListener(MEDIA_TIMED_TEXT, 0, 0, &parcel); |
| } else { // send an empty timed text |
| notifyListener(MEDIA_TIMED_TEXT, 0, 0); |
| } |
| } |
| |
| void NuPlayer::sendIMSRxNotice(const sp<AMessage> &msg) { |
| int32_t payloadType; |
| |
| CHECK(msg->findInt32("payload-type", &payloadType)); |
| |
| int32_t rtpSeq = 0, rtpTime = 0; |
| int64_t ntpTime = 0, recvTimeUs = 0; |
| |
| Parcel in; |
| in.writeInt32(payloadType); |
| |
| switch (payloadType) { |
| case ARTPSource::RTP_FIRST_PACKET: |
| { |
| CHECK(msg->findInt32("rtp-time", &rtpTime)); |
| CHECK(msg->findInt32("rtp-seq-num", &rtpSeq)); |
| CHECK(msg->findInt64("recv-time-us", &recvTimeUs)); |
| in.writeInt32(rtpTime); |
| in.writeInt32(rtpSeq); |
| in.writeInt32(recvTimeUs >> 32); |
| in.writeInt32(recvTimeUs & 0xFFFFFFFF); |
| break; |
| } |
| case ARTPSource::RTCP_FIRST_PACKET: |
| { |
| CHECK(msg->findInt64("recv-time-us", &recvTimeUs)); |
| in.writeInt32(recvTimeUs >> 32); |
| in.writeInt32(recvTimeUs & 0xFFFFFFFF); |
| break; |
| } |
| case ARTPSource::RTCP_SR: |
| { |
| CHECK(msg->findInt32("rtp-time", &rtpTime)); |
| CHECK(msg->findInt64("ntp-time", &ntpTime)); |
| CHECK(msg->findInt64("recv-time-us", &recvTimeUs)); |
| in.writeInt32(rtpTime); |
| in.writeInt32(ntpTime >> 32); |
| in.writeInt32(ntpTime & 0xFFFFFFFF); |
| in.writeInt32(recvTimeUs >> 32); |
| in.writeInt32(recvTimeUs & 0xFFFFFFFF); |
| break; |
| } |
| case ARTPSource::RTCP_RR: |
| { |
| int64_t recvTimeUs; |
| int32_t senderId; |
| int32_t ssrc; |
| int32_t fraction; |
| int32_t lost; |
| int32_t lastSeq; |
| int32_t jitter; |
| int32_t lsr; |
| int32_t dlsr; |
| CHECK(msg->findInt64("recv-time-us", &recvTimeUs)); |
| CHECK(msg->findInt32("rtcp-rr-ssrc", &senderId)); |
| CHECK(msg->findInt32("rtcp-rrb-ssrc", &ssrc)); |
| CHECK(msg->findInt32("rtcp-rrb-fraction", &fraction)); |
| CHECK(msg->findInt32("rtcp-rrb-lost", &lost)); |
| CHECK(msg->findInt32("rtcp-rrb-lastSeq", &lastSeq)); |
| CHECK(msg->findInt32("rtcp-rrb-jitter", &jitter)); |
| CHECK(msg->findInt32("rtcp-rrb-lsr", &lsr)); |
| CHECK(msg->findInt32("rtcp-rrb-dlsr", &dlsr)); |
| in.writeInt32(recvTimeUs >> 32); |
| in.writeInt32(recvTimeUs & 0xFFFFFFFF); |
| in.writeInt32(senderId); |
| in.writeInt32(ssrc); |
| in.writeInt32(fraction); |
| in.writeInt32(lost); |
| in.writeInt32(lastSeq); |
| in.writeInt32(jitter); |
| in.writeInt32(lsr); |
| in.writeInt32(dlsr); |
| break; |
| } |
| case ARTPSource::RTCP_TSFB: // RTCP TSFB |
| case ARTPSource::RTCP_PSFB: // RTCP PSFB |
| case ARTPSource::RTP_AUTODOWN: |
| { |
| int32_t feedbackType, id; |
| CHECK(msg->findInt32("feedback-type", &feedbackType)); |
| CHECK(msg->findInt32("sender", &id)); |
| in.writeInt32(feedbackType); |
| in.writeInt32(id); |
| if (payloadType == ARTPSource::RTCP_TSFB) { |
| int32_t bitrate; |
| CHECK(msg->findInt32("bit-rate", &bitrate)); |
| in.writeInt32(bitrate); |
| } |
| break; |
| } |
| case ARTPSource::RTP_QUALITY: |
| case ARTPSource::RTP_QUALITY_EMC: |
| { |
| int32_t feedbackType, bitrate; |
| int32_t highestSeqNum, baseSeqNum, prevExpected; |
| int32_t numBufRecv, prevNumBufRecv; |
| int32_t latestRtpTime, jbTimeMs, rtpRtcpSrTimeGapMs; |
| int64_t recvTimeUs; |
| CHECK(msg->findInt32("feedback-type", &feedbackType)); |
| CHECK(msg->findInt32("bit-rate", &bitrate)); |
| CHECK(msg->findInt32("highest-seq-num", &highestSeqNum)); |
| CHECK(msg->findInt32("base-seq-num", &baseSeqNum)); |
| CHECK(msg->findInt32("prev-expected", &prevExpected)); |
| CHECK(msg->findInt32("num-buf-recv", &numBufRecv)); |
| CHECK(msg->findInt32("prev-num-buf-recv", &prevNumBufRecv)); |
| CHECK(msg->findInt32("latest-rtp-time", &latestRtpTime)); |
| CHECK(msg->findInt64("recv-time-us", &recvTimeUs)); |
| CHECK(msg->findInt32("rtp-jitter-time-ms", &jbTimeMs)); |
| CHECK(msg->findInt32("rtp-rtcpsr-time-gap-ms", &rtpRtcpSrTimeGapMs)); |
| in.writeInt32(feedbackType); |
| in.writeInt32(bitrate); |
| in.writeInt32(highestSeqNum); |
| in.writeInt32(baseSeqNum); |
| in.writeInt32(prevExpected); |
| in.writeInt32(numBufRecv); |
| in.writeInt32(prevNumBufRecv); |
| in.writeInt32(latestRtpTime); |
| in.writeInt32(recvTimeUs >> 32); |
| in.writeInt32(recvTimeUs & 0xFFFFFFFF); |
| in.writeInt32(jbTimeMs); |
| in.writeInt32(rtpRtcpSrTimeGapMs); |
| break; |
| } |
| case ARTPSource::RTP_CVO: |
| { |
| int32_t cvo; |
| CHECK(msg->findInt32("cvo", &cvo)); |
| in.writeInt32(cvo); |
| break; |
| } |
| default: |
| break; |
| } |
| |
| notifyListener(MEDIA_IMS_RX_NOTICE, 0, 0, &in); |
| } |
| |
| const char *NuPlayer::getDataSourceType() { |
| switch (mDataSourceType) { |
| case DATA_SOURCE_TYPE_HTTP_LIVE: |
| return "HTTPLive"; |
| |
| case DATA_SOURCE_TYPE_RTP: |
| return "RTP"; |
| |
| case DATA_SOURCE_TYPE_RTSP: |
| return "RTSP"; |
| |
| case DATA_SOURCE_TYPE_GENERIC_URL: |
| return "GenURL"; |
| |
| case DATA_SOURCE_TYPE_GENERIC_FD: |
| return "GenFD"; |
| |
| case DATA_SOURCE_TYPE_MEDIA: |
| return "Media"; |
| |
| case DATA_SOURCE_TYPE_STREAM: |
| return "Stream"; |
| |
| case DATA_SOURCE_TYPE_NONE: |
| default: |
| return "None"; |
| } |
| } |
| |
| void NuPlayer::dump(AString& logString) { |
| logString.append("renderer("); |
| if (mRenderer != nullptr) { |
| mRenderer->dump(logString); |
| } else { |
| logString.append("null"); |
| } |
| logString.append(")"); |
| } |
| |
| // Modular DRM begin |
| status_t NuPlayer::prepareDrm(const uint8_t uuid[16], const Vector<uint8_t> &drmSessionId) |
| { |
| ALOGV("prepareDrm "); |
| |
| // Passing to the looper anyway; called in a pre-config prepared state so no race on mCrypto |
| sp<AMessage> msg = new AMessage(kWhatPrepareDrm, this); |
| // synchronous call so just passing the address but with local copies of "const" args |
| uint8_t UUID[16]; |
| memcpy(UUID, uuid, sizeof(UUID)); |
| Vector<uint8_t> sessionId = drmSessionId; |
| msg->setPointer("uuid", (void*)UUID); |
| msg->setPointer("drmSessionId", (void*)&sessionId); |
| |
| sp<AMessage> response; |
| status_t status = msg->postAndAwaitResponse(&response); |
| |
| if (status == OK && response != NULL) { |
| CHECK(response->findInt32("status", &status)); |
| ALOGV("prepareDrm ret: %d ", status); |
| } else { |
| ALOGE("prepareDrm err: %d", status); |
| } |
| |
| return status; |
| } |
| |
| status_t NuPlayer::releaseDrm() |
| { |
| ALOGV("releaseDrm "); |
| |
| sp<AMessage> msg = new AMessage(kWhatReleaseDrm, this); |
| |
| sp<AMessage> response; |
| status_t status = msg->postAndAwaitResponse(&response); |
| |
| if (status == OK && response != NULL) { |
| CHECK(response->findInt32("status", &status)); |
| ALOGV("releaseDrm ret: %d ", status); |
| } else { |
| ALOGE("releaseDrm err: %d", status); |
| } |
| |
| return status; |
| } |
| |
| status_t NuPlayer::onPrepareDrm(const sp<AMessage> &msg) |
| { |
| // TODO change to ALOGV |
| ALOGD("onPrepareDrm "); |
| |
| status_t status = INVALID_OPERATION; |
| if (mSource == NULL) { |
| ALOGE("onPrepareDrm: No source. onPrepareDrm failed with %d.", status); |
| return status; |
| } |
| |
| uint8_t *uuid; |
| Vector<uint8_t> *drmSessionId; |
| CHECK(msg->findPointer("uuid", (void**)&uuid)); |
| CHECK(msg->findPointer("drmSessionId", (void**)&drmSessionId)); |
| |
| status = OK; |
| sp<ICrypto> crypto = NULL; |
| |
| status = mSource->prepareDrm(uuid, *drmSessionId, &crypto); |
| if (crypto == NULL) { |
| ALOGE("onPrepareDrm: mSource->prepareDrm failed. status: %d", status); |
| return status; |
| } |
| ALOGV("onPrepareDrm: mSource->prepareDrm succeeded"); |
| |
| if (mCrypto != NULL) { |
| ALOGE("onPrepareDrm: Unexpected. Already having mCrypto: %p (%d)", |
| mCrypto.get(), mCrypto->getStrongCount()); |
| mCrypto.clear(); |
| } |
| |
| mCrypto = crypto; |
| mIsDrmProtected = true; |
| // TODO change to ALOGV |
| ALOGD("onPrepareDrm: mCrypto: %p (%d)", mCrypto.get(), |
| (mCrypto != NULL ? mCrypto->getStrongCount() : 0)); |
| |
| return status; |
| } |
| |
| status_t NuPlayer::onReleaseDrm() |
| { |
| // TODO change to ALOGV |
| ALOGD("onReleaseDrm "); |
| |
| if (!mIsDrmProtected) { |
| ALOGW("onReleaseDrm: Unexpected. mIsDrmProtected is already false."); |
| } |
| |
| mIsDrmProtected = false; |
| |
| status_t status; |
| if (mCrypto != NULL) { |
| // notifying the source first before removing crypto from codec |
| if (mSource != NULL) { |
| mSource->releaseDrm(); |
| } |
| |
| status=OK; |
| // first making sure the codecs have released their crypto reference |
| const sp<DecoderBase> &videoDecoder = getDecoder(false/*audio*/); |
| if (videoDecoder != NULL) { |
| status = videoDecoder->releaseCrypto(); |
| ALOGV("onReleaseDrm: video decoder ret: %d", status); |
| } |
| |
| const sp<DecoderBase> &audioDecoder = getDecoder(true/*audio*/); |
| if (audioDecoder != NULL) { |
| status_t status_audio = audioDecoder->releaseCrypto(); |
| if (status == OK) { // otherwise, returning the first error |
| status = status_audio; |
| } |
| ALOGV("onReleaseDrm: audio decoder ret: %d", status_audio); |
| } |
| |
| // TODO change to ALOGV |
| ALOGD("onReleaseDrm: mCrypto: %p (%d)", mCrypto.get(), |
| (mCrypto != NULL ? mCrypto->getStrongCount() : 0)); |
| mCrypto.clear(); |
| } else { // mCrypto == NULL |
| ALOGE("onReleaseDrm: Unexpected. There is no crypto."); |
| status = INVALID_OPERATION; |
| } |
| |
| return status; |
| } |
| // Modular DRM end |
| //////////////////////////////////////////////////////////////////////////////// |
| |
| sp<AMessage> NuPlayer::Source::getFormat(bool audio) { |
| sp<MetaData> meta = getFormatMeta(audio); |
| |
| if (meta == NULL) { |
| return NULL; |
| } |
| |
| sp<AMessage> msg = new AMessage; |
| |
| if(convertMetaDataToMessage(meta, &msg) == OK) { |
| return msg; |
| } |
| return NULL; |
| } |
| |
| void NuPlayer::Source::notifyFlagsChanged(uint32_t flags) { |
| sp<AMessage> notify = dupNotify(); |
| notify->setInt32("what", kWhatFlagsChanged); |
| notify->setInt32("flags", flags); |
| notify->post(); |
| } |
| |
| void NuPlayer::Source::notifyVideoSizeChanged(const sp<AMessage> &format) { |
| sp<AMessage> notify = dupNotify(); |
| notify->setInt32("what", kWhatVideoSizeChanged); |
| notify->setMessage("format", format); |
| notify->post(); |
| } |
| |
| void NuPlayer::Source::notifyPrepared(status_t err) { |
| ALOGV("Source::notifyPrepared %d", err); |
| sp<AMessage> notify = dupNotify(); |
| notify->setInt32("what", kWhatPrepared); |
| notify->setInt32("err", err); |
| notify->post(); |
| } |
| |
| void NuPlayer::Source::notifyDrmInfo(const sp<ABuffer> &drmInfoBuffer) |
| { |
| ALOGV("Source::notifyDrmInfo"); |
| |
| sp<AMessage> notify = dupNotify(); |
| notify->setInt32("what", kWhatDrmInfo); |
| notify->setBuffer("drmInfo", drmInfoBuffer); |
| |
| notify->post(); |
| } |
| |
| void NuPlayer::Source::notifyInstantiateSecureDecoders(const sp<AMessage> &reply) { |
| sp<AMessage> notify = dupNotify(); |
| notify->setInt32("what", kWhatInstantiateSecureDecoders); |
| notify->setMessage("reply", reply); |
| notify->post(); |
| } |
| |
| void NuPlayer::Source::onMessageReceived(const sp<AMessage> & /* msg */) { |
| TRESPASS(); |
| } |
| |
| } // namespace android |