| /* |
| * Copyright (C) 2011 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| |
| // #define LOG_NDEBUG 0 |
| #define LOG_TAG "PreviewPlayer" |
| #include <utils/Log.h> |
| |
| #include <binder/IPCThreadState.h> |
| #include <binder/IServiceManager.h> |
| #include <media/IMediaPlayerService.h> |
| #include <media/stagefright/DataSource.h> |
| #include <media/stagefright/MediaBuffer.h> |
| #include <media/stagefright/MediaDefs.h> |
| #include <media/stagefright/MediaExtractor.h> |
| #include <media/stagefright/MediaSource.h> |
| #include <media/stagefright/MetaData.h> |
| #include <media/stagefright/OMXCodec.h> |
| #include <media/stagefright/foundation/ADebug.h> |
| #include <gui/Surface.h> |
| #include <gui/ISurfaceTexture.h> |
| #include <gui/SurfaceTextureClient.h> |
| |
| #include "VideoEditorPreviewController.h" |
| #include "DummyAudioSource.h" |
| #include "DummyVideoSource.h" |
| #include "VideoEditorSRC.h" |
| #include "PreviewPlayer.h" |
| |
| namespace android { |
| |
| |
| void addBatteryData(uint32_t params) { |
| sp<IBinder> binder = |
| defaultServiceManager()->getService(String16("media.player")); |
| sp<IMediaPlayerService> service = interface_cast<IMediaPlayerService>(binder); |
| CHECK(service.get() != NULL); |
| |
| service->addBatteryData(params); |
| } |
| |
| struct PreviewPlayerEvent : public TimedEventQueue::Event { |
| PreviewPlayerEvent( |
| PreviewPlayer *player, |
| void (PreviewPlayer::*method)()) |
| : mPlayer(player), |
| mMethod(method) { |
| } |
| |
| protected: |
| virtual ~PreviewPlayerEvent() {} |
| |
| virtual void fire(TimedEventQueue *queue, int64_t /* now_us */) { |
| (mPlayer->*mMethod)(); |
| } |
| |
| private: |
| PreviewPlayer *mPlayer; |
| void (PreviewPlayer::*mMethod)(); |
| |
| PreviewPlayerEvent(const PreviewPlayerEvent &); |
| PreviewPlayerEvent &operator=(const PreviewPlayerEvent &); |
| }; |
| |
| PreviewPlayer::PreviewPlayer(NativeWindowRenderer* renderer) |
| : mQueueStarted(false), |
| mTimeSource(NULL), |
| mVideoRendererIsPreview(false), |
| mAudioPlayer(NULL), |
| mDisplayWidth(0), |
| mDisplayHeight(0), |
| mFlags(0), |
| mExtractorFlags(0), |
| mVideoBuffer(NULL), |
| mLastVideoTimeUs(-1), |
| mNativeWindowRenderer(renderer), |
| mCurrFramingEffectIndex(0), |
| mFrameRGBBuffer(NULL), |
| mFrameYUVBuffer(NULL) { |
| |
| CHECK_EQ(mClient.connect(), (status_t)OK); |
| DataSource::RegisterDefaultSniffers(); |
| |
| |
| mVideoRenderer = NULL; |
| mEffectsSettings = NULL; |
| mAudioPlayer = NULL; |
| mAudioMixStoryBoardTS = 0; |
| mCurrentMediaBeginCutTime = 0; |
| mCurrentMediaVolumeValue = 0; |
| mNumberEffects = 0; |
| mDecodedVideoTs = 0; |
| mDecVideoTsStoryBoard = 0; |
| mCurrentVideoEffect = VIDEO_EFFECT_NONE; |
| mProgressCbInterval = 0; |
| mNumberDecVideoFrames = 0; |
| mOverlayUpdateEventPosted = false; |
| mIsChangeSourceRequired = true; |
| |
| mVideoEvent = new PreviewPlayerEvent(this, &PreviewPlayer::onVideoEvent); |
| mVideoEventPending = false; |
| mVideoLagEvent = new PreviewPlayerEvent(this, &PreviewPlayer::onVideoLagUpdate); |
| mVideoEventPending = false; |
| mCheckAudioStatusEvent = new PreviewPlayerEvent( |
| this, &PreviewPlayer::onCheckAudioStatus); |
| mAudioStatusEventPending = false; |
| mStreamDoneEvent = new PreviewPlayerEvent( |
| this, &PreviewPlayer::onStreamDone); |
| mStreamDoneEventPending = false; |
| mProgressCbEvent = new PreviewPlayerEvent(this, |
| &PreviewPlayer::onProgressCbEvent); |
| |
| mOverlayUpdateEvent = new PreviewPlayerEvent(this, |
| &PreviewPlayer::onUpdateOverlayEvent); |
| mProgressCbEventPending = false; |
| |
| mOverlayUpdateEventPending = false; |
| mRenderingMode = (M4xVSS_MediaRendering)MEDIA_RENDERING_INVALID; |
| mIsFiftiesEffectStarted = false; |
| reset(); |
| } |
| |
| PreviewPlayer::~PreviewPlayer() { |
| |
| if (mQueueStarted) { |
| mQueue.stop(); |
| } |
| |
| reset(); |
| |
| if (mVideoRenderer) { |
| mNativeWindowRenderer->destroyRenderInput(mVideoRenderer); |
| } |
| |
| Mutex::Autolock lock(mLock); |
| clear_l(); |
| mClient.disconnect(); |
| } |
| |
| void PreviewPlayer::cancelPlayerEvents_l(bool updateProgressCb) { |
| mQueue.cancelEvent(mVideoEvent->eventID()); |
| mVideoEventPending = false; |
| mQueue.cancelEvent(mStreamDoneEvent->eventID()); |
| mStreamDoneEventPending = false; |
| mQueue.cancelEvent(mCheckAudioStatusEvent->eventID()); |
| mAudioStatusEventPending = false; |
| mQueue.cancelEvent(mVideoLagEvent->eventID()); |
| mVideoLagEventPending = false; |
| if (updateProgressCb) { |
| mQueue.cancelEvent(mProgressCbEvent->eventID()); |
| mProgressCbEventPending = false; |
| } |
| } |
| |
| status_t PreviewPlayer::setDataSource(const char *path) { |
| Mutex::Autolock autoLock(mLock); |
| return setDataSource_l(path); |
| } |
| |
| status_t PreviewPlayer::setDataSource_l(const char *path) { |
| reset_l(); |
| |
| mUri = path; |
| |
| // The actual work will be done during preparation in the call to |
| // ::finishSetDataSource_l to avoid blocking the calling thread in |
| // setDataSource for any significant time. |
| return OK; |
| } |
| |
| status_t PreviewPlayer::setDataSource_l(const sp<MediaExtractor> &extractor) { |
| bool haveAudio = false; |
| bool haveVideo = false; |
| for (size_t i = 0; i < extractor->countTracks(); ++i) { |
| sp<MetaData> meta = extractor->getTrackMetaData(i); |
| |
| const char *mime; |
| CHECK(meta->findCString(kKeyMIMEType, &mime)); |
| |
| if (!haveVideo && !strncasecmp(mime, "video/", 6)) { |
| setVideoSource(extractor->getTrack(i)); |
| haveVideo = true; |
| } else if (!haveAudio && !strncasecmp(mime, "audio/", 6)) { |
| setAudioSource(extractor->getTrack(i)); |
| haveAudio = true; |
| |
| if (!strcasecmp(mime, MEDIA_MIMETYPE_AUDIO_VORBIS)) { |
| // Only do this for vorbis audio, none of the other audio |
| // formats even support this ringtone specific hack and |
| // retrieving the metadata on some extractors may turn out |
| // to be very expensive. |
| sp<MetaData> fileMeta = extractor->getMetaData(); |
| int32_t loop; |
| if (fileMeta != NULL |
| && fileMeta->findInt32(kKeyAutoLoop, &loop) |
| && loop != 0) { |
| mFlags |= AUTO_LOOPING; |
| } |
| } |
| } |
| |
| if (haveAudio && haveVideo) { |
| break; |
| } |
| } |
| |
| /* Add the support for Dummy audio*/ |
| if( !haveAudio ){ |
| mAudioTrack = DummyAudioSource::Create(32000, 2, 20000, |
| ((mPlayEndTimeMsec)*1000LL)); |
| if(mAudioTrack != NULL) { |
| haveAudio = true; |
| } |
| } |
| |
| if (!haveAudio && !haveVideo) { |
| return UNKNOWN_ERROR; |
| } |
| |
| mExtractorFlags = extractor->flags(); |
| return OK; |
| } |
| |
| status_t PreviewPlayer::setDataSource_l_jpg() { |
| ALOGV("setDataSource_l_jpg"); |
| |
| M4OSA_ERR err = M4NO_ERROR; |
| |
| mAudioSource = DummyAudioSource::Create(32000, 2, 20000, |
| ((mPlayEndTimeMsec)*1000LL)); |
| if(mAudioSource != NULL) { |
| setAudioSource(mAudioSource); |
| } |
| status_t error = mAudioSource->start(); |
| if (error != OK) { |
| ALOGE("Error starting dummy audio source"); |
| mAudioSource.clear(); |
| return err; |
| } |
| |
| mDurationUs = (mPlayEndTimeMsec - mPlayBeginTimeMsec)*1000LL; |
| |
| mVideoSource = DummyVideoSource::Create(mVideoWidth, mVideoHeight, |
| mDurationUs, mUri); |
| |
| updateSizeToRender(mVideoSource->getFormat()); |
| setVideoSource(mVideoSource); |
| status_t err1 = mVideoSource->start(); |
| if (err1 != OK) { |
| mVideoSource.clear(); |
| return err; |
| } |
| |
| mIsVideoSourceJpg = true; |
| return OK; |
| } |
| |
| void PreviewPlayer::reset_l() { |
| |
| if (mFlags & PREPARING) { |
| mFlags |= PREPARE_CANCELLED; |
| } |
| |
| while (mFlags & PREPARING) { |
| mPreparedCondition.wait(mLock); |
| } |
| |
| cancelPlayerEvents_l(); |
| mAudioTrack.clear(); |
| mVideoTrack.clear(); |
| |
| // Shutdown audio first, so that the respone to the reset request |
| // appears to happen instantaneously as far as the user is concerned |
| // If we did this later, audio would continue playing while we |
| // shutdown the video-related resources and the player appear to |
| // not be as responsive to a reset request. |
| if (mAudioPlayer == NULL && mAudioSource != NULL) { |
| // If we had an audio player, it would have effectively |
| // taken possession of the audio source and stopped it when |
| // _it_ is stopped. Otherwise this is still our responsibility. |
| mAudioSource->stop(); |
| } |
| mAudioSource.clear(); |
| |
| mTimeSource = NULL; |
| |
| //Single audio player instance used |
| //So donot delete it here |
| //It is deleted from PreviewController class |
| //delete mAudioPlayer; |
| mAudioPlayer = NULL; |
| |
| if (mVideoBuffer) { |
| mVideoBuffer->release(); |
| mVideoBuffer = NULL; |
| } |
| |
| if (mVideoSource != NULL) { |
| mVideoSource->stop(); |
| |
| // The following hack is necessary to ensure that the OMX |
| // component is completely released by the time we may try |
| // to instantiate it again. |
| wp<MediaSource> tmp = mVideoSource; |
| mVideoSource.clear(); |
| while (tmp.promote() != NULL) { |
| usleep(1000); |
| } |
| IPCThreadState::self()->flushCommands(); |
| } |
| |
| mDurationUs = -1; |
| mFlags = 0; |
| mExtractorFlags = 0; |
| mVideoWidth = mVideoHeight = -1; |
| mTimeSourceDeltaUs = 0; |
| mVideoTimeUs = 0; |
| |
| mSeeking = NO_SEEK; |
| mSeekNotificationSent = false; |
| mSeekTimeUs = 0; |
| |
| mUri.setTo(""); |
| |
| mCurrentVideoEffect = VIDEO_EFFECT_NONE; |
| mIsVideoSourceJpg = false; |
| mFrameRGBBuffer = NULL; |
| if(mFrameYUVBuffer != NULL) { |
| free(mFrameYUVBuffer); |
| mFrameYUVBuffer = NULL; |
| } |
| } |
| |
| status_t PreviewPlayer::play() { |
| ALOGV("play"); |
| Mutex::Autolock autoLock(mLock); |
| |
| mFlags &= ~CACHE_UNDERRUN; |
| mFlags &= ~INFORMED_AV_EOS; |
| return play_l(); |
| } |
| |
| status_t PreviewPlayer::startAudioPlayer_l() { |
| ALOGV("startAudioPlayer_l"); |
| CHECK(!(mFlags & AUDIO_RUNNING)); |
| |
| if (mAudioSource == NULL || mAudioPlayer == NULL) { |
| return OK; |
| } |
| |
| if (!(mFlags & AUDIOPLAYER_STARTED)) { |
| mFlags |= AUDIOPLAYER_STARTED; |
| |
| // We've already started the MediaSource in order to enable |
| // the prefetcher to read its data. |
| status_t err = mAudioPlayer->start( |
| true /* sourceAlreadyStarted */); |
| |
| if (err != OK) { |
| notifyListener_l(MEDIA_ERROR, MEDIA_ERROR_UNKNOWN, err); |
| return err; |
| } |
| } else { |
| mAudioPlayer->resume(); |
| } |
| |
| mFlags |= AUDIO_RUNNING; |
| |
| mWatchForAudioEOS = true; |
| |
| return OK; |
| } |
| |
| status_t PreviewPlayer::setAudioPlayer(VideoEditorAudioPlayer *audioPlayer) { |
| ALOGV("setAudioPlayer"); |
| Mutex::Autolock autoLock(mLock); |
| CHECK(!(mFlags & PLAYING)); |
| mAudioPlayer = audioPlayer; |
| |
| ALOGV("SetAudioPlayer"); |
| mIsChangeSourceRequired = true; |
| |
| // check if the new and old source are dummy |
| sp<MediaSource> anAudioSource = mAudioPlayer->getSource(); |
| if (anAudioSource == NULL) { |
| // Audio player does not have any source set. |
| ALOGV("setAudioPlayer: Audio player does not have any source set"); |
| return OK; |
| } |
| |
| // If new video source is not dummy, then always change source |
| // Else audio player continues using old audio source and there are |
| // frame drops to maintain AV sync |
| sp<MetaData> meta; |
| if (mVideoSource != NULL) { |
| meta = mVideoSource->getFormat(); |
| const char *pVidSrcType; |
| if (meta->findCString(kKeyDecoderComponent, &pVidSrcType)) { |
| if (strcmp(pVidSrcType, "DummyVideoSource") != 0) { |
| ALOGV(" Video clip with silent audio; need to change source"); |
| return OK; |
| } |
| } |
| } |
| |
| const char *pSrcType1; |
| const char *pSrcType2; |
| meta = anAudioSource->getFormat(); |
| |
| if (meta->findCString(kKeyDecoderComponent, &pSrcType1)) { |
| if (strcmp(pSrcType1, "DummyAudioSource") == 0) { |
| meta = mAudioSource->getFormat(); |
| if (meta->findCString(kKeyDecoderComponent, &pSrcType2)) { |
| if (strcmp(pSrcType2, "DummyAudioSource") == 0) { |
| mIsChangeSourceRequired = false; |
| // Just set the new play duration for the existing source |
| MediaSource *pMediaSrc = anAudioSource.get(); |
| DummyAudioSource *pDummyAudioSource = (DummyAudioSource*)pMediaSrc; |
| //Increment the duration of audio source |
| pDummyAudioSource->setDuration( |
| (int64_t)((mPlayEndTimeMsec)*1000LL)); |
| |
| // Stop the new audio source |
| // since we continue using old source |
| ALOGV("setAudioPlayer: stop new audio source"); |
| mAudioSource->stop(); |
| } |
| } |
| } |
| } |
| |
| return OK; |
| } |
| |
| void PreviewPlayer::onStreamDone() { |
| ALOGV("onStreamDone"); |
| // Posted whenever any stream finishes playing. |
| |
| Mutex::Autolock autoLock(mLock); |
| if (!mStreamDoneEventPending) { |
| return; |
| } |
| mStreamDoneEventPending = false; |
| |
| if (mStreamDoneStatus != ERROR_END_OF_STREAM) { |
| ALOGV("MEDIA_ERROR %d", mStreamDoneStatus); |
| |
| notifyListener_l( |
| MEDIA_ERROR, MEDIA_ERROR_UNKNOWN, mStreamDoneStatus); |
| |
| pause_l(true /* at eos */); |
| |
| mFlags |= AT_EOS; |
| return; |
| } |
| |
| const bool allDone = |
| (mVideoSource == NULL || (mFlags & VIDEO_AT_EOS)) |
| && (mAudioSource == NULL || (mFlags & AUDIO_AT_EOS)); |
| |
| if (!allDone) { |
| return; |
| } |
| |
| if (mFlags & (LOOPING | AUTO_LOOPING)) { |
| seekTo_l(0); |
| |
| if (mVideoSource != NULL) { |
| postVideoEvent_l(); |
| } |
| } else { |
| ALOGV("MEDIA_PLAYBACK_COMPLETE"); |
| //pause before sending event |
| pause_l(true /* at eos */); |
| |
| //This lock is used to syncronize onStreamDone() in PreviewPlayer and |
| //stopPreview() in PreviewController |
| Mutex::Autolock autoLock(mLockControl); |
| /* Make sure PreviewPlayer only notifies MEDIA_PLAYBACK_COMPLETE once for each clip! |
| * It happens twice in following scenario. |
| * To make the clips in preview storyboard are played and switched smoothly, |
| * PreviewController uses two PreviewPlayer instances and one AudioPlayer. |
| * The two PreviewPlayer use the same AudioPlayer to play the audio, |
| * and change the audio source of the AudioPlayer. |
| * If the audio source of current playing clip and next clip are dummy |
| * audio source(image or video without audio), it will not change the audio source |
| * to avoid the "audio glitch", and keep using the current audio source. |
| * When the video of current clip reached the EOS, PreviewPlayer will set EOS flag |
| * for video and audio, and it will notify MEDIA_PLAYBACK_COMPLETE. |
| * But the audio(dummy audio source) is still playing(for next clip), |
| * and when it reached the EOS, and video reached EOS, |
| * PreviewPlayer will notify MEDIA_PLAYBACK_COMPLETE again. */ |
| if (!(mFlags & INFORMED_AV_EOS)) { |
| notifyListener_l(MEDIA_PLAYBACK_COMPLETE); |
| mFlags |= INFORMED_AV_EOS; |
| } |
| mFlags |= AT_EOS; |
| ALOGV("onStreamDone end"); |
| return; |
| } |
| } |
| |
| |
| status_t PreviewPlayer::play_l() { |
| ALOGV("play_l"); |
| |
| mFlags &= ~SEEK_PREVIEW; |
| |
| if (mFlags & PLAYING) { |
| return OK; |
| } |
| mStartNextPlayer = false; |
| |
| if (!(mFlags & PREPARED)) { |
| status_t err = prepare_l(); |
| |
| if (err != OK) { |
| return err; |
| } |
| } |
| |
| mFlags |= PLAYING; |
| mFlags |= FIRST_FRAME; |
| |
| bool deferredAudioSeek = false; |
| |
| if (mAudioSource != NULL) { |
| if (mAudioPlayer == NULL) { |
| if (mAudioSink != NULL) { |
| |
| mAudioPlayer = new VideoEditorAudioPlayer(mAudioSink, this); |
| mAudioPlayer->setSource(mAudioSource); |
| |
| mAudioPlayer->setAudioMixSettings( |
| mPreviewPlayerAudioMixSettings); |
| |
| mAudioPlayer->setAudioMixPCMFileHandle( |
| mAudioMixPCMFileHandle); |
| |
| mAudioPlayer->setAudioMixStoryBoardSkimTimeStamp( |
| mAudioMixStoryBoardTS, mCurrentMediaBeginCutTime, |
| mCurrentMediaVolumeValue); |
| |
| mFlags |= AUDIOPLAYER_STARTED; |
| // We've already started the MediaSource in order to enable |
| // the prefetcher to read its data. |
| status_t err = mAudioPlayer->start( |
| true /* sourceAlreadyStarted */); |
| |
| if (err != OK) { |
| //delete mAudioPlayer; |
| mAudioPlayer = NULL; |
| |
| mFlags &= ~(PLAYING | FIRST_FRAME); |
| return err; |
| } |
| |
| mTimeSource = mAudioPlayer; |
| mFlags |= AUDIO_RUNNING; |
| deferredAudioSeek = true; |
| mWatchForAudioSeekComplete = false; |
| mWatchForAudioEOS = true; |
| } |
| } else { |
| bool isAudioPlayerStarted = mAudioPlayer->isStarted(); |
| |
| if (mIsChangeSourceRequired == true) { |
| ALOGV("play_l: Change audio source required"); |
| |
| if (isAudioPlayerStarted == true) { |
| mAudioPlayer->pause(); |
| } |
| |
| mAudioPlayer->setSource(mAudioSource); |
| mAudioPlayer->setObserver(this); |
| |
| mAudioPlayer->setAudioMixSettings( |
| mPreviewPlayerAudioMixSettings); |
| |
| mAudioPlayer->setAudioMixStoryBoardSkimTimeStamp( |
| mAudioMixStoryBoardTS, mCurrentMediaBeginCutTime, |
| mCurrentMediaVolumeValue); |
| |
| if (isAudioPlayerStarted == true) { |
| mAudioPlayer->resume(); |
| } else { |
| status_t err = OK; |
| err = mAudioPlayer->start(true); |
| if (err != OK) { |
| mAudioPlayer = NULL; |
| mAudioPlayer = NULL; |
| |
| mFlags &= ~(PLAYING | FIRST_FRAME); |
| return err; |
| } |
| } |
| } else { |
| ALOGV("play_l: No Source change required"); |
| mAudioPlayer->setAudioMixStoryBoardSkimTimeStamp( |
| mAudioMixStoryBoardTS, mCurrentMediaBeginCutTime, |
| mCurrentMediaVolumeValue); |
| |
| mAudioPlayer->resume(); |
| } |
| |
| mFlags |= AUDIOPLAYER_STARTED; |
| mFlags |= AUDIO_RUNNING; |
| mTimeSource = mAudioPlayer; |
| deferredAudioSeek = true; |
| mWatchForAudioSeekComplete = false; |
| mWatchForAudioEOS = true; |
| } |
| } |
| |
| if (mTimeSource == NULL && mAudioPlayer == NULL) { |
| mTimeSource = &mSystemTimeSource; |
| } |
| |
| // Set the seek option for Image source files and read. |
| // This resets the timestamping for image play |
| if (mIsVideoSourceJpg) { |
| MediaSource::ReadOptions options; |
| MediaBuffer *aLocalBuffer; |
| options.setSeekTo(mSeekTimeUs); |
| mVideoSource->read(&aLocalBuffer, &options); |
| aLocalBuffer->release(); |
| } |
| |
| if (mVideoSource != NULL) { |
| // Kick off video playback |
| postVideoEvent_l(); |
| } |
| |
| if (deferredAudioSeek) { |
| // If there was a seek request while we were paused |
| // and we're just starting up again, honor the request now. |
| seekAudioIfNecessary_l(); |
| } |
| |
| if (mFlags & AT_EOS) { |
| // Legacy behaviour, if a stream finishes playing and then |
| // is started again, we play from the start... |
| seekTo_l(0); |
| } |
| |
| return OK; |
| } |
| |
| |
| status_t PreviewPlayer::initRenderer_l() { |
| if (mSurface != NULL) { |
| if(mVideoRenderer == NULL) { |
| mVideoRenderer = mNativeWindowRenderer->createRenderInput(); |
| if (mVideoSource != NULL) { |
| updateSizeToRender(mVideoSource->getFormat()); |
| } |
| } |
| } |
| return OK; |
| } |
| |
| |
| status_t PreviewPlayer::seekTo(int64_t timeUs) { |
| Mutex::Autolock autoLock(mLock); |
| if ((mExtractorFlags & MediaExtractor::CAN_SEEK) || (mIsVideoSourceJpg)) { |
| return seekTo_l(timeUs); |
| } |
| |
| return OK; |
| } |
| |
| |
| status_t PreviewPlayer::getVideoDimensions( |
| int32_t *width, int32_t *height) const { |
| Mutex::Autolock autoLock(mLock); |
| |
| if (mVideoWidth < 0 || mVideoHeight < 0) { |
| return UNKNOWN_ERROR; |
| } |
| |
| *width = mVideoWidth; |
| *height = mVideoHeight; |
| |
| return OK; |
| } |
| |
| |
| status_t PreviewPlayer::initAudioDecoder_l() { |
| sp<MetaData> meta = mAudioTrack->getFormat(); |
| const char *mime; |
| CHECK(meta->findCString(kKeyMIMEType, &mime)); |
| |
| if (!strcasecmp(mime, MEDIA_MIMETYPE_AUDIO_RAW)) { |
| mAudioSource = mAudioTrack; |
| } else { |
| sp<MediaSource> aRawSource; |
| aRawSource = OMXCodec::Create( |
| mClient.interface(), mAudioTrack->getFormat(), |
| false, // createEncoder |
| mAudioTrack); |
| |
| if(aRawSource != NULL) { |
| mAudioSource = new VideoEditorSRC(aRawSource); |
| } |
| } |
| |
| if (mAudioSource != NULL) { |
| int64_t durationUs; |
| if (mAudioTrack->getFormat()->findInt64(kKeyDuration, &durationUs)) { |
| setDuration_l(durationUs); |
| } |
| status_t err = mAudioSource->start(); |
| |
| if (err != OK) { |
| mAudioSource.clear(); |
| return err; |
| } |
| } else if (!strcasecmp(mime, MEDIA_MIMETYPE_AUDIO_QCELP)) { |
| // For legacy reasons we're simply going to ignore the absence |
| // of an audio decoder for QCELP instead of aborting playback |
| // altogether. |
| return OK; |
| } |
| |
| return mAudioSource != NULL ? OK : UNKNOWN_ERROR; |
| } |
| |
| status_t PreviewPlayer::initVideoDecoder_l(uint32_t flags) { |
| initRenderer_l(); |
| |
| if (mVideoRenderer == NULL) { |
| ALOGE("Cannot create renderer"); |
| return UNKNOWN_ERROR; |
| } |
| |
| mVideoSource = OMXCodec::Create( |
| mClient.interface(), mVideoTrack->getFormat(), |
| false, |
| mVideoTrack, |
| NULL, flags, mVideoRenderer->getTargetWindow()); |
| |
| if (mVideoSource != NULL) { |
| int64_t durationUs; |
| if (mVideoTrack->getFormat()->findInt64(kKeyDuration, &durationUs)) { |
| setDuration_l(durationUs); |
| } |
| |
| updateSizeToRender(mVideoTrack->getFormat()); |
| |
| status_t err = mVideoSource->start(); |
| |
| if (err != OK) { |
| mVideoSource.clear(); |
| return err; |
| } |
| } |
| |
| return mVideoSource != NULL ? OK : UNKNOWN_ERROR; |
| } |
| |
| |
| void PreviewPlayer::onVideoEvent() { |
| uint32_t i=0; |
| M4OSA_ERR err1 = M4NO_ERROR; |
| int64_t imageFrameTimeUs = 0; |
| |
| Mutex::Autolock autoLock(mLock); |
| if (!mVideoEventPending) { |
| // The event has been cancelled in reset_l() but had already |
| // been scheduled for execution at that time. |
| return; |
| } |
| mVideoEventPending = false; |
| |
| if (mFlags & SEEK_PREVIEW) { |
| mFlags &= ~SEEK_PREVIEW; |
| return; |
| } |
| |
| TimeSource *ts_st = &mSystemTimeSource; |
| int64_t timeStartUs = ts_st->getRealTimeUs(); |
| |
| if (mSeeking != NO_SEEK) { |
| |
| if(mAudioSource != NULL) { |
| |
| // We're going to seek the video source first, followed by |
| // the audio source. |
| // In order to avoid jumps in the DataSource offset caused by |
| // the audio codec prefetching data from the old locations |
| // while the video codec is already reading data from the new |
| // locations, we'll "pause" the audio source, causing it to |
| // stop reading input data until a subsequent seek. |
| |
| if (mAudioPlayer != NULL && (mFlags & AUDIO_RUNNING)) { |
| mAudioPlayer->pause(); |
| mFlags &= ~AUDIO_RUNNING; |
| } |
| mAudioSource->pause(); |
| } |
| } |
| |
| if (!mVideoBuffer) { |
| MediaSource::ReadOptions options; |
| if (mSeeking != NO_SEEK) { |
| ALOGV("LV PLAYER seeking to %lld us (%.2f secs)", mSeekTimeUs, |
| mSeekTimeUs / 1E6); |
| |
| options.setSeekTo( |
| mSeekTimeUs, MediaSource::ReadOptions::SEEK_CLOSEST); |
| } |
| for (;;) { |
| status_t err = mVideoSource->read(&mVideoBuffer, &options); |
| options.clearSeekTo(); |
| |
| if (err != OK) { |
| CHECK(!mVideoBuffer); |
| |
| if (err == INFO_FORMAT_CHANGED) { |
| ALOGV("LV PLAYER VideoSource signalled format change"); |
| notifyVideoSize_l(); |
| |
| if (mVideoRenderer != NULL) { |
| mVideoRendererIsPreview = false; |
| err = initRenderer_l(); |
| if (err != OK) { |
| postStreamDoneEvent_l(err); |
| } |
| |
| } |
| |
| updateSizeToRender(mVideoSource->getFormat()); |
| continue; |
| } |
| // So video playback is complete, but we may still have |
| // a seek request pending that needs to be applied to the audio track |
| if (mSeeking != NO_SEEK) { |
| ALOGV("video stream ended while seeking!"); |
| } |
| finishSeekIfNecessary(-1); |
| ALOGV("PreviewPlayer: onVideoEvent EOS reached."); |
| mFlags |= VIDEO_AT_EOS; |
| mFlags |= AUDIO_AT_EOS; |
| mOverlayUpdateEventPosted = false; |
| postStreamDoneEvent_l(err); |
| // Set the last decoded timestamp to duration |
| mDecodedVideoTs = (mPlayEndTimeMsec*1000LL); |
| return; |
| } |
| |
| if (mVideoBuffer->range_length() == 0) { |
| // Some decoders, notably the PV AVC software decoder |
| // return spurious empty buffers that we just want to ignore. |
| |
| mVideoBuffer->release(); |
| mVideoBuffer = NULL; |
| continue; |
| } |
| |
| int64_t videoTimeUs; |
| CHECK(mVideoBuffer->meta_data()->findInt64(kKeyTime, &videoTimeUs)); |
| |
| if (mSeeking != NO_SEEK) { |
| if (videoTimeUs < mSeekTimeUs) { |
| // buffers are before seek time |
| // ignore them |
| mVideoBuffer->release(); |
| mVideoBuffer = NULL; |
| continue; |
| } |
| } else { |
| if((videoTimeUs/1000) < mPlayBeginTimeMsec) { |
| // Frames are before begin cut time |
| // Donot render |
| mVideoBuffer->release(); |
| mVideoBuffer = NULL; |
| continue; |
| } |
| } |
| break; |
| } |
| } |
| |
| mNumberDecVideoFrames++; |
| |
| int64_t timeUs; |
| CHECK(mVideoBuffer->meta_data()->findInt64(kKeyTime, &timeUs)); |
| setPosition_l(timeUs); |
| |
| if (!mStartNextPlayer) { |
| int64_t playbackTimeRemaining = (mPlayEndTimeMsec * 1000LL) - timeUs; |
| if (playbackTimeRemaining <= 1500000) { |
| //When less than 1.5 sec of playback left |
| // send notification to start next player |
| |
| mStartNextPlayer = true; |
| notifyListener_l(0xAAAAAAAA); |
| } |
| } |
| |
| SeekType wasSeeking = mSeeking; |
| finishSeekIfNecessary(timeUs); |
| if (mAudioPlayer != NULL && !(mFlags & (AUDIO_RUNNING))) { |
| status_t err = startAudioPlayer_l(); |
| if (err != OK) { |
| ALOGE("Starting the audio player failed w/ err %d", err); |
| return; |
| } |
| } |
| |
| TimeSource *ts = (mFlags & AUDIO_AT_EOS) ? &mSystemTimeSource : mTimeSource; |
| |
| if(ts == NULL) { |
| mVideoBuffer->release(); |
| mVideoBuffer = NULL; |
| return; |
| } |
| |
| if(!mIsVideoSourceJpg) { |
| if (mFlags & FIRST_FRAME) { |
| mFlags &= ~FIRST_FRAME; |
| |
| mTimeSourceDeltaUs = ts->getRealTimeUs() - timeUs; |
| } |
| |
| int64_t realTimeUs, mediaTimeUs; |
| if (!(mFlags & AUDIO_AT_EOS) && mAudioPlayer != NULL |
| && mAudioPlayer->getMediaTimeMapping(&realTimeUs, &mediaTimeUs)) { |
| mTimeSourceDeltaUs = realTimeUs - mediaTimeUs; |
| } |
| |
| int64_t nowUs = ts->getRealTimeUs() - mTimeSourceDeltaUs; |
| |
| int64_t latenessUs = nowUs - timeUs; |
| |
| if (wasSeeking != NO_SEEK) { |
| // Let's display the first frame after seeking right away. |
| latenessUs = 0; |
| } |
| ALOGV("Audio time stamp = %lld and video time stamp = %lld", |
| ts->getRealTimeUs(),timeUs); |
| if (latenessUs > 40000) { |
| // We're more than 40ms late. |
| |
| ALOGV("LV PLAYER we're late by %lld us (%.2f secs)", |
| latenessUs, latenessUs / 1E6); |
| |
| mVideoBuffer->release(); |
| mVideoBuffer = NULL; |
| postVideoEvent_l(0); |
| return; |
| } |
| |
| if (latenessUs < -25000) { |
| // We're more than 25ms early. |
| ALOGV("We're more than 25ms early, lateness %lld", latenessUs); |
| |
| postVideoEvent_l(25000); |
| return; |
| } |
| } |
| |
| if (mVideoRendererIsPreview || mVideoRenderer == NULL) { |
| mVideoRendererIsPreview = false; |
| |
| status_t err = initRenderer_l(); |
| if (err != OK) { |
| postStreamDoneEvent_l(err); |
| } |
| } |
| |
| // If timestamp exceeds endCutTime of clip, donot render |
| if((timeUs/1000) > mPlayEndTimeMsec) { |
| mVideoBuffer->release(); |
| mVideoBuffer = NULL; |
| mFlags |= VIDEO_AT_EOS; |
| mFlags |= AUDIO_AT_EOS; |
| ALOGV("PreviewPlayer: onVideoEvent timeUs > mPlayEndTime; send EOS.."); |
| mOverlayUpdateEventPosted = false; |
| // Set the last decoded timestamp to duration |
| mDecodedVideoTs = (mPlayEndTimeMsec*1000LL); |
| postStreamDoneEvent_l(ERROR_END_OF_STREAM); |
| return; |
| } |
| // Capture the frame timestamp to be rendered |
| mDecodedVideoTs = timeUs; |
| |
| // Post processing to apply video effects |
| for(i=0;i<mNumberEffects;i++) { |
| // First check if effect starttime matches the clip being previewed |
| if((mEffectsSettings[i].uiStartTime < (mDecVideoTsStoryBoard/1000)) || |
| (mEffectsSettings[i].uiStartTime >= |
| ((mDecVideoTsStoryBoard/1000) + mPlayEndTimeMsec - mPlayBeginTimeMsec))) |
| { |
| // This effect doesn't belong to this clip, check next one |
| continue; |
| } |
| // Check if effect applies to this particular frame timestamp |
| if((mEffectsSettings[i].uiStartTime <= |
| (((timeUs+mDecVideoTsStoryBoard)/1000)-mPlayBeginTimeMsec)) && |
| ((mEffectsSettings[i].uiStartTime+mEffectsSettings[i].uiDuration) >= |
| (((timeUs+mDecVideoTsStoryBoard)/1000)-mPlayBeginTimeMsec)) |
| && (mEffectsSettings[i].uiDuration != 0)) { |
| setVideoPostProcessingNode( |
| mEffectsSettings[i].VideoEffectType, TRUE); |
| } |
| else { |
| setVideoPostProcessingNode( |
| mEffectsSettings[i].VideoEffectType, FALSE); |
| } |
| } |
| |
| //Provide the overlay Update indication when there is an overlay effect |
| if (mCurrentVideoEffect & VIDEO_EFFECT_FRAMING) { |
| mCurrentVideoEffect &= ~VIDEO_EFFECT_FRAMING; //never apply framing here. |
| if (!mOverlayUpdateEventPosted) { |
| // Find the effect in effectSettings array |
| M4OSA_UInt32 index; |
| for (index = 0; index < mNumberEffects; index++) { |
| M4OSA_UInt32 timeMs = mDecodedVideoTs/1000; |
| M4OSA_UInt32 timeOffset = mDecVideoTsStoryBoard/1000; |
| if(mEffectsSettings[index].VideoEffectType == |
| (M4VSS3GPP_VideoEffectType)M4xVSS_kVideoEffectType_Framing) { |
| if (((mEffectsSettings[index].uiStartTime + 1) <= |
| timeMs + timeOffset - mPlayBeginTimeMsec) && |
| ((mEffectsSettings[index].uiStartTime - 1 + |
| mEffectsSettings[index].uiDuration) >= |
| timeMs + timeOffset - mPlayBeginTimeMsec)) |
| { |
| break; |
| } |
| } |
| } |
| if (index < mNumberEffects) { |
| mCurrFramingEffectIndex = index; |
| mOverlayUpdateEventPosted = true; |
| postOverlayUpdateEvent_l(); |
| ALOGV("Framing index = %ld", mCurrFramingEffectIndex); |
| } else { |
| ALOGV("No framing effects found"); |
| } |
| } |
| |
| } else if (mOverlayUpdateEventPosted) { |
| //Post the event when the overlay is no more valid |
| ALOGV("Overlay is Done"); |
| mOverlayUpdateEventPosted = false; |
| postOverlayUpdateEvent_l(); |
| } |
| |
| if (mVideoRenderer != NULL) { |
| mVideoRenderer->render(mVideoBuffer, mCurrentVideoEffect, |
| mRenderingMode, mIsVideoSourceJpg); |
| } |
| |
| mVideoBuffer->release(); |
| mVideoBuffer = NULL; |
| |
| // Post progress callback based on callback interval set |
| if(mNumberDecVideoFrames >= mProgressCbInterval) { |
| postProgressCallbackEvent_l(); |
| mNumberDecVideoFrames = 0; // reset counter |
| } |
| |
| // if reached EndCutTime of clip, post EOS event |
| if((timeUs/1000) >= mPlayEndTimeMsec) { |
| ALOGV("PreviewPlayer: onVideoEvent EOS."); |
| mFlags |= VIDEO_AT_EOS; |
| mFlags |= AUDIO_AT_EOS; |
| mOverlayUpdateEventPosted = false; |
| // Set the last decoded timestamp to duration |
| mDecodedVideoTs = (mPlayEndTimeMsec*1000LL); |
| postStreamDoneEvent_l(ERROR_END_OF_STREAM); |
| } |
| else { |
| if ((wasSeeking != NO_SEEK) && (mFlags & SEEK_PREVIEW)) { |
| mFlags &= ~SEEK_PREVIEW; |
| return; |
| } |
| |
| if(!mIsVideoSourceJpg) { |
| postVideoEvent_l(0); |
| } |
| else { |
| postVideoEvent_l(33000); |
| } |
| } |
| } |
| |
| status_t PreviewPlayer::prepare() { |
| ALOGV("prepare"); |
| Mutex::Autolock autoLock(mLock); |
| return prepare_l(); |
| } |
| |
| status_t PreviewPlayer::prepare_l() { |
| ALOGV("prepare_l"); |
| if (mFlags & PREPARED) { |
| return OK; |
| } |
| |
| if (mFlags & PREPARING) { |
| return UNKNOWN_ERROR; |
| } |
| |
| mIsAsyncPrepare = false; |
| status_t err = prepareAsync_l(); |
| |
| if (err != OK) { |
| return err; |
| } |
| |
| while (mFlags & PREPARING) { |
| mPreparedCondition.wait(mLock); |
| } |
| |
| return mPrepareResult; |
| } |
| |
| status_t PreviewPlayer::prepareAsync() { |
| ALOGV("prepareAsync"); |
| Mutex::Autolock autoLock(mLock); |
| return prepareAsync_l(); |
| } |
| |
| status_t PreviewPlayer::prepareAsync_l() { |
| ALOGV("prepareAsync_l"); |
| if (mFlags & PREPARING) { |
| return UNKNOWN_ERROR; // async prepare already pending |
| } |
| |
| if (!mQueueStarted) { |
| mQueue.start(); |
| mQueueStarted = true; |
| } |
| |
| mFlags |= PREPARING; |
| mAsyncPrepareEvent = new PreviewPlayerEvent( |
| this, &PreviewPlayer::onPrepareAsyncEvent); |
| |
| mQueue.postEvent(mAsyncPrepareEvent); |
| |
| return OK; |
| } |
| |
| status_t PreviewPlayer::finishSetDataSource_l() { |
| sp<DataSource> dataSource; |
| sp<MediaExtractor> extractor; |
| |
| dataSource = DataSource::CreateFromURI(mUri.string(), NULL); |
| |
| if (dataSource == NULL) { |
| return UNKNOWN_ERROR; |
| } |
| |
| //If file type is .rgb, then no need to check for Extractor |
| int uriLen = strlen(mUri); |
| int startOffset = uriLen - 4; |
| if(!strncasecmp(mUri+startOffset, ".rgb", 4)) { |
| extractor = NULL; |
| } |
| else { |
| extractor = MediaExtractor::Create(dataSource, |
| MEDIA_MIMETYPE_CONTAINER_MPEG4); |
| } |
| |
| if (extractor == NULL) { |
| ALOGV("finishSetDataSource_l: failed to create extractor"); |
| return setDataSource_l_jpg(); |
| } |
| |
| return setDataSource_l(extractor); |
| } |
| |
| void PreviewPlayer::onPrepareAsyncEvent() { |
| Mutex::Autolock autoLock(mLock); |
| ALOGV("onPrepareAsyncEvent"); |
| |
| if (mFlags & PREPARE_CANCELLED) { |
| ALOGV("prepare was cancelled before doing anything"); |
| abortPrepare(UNKNOWN_ERROR); |
| return; |
| } |
| |
| if (mUri.size() > 0) { |
| status_t err = finishSetDataSource_l(); |
| |
| if (err != OK) { |
| abortPrepare(err); |
| return; |
| } |
| } |
| |
| if (mVideoTrack != NULL && mVideoSource == NULL) { |
| status_t err = initVideoDecoder_l(OMXCodec::kHardwareCodecsOnly); |
| |
| if (err != OK) { |
| abortPrepare(err); |
| return; |
| } |
| } |
| |
| if (mAudioTrack != NULL && mAudioSource == NULL) { |
| status_t err = initAudioDecoder_l(); |
| |
| if (err != OK) { |
| abortPrepare(err); |
| return; |
| } |
| } |
| finishAsyncPrepare_l(); |
| |
| } |
| |
| void PreviewPlayer::finishAsyncPrepare_l() { |
| ALOGV("finishAsyncPrepare_l"); |
| if (mIsAsyncPrepare) { |
| if (mVideoSource == NULL) { |
| notifyListener_l(MEDIA_SET_VIDEO_SIZE, 0, 0); |
| } else { |
| notifyVideoSize_l(); |
| } |
| notifyListener_l(MEDIA_PREPARED); |
| } |
| |
| mPrepareResult = OK; |
| mFlags &= ~(PREPARING|PREPARE_CANCELLED); |
| mFlags |= PREPARED; |
| mAsyncPrepareEvent = NULL; |
| mPreparedCondition.broadcast(); |
| } |
| |
| void PreviewPlayer::acquireLock() { |
| ALOGV("acquireLock"); |
| mLockControl.lock(); |
| } |
| |
| void PreviewPlayer::releaseLock() { |
| ALOGV("releaseLock"); |
| mLockControl.unlock(); |
| } |
| |
| status_t PreviewPlayer::loadEffectsSettings( |
| M4VSS3GPP_EffectSettings* pEffectSettings, int nEffects) { |
| |
| ALOGV("loadEffectsSettings"); |
| mNumberEffects = nEffects; |
| mEffectsSettings = pEffectSettings; |
| return OK; |
| } |
| |
| status_t PreviewPlayer::loadAudioMixSettings( |
| M4xVSS_AudioMixingSettings* pAudioMixSettings) { |
| |
| ALOGV("loadAudioMixSettings"); |
| mPreviewPlayerAudioMixSettings = pAudioMixSettings; |
| return OK; |
| } |
| |
| status_t PreviewPlayer::setAudioMixPCMFileHandle( |
| M4OSA_Context pAudioMixPCMFileHandle) { |
| |
| ALOGV("setAudioMixPCMFileHandle"); |
| mAudioMixPCMFileHandle = pAudioMixPCMFileHandle; |
| return OK; |
| } |
| |
| status_t PreviewPlayer::setAudioMixStoryBoardParam( |
| M4OSA_UInt32 audioMixStoryBoardTS, |
| M4OSA_UInt32 currentMediaBeginCutTime, |
| M4OSA_UInt32 primaryTrackVolValue ) { |
| |
| ALOGV("setAudioMixStoryBoardParam"); |
| mAudioMixStoryBoardTS = audioMixStoryBoardTS; |
| mCurrentMediaBeginCutTime = currentMediaBeginCutTime; |
| mCurrentMediaVolumeValue = primaryTrackVolValue; |
| return OK; |
| } |
| |
| status_t PreviewPlayer::setPlaybackBeginTime(uint32_t msec) { |
| |
| mPlayBeginTimeMsec = msec; |
| return OK; |
| } |
| |
| status_t PreviewPlayer::setPlaybackEndTime(uint32_t msec) { |
| |
| mPlayEndTimeMsec = msec; |
| return OK; |
| } |
| |
| status_t PreviewPlayer::setStoryboardStartTime(uint32_t msec) { |
| |
| mStoryboardStartTimeMsec = msec; |
| mDecVideoTsStoryBoard = mStoryboardStartTimeMsec * 1000LL; |
| return OK; |
| } |
| |
| status_t PreviewPlayer::setProgressCallbackInterval(uint32_t cbInterval) { |
| |
| mProgressCbInterval = cbInterval; |
| return OK; |
| } |
| |
| |
| status_t PreviewPlayer::setMediaRenderingMode( |
| M4xVSS_MediaRendering mode, |
| M4VIDEOEDITING_VideoFrameSize outputVideoSize) { |
| |
| mRenderingMode = mode; |
| |
| /* get the video width and height by resolution */ |
| return getVideoSizeByResolution( |
| outputVideoSize, |
| &mOutputVideoWidth, &mOutputVideoHeight); |
| |
| } |
| |
| status_t PreviewPlayer::resetJniCallbackTimeStamp() { |
| |
| mDecVideoTsStoryBoard = mStoryboardStartTimeMsec * 1000LL; |
| return OK; |
| } |
| |
| void PreviewPlayer::postProgressCallbackEvent_l() { |
| if (mProgressCbEventPending) { |
| return; |
| } |
| mProgressCbEventPending = true; |
| |
| mQueue.postEvent(mProgressCbEvent); |
| } |
| |
| |
| void PreviewPlayer::onProgressCbEvent() { |
| Mutex::Autolock autoLock(mLock); |
| if (!mProgressCbEventPending) { |
| return; |
| } |
| mProgressCbEventPending = false; |
| // If playback starts from previous I-frame, |
| // then send frame storyboard duration |
| if ((mDecodedVideoTs/1000) < mPlayBeginTimeMsec) { |
| notifyListener_l(MEDIA_INFO, 0, mDecVideoTsStoryBoard/1000); |
| } else { |
| notifyListener_l(MEDIA_INFO, 0, |
| (((mDecodedVideoTs+mDecVideoTsStoryBoard)/1000)-mPlayBeginTimeMsec)); |
| } |
| } |
| |
| void PreviewPlayer::postOverlayUpdateEvent_l() { |
| if (mOverlayUpdateEventPending) { |
| return; |
| } |
| mOverlayUpdateEventPending = true; |
| mQueue.postEvent(mOverlayUpdateEvent); |
| } |
| |
| void PreviewPlayer::onUpdateOverlayEvent() { |
| Mutex::Autolock autoLock(mLock); |
| |
| if (!mOverlayUpdateEventPending) { |
| return; |
| } |
| mOverlayUpdateEventPending = false; |
| |
| int updateState = mOverlayUpdateEventPosted? 1: 0; |
| notifyListener_l(0xBBBBBBBB, updateState, mCurrFramingEffectIndex); |
| } |
| |
| |
| void PreviewPlayer::setVideoPostProcessingNode( |
| M4VSS3GPP_VideoEffectType type, M4OSA_Bool enable) { |
| |
| uint32_t effect = VIDEO_EFFECT_NONE; |
| |
| //Map M4VSS3GPP_VideoEffectType to local enum |
| switch(type) { |
| case M4VSS3GPP_kVideoEffectType_FadeFromBlack: |
| effect = VIDEO_EFFECT_FADEFROMBLACK; |
| break; |
| |
| case M4VSS3GPP_kVideoEffectType_FadeToBlack: |
| effect = VIDEO_EFFECT_FADETOBLACK; |
| break; |
| |
| case M4xVSS_kVideoEffectType_BlackAndWhite: |
| effect = VIDEO_EFFECT_BLACKANDWHITE; |
| break; |
| |
| case M4xVSS_kVideoEffectType_Pink: |
| effect = VIDEO_EFFECT_PINK; |
| break; |
| |
| case M4xVSS_kVideoEffectType_Green: |
| effect = VIDEO_EFFECT_GREEN; |
| break; |
| |
| case M4xVSS_kVideoEffectType_Sepia: |
| effect = VIDEO_EFFECT_SEPIA; |
| break; |
| |
| case M4xVSS_kVideoEffectType_Negative: |
| effect = VIDEO_EFFECT_NEGATIVE; |
| break; |
| |
| case M4xVSS_kVideoEffectType_Framing: |
| effect = VIDEO_EFFECT_FRAMING; |
| break; |
| |
| case M4xVSS_kVideoEffectType_Fifties: |
| effect = VIDEO_EFFECT_FIFTIES; |
| break; |
| |
| case M4xVSS_kVideoEffectType_ColorRGB16: |
| effect = VIDEO_EFFECT_COLOR_RGB16; |
| break; |
| |
| case M4xVSS_kVideoEffectType_Gradient: |
| effect = VIDEO_EFFECT_GRADIENT; |
| break; |
| |
| default: |
| effect = VIDEO_EFFECT_NONE; |
| break; |
| } |
| |
| if (enable == M4OSA_TRUE) { |
| //If already set, then no need to set again |
| if (!(mCurrentVideoEffect & effect)) { |
| mCurrentVideoEffect |= effect; |
| if (effect == VIDEO_EFFECT_FIFTIES) { |
| mIsFiftiesEffectStarted = true; |
| } |
| } |
| } else { |
| //Reset only if already set |
| if (mCurrentVideoEffect & effect) { |
| mCurrentVideoEffect &= ~effect; |
| } |
| } |
| } |
| |
| status_t PreviewPlayer::setImageClipProperties(uint32_t width,uint32_t height) { |
| mVideoWidth = width; |
| mVideoHeight = height; |
| return OK; |
| } |
| |
| status_t PreviewPlayer::readFirstVideoFrame() { |
| ALOGV("readFirstVideoFrame"); |
| |
| if (!mVideoBuffer) { |
| MediaSource::ReadOptions options; |
| if (mSeeking != NO_SEEK) { |
| ALOGV("seeking to %lld us (%.2f secs)", mSeekTimeUs, |
| mSeekTimeUs / 1E6); |
| |
| options.setSeekTo( |
| mSeekTimeUs, MediaSource::ReadOptions::SEEK_CLOSEST); |
| } |
| for (;;) { |
| status_t err = mVideoSource->read(&mVideoBuffer, &options); |
| options.clearSeekTo(); |
| |
| if (err != OK) { |
| CHECK(!mVideoBuffer); |
| |
| if (err == INFO_FORMAT_CHANGED) { |
| ALOGV("VideoSource signalled format change"); |
| notifyVideoSize_l(); |
| |
| if (mVideoRenderer != NULL) { |
| mVideoRendererIsPreview = false; |
| err = initRenderer_l(); |
| if (err != OK) { |
| postStreamDoneEvent_l(err); |
| } |
| } |
| |
| updateSizeToRender(mVideoSource->getFormat()); |
| continue; |
| } |
| ALOGV("EOS reached."); |
| mFlags |= VIDEO_AT_EOS; |
| mFlags |= AUDIO_AT_EOS; |
| postStreamDoneEvent_l(err); |
| return OK; |
| } |
| |
| if (mVideoBuffer->range_length() == 0) { |
| // Some decoders, notably the PV AVC software decoder |
| // return spurious empty buffers that we just want to ignore. |
| |
| mVideoBuffer->release(); |
| mVideoBuffer = NULL; |
| continue; |
| } |
| |
| int64_t videoTimeUs; |
| CHECK(mVideoBuffer->meta_data()->findInt64(kKeyTime, &videoTimeUs)); |
| if (mSeeking != NO_SEEK) { |
| if (videoTimeUs < mSeekTimeUs) { |
| // buffers are before seek time |
| // ignore them |
| mVideoBuffer->release(); |
| mVideoBuffer = NULL; |
| continue; |
| } |
| } else { |
| if ((videoTimeUs/1000) < mPlayBeginTimeMsec) { |
| // buffers are before begin cut time |
| // ignore them |
| mVideoBuffer->release(); |
| mVideoBuffer = NULL; |
| continue; |
| } |
| } |
| break; |
| } |
| } |
| |
| int64_t timeUs; |
| CHECK(mVideoBuffer->meta_data()->findInt64(kKeyTime, &timeUs)); |
| setPosition_l(timeUs); |
| |
| mDecodedVideoTs = timeUs; |
| |
| return OK; |
| |
| } |
| |
| status_t PreviewPlayer::getLastRenderedTimeMs(uint32_t *lastRenderedTimeMs) { |
| *lastRenderedTimeMs = (((mDecodedVideoTs+mDecVideoTsStoryBoard)/1000)-mPlayBeginTimeMsec); |
| return OK; |
| } |
| |
| void PreviewPlayer::updateSizeToRender(sp<MetaData> meta) { |
| if (mVideoRenderer) { |
| mVideoRenderer->updateVideoSize(meta); |
| } |
| } |
| |
| void PreviewPlayer::setListener(const wp<MediaPlayerBase> &listener) { |
| Mutex::Autolock autoLock(mLock); |
| mListener = listener; |
| } |
| |
| status_t PreviewPlayer::setDataSource(const sp<IStreamSource> &source) { |
| return INVALID_OPERATION; |
| } |
| |
| void PreviewPlayer::reset() { |
| Mutex::Autolock autoLock(mLock); |
| reset_l(); |
| } |
| |
| void PreviewPlayer::clear_l() { |
| mDisplayWidth = 0; |
| mDisplayHeight = 0; |
| |
| if (mFlags & PLAYING) { |
| updateBatteryUsage_l(); |
| } |
| |
| if (mFlags & PREPARING) { |
| mFlags |= PREPARE_CANCELLED; |
| |
| if (mFlags & PREPARING_CONNECTED) { |
| // We are basically done preparing, we're just buffering |
| // enough data to start playback, we can safely interrupt that. |
| finishAsyncPrepare_l(); |
| } |
| } |
| |
| while (mFlags & PREPARING) { |
| mPreparedCondition.wait(mLock); |
| } |
| |
| cancelPlayerEvents_l(true); |
| |
| mAudioTrack.clear(); |
| mVideoTrack.clear(); |
| |
| // Shutdown audio first, so that the respone to the reset request |
| // appears to happen instantaneously as far as the user is concerned |
| // If we did this later, audio would continue playing while we |
| // shutdown the video-related resources and the player appear to |
| // not be as responsive to a reset request. |
| if (mAudioPlayer == NULL && mAudioSource != NULL) { |
| // If we had an audio player, it would have effectively |
| // taken possession of the audio source and stopped it when |
| // _it_ is stopped. Otherwise this is still our responsibility. |
| mAudioSource->stop(); |
| } |
| mAudioSource.clear(); |
| |
| mTimeSource = NULL; |
| |
| delete mAudioPlayer; |
| mAudioPlayer = NULL; |
| |
| if (mVideoSource != NULL) { |
| shutdownVideoDecoder_l(); |
| } |
| |
| mDurationUs = -1; |
| mFlags = 0; |
| mExtractorFlags = 0; |
| mTimeSourceDeltaUs = 0; |
| mVideoTimeUs = 0; |
| |
| mSeeking = NO_SEEK; |
| mSeekNotificationSent = false; |
| mSeekTimeUs = 0; |
| |
| mUri.setTo(""); |
| |
| mBitrate = -1; |
| mLastVideoTimeUs = -1; |
| } |
| |
| void PreviewPlayer::notifyListener_l(int msg, int ext1, int ext2) { |
| if (mListener != NULL) { |
| sp<MediaPlayerBase> listener = mListener.promote(); |
| |
| if (listener != NULL) { |
| listener->sendEvent(msg, ext1, ext2); |
| } |
| } |
| } |
| |
| void PreviewPlayer::onVideoLagUpdate() { |
| Mutex::Autolock autoLock(mLock); |
| if (!mVideoLagEventPending) { |
| return; |
| } |
| mVideoLagEventPending = false; |
| |
| int64_t audioTimeUs = mAudioPlayer->getMediaTimeUs(); |
| int64_t videoLateByUs = audioTimeUs - mVideoTimeUs; |
| |
| if (!(mFlags & VIDEO_AT_EOS) && videoLateByUs > 300000ll) { |
| ALOGV("video late by %lld ms.", videoLateByUs / 1000ll); |
| |
| notifyListener_l( |
| MEDIA_INFO, |
| MEDIA_INFO_VIDEO_TRACK_LAGGING, |
| videoLateByUs / 1000ll); |
| } |
| |
| postVideoLagEvent_l(); |
| } |
| |
| void PreviewPlayer::notifyVideoSize_l() { |
| sp<MetaData> meta = mVideoSource->getFormat(); |
| |
| int32_t vWidth, vHeight; |
| int32_t cropLeft, cropTop, cropRight, cropBottom; |
| |
| CHECK(meta->findInt32(kKeyWidth, &vWidth)); |
| CHECK(meta->findInt32(kKeyHeight, &vHeight)); |
| |
| mGivenWidth = vWidth; |
| mGivenHeight = vHeight; |
| |
| if (!meta->findRect( |
| kKeyCropRect, &cropLeft, &cropTop, &cropRight, &cropBottom)) { |
| |
| cropLeft = cropTop = 0; |
| cropRight = vWidth - 1; |
| cropBottom = vHeight - 1; |
| |
| ALOGD("got dimensions only %d x %d", vWidth, vHeight); |
| } else { |
| ALOGD("got crop rect %d, %d, %d, %d", |
| cropLeft, cropTop, cropRight, cropBottom); |
| } |
| |
| mCropRect.left = cropLeft; |
| mCropRect.right = cropRight; |
| mCropRect.top = cropTop; |
| mCropRect.bottom = cropBottom; |
| |
| int32_t displayWidth; |
| if (meta->findInt32(kKeyDisplayWidth, &displayWidth)) { |
| ALOGV("Display width changed (%d=>%d)", mDisplayWidth, displayWidth); |
| mDisplayWidth = displayWidth; |
| } |
| int32_t displayHeight; |
| if (meta->findInt32(kKeyDisplayHeight, &displayHeight)) { |
| ALOGV("Display height changed (%d=>%d)", mDisplayHeight, displayHeight); |
| mDisplayHeight = displayHeight; |
| } |
| |
| int32_t usableWidth = cropRight - cropLeft + 1; |
| int32_t usableHeight = cropBottom - cropTop + 1; |
| if (mDisplayWidth != 0) { |
| usableWidth = mDisplayWidth; |
| } |
| if (mDisplayHeight != 0) { |
| usableHeight = mDisplayHeight; |
| } |
| |
| int32_t rotationDegrees; |
| if (!mVideoTrack->getFormat()->findInt32( |
| kKeyRotation, &rotationDegrees)) { |
| rotationDegrees = 0; |
| } |
| |
| if (rotationDegrees == 90 || rotationDegrees == 270) { |
| notifyListener_l( |
| MEDIA_SET_VIDEO_SIZE, usableHeight, usableWidth); |
| } else { |
| notifyListener_l( |
| MEDIA_SET_VIDEO_SIZE, usableWidth, usableHeight); |
| } |
| } |
| |
| status_t PreviewPlayer::pause() { |
| Mutex::Autolock autoLock(mLock); |
| |
| mFlags &= ~CACHE_UNDERRUN; |
| |
| return pause_l(); |
| } |
| |
| status_t PreviewPlayer::pause_l(bool at_eos) { |
| if (!(mFlags & PLAYING)) { |
| return OK; |
| } |
| |
| cancelPlayerEvents_l(); |
| |
| if (mAudioPlayer != NULL && (mFlags & AUDIO_RUNNING)) { |
| if (at_eos) { |
| // If we played the audio stream to completion we |
| // want to make sure that all samples remaining in the audio |
| // track's queue are played out. |
| mAudioPlayer->pause(true /* playPendingSamples */); |
| } else { |
| mAudioPlayer->pause(); |
| } |
| |
| mFlags &= ~AUDIO_RUNNING; |
| } |
| |
| mFlags &= ~PLAYING; |
| updateBatteryUsage_l(); |
| |
| return OK; |
| } |
| |
| bool PreviewPlayer::isPlaying() const { |
| return (mFlags & PLAYING) || (mFlags & CACHE_UNDERRUN); |
| } |
| |
| void PreviewPlayer::setSurface(const sp<Surface> &surface) { |
| Mutex::Autolock autoLock(mLock); |
| |
| mSurface = surface; |
| setNativeWindow_l(surface); |
| } |
| |
| void PreviewPlayer::setSurfaceTexture(const sp<ISurfaceTexture> &surfaceTexture) { |
| Mutex::Autolock autoLock(mLock); |
| |
| mSurface.clear(); |
| if (surfaceTexture != NULL) { |
| setNativeWindow_l(new SurfaceTextureClient(surfaceTexture)); |
| } |
| } |
| |
| void PreviewPlayer::shutdownVideoDecoder_l() { |
| if (mVideoBuffer) { |
| mVideoBuffer->release(); |
| mVideoBuffer = NULL; |
| } |
| |
| mVideoSource->stop(); |
| |
| // The following hack is necessary to ensure that the OMX |
| // component is completely released by the time we may try |
| // to instantiate it again. |
| wp<MediaSource> tmp = mVideoSource; |
| mVideoSource.clear(); |
| while (tmp.promote() != NULL) { |
| usleep(1000); |
| } |
| IPCThreadState::self()->flushCommands(); |
| } |
| |
| void PreviewPlayer::setNativeWindow_l(const sp<ANativeWindow> &native) { |
| mNativeWindow = native; |
| |
| if (mVideoSource == NULL) { |
| return; |
| } |
| |
| ALOGI("attempting to reconfigure to use new surface"); |
| |
| bool wasPlaying = (mFlags & PLAYING) != 0; |
| |
| pause_l(); |
| |
| shutdownVideoDecoder_l(); |
| |
| CHECK_EQ(initVideoDecoder_l(), (status_t)OK); |
| |
| if (mLastVideoTimeUs >= 0) { |
| mSeeking = SEEK; |
| mSeekNotificationSent = true; |
| mSeekTimeUs = mLastVideoTimeUs; |
| mFlags &= ~(AT_EOS | AUDIO_AT_EOS | VIDEO_AT_EOS); |
| } |
| |
| if (wasPlaying) { |
| play_l(); |
| } |
| } |
| |
| void PreviewPlayer::setAudioSink( |
| const sp<MediaPlayerBase::AudioSink> &audioSink) { |
| Mutex::Autolock autoLock(mLock); |
| |
| mAudioSink = audioSink; |
| } |
| |
| status_t PreviewPlayer::setLooping(bool shouldLoop) { |
| Mutex::Autolock autoLock(mLock); |
| |
| mFlags = mFlags & ~LOOPING; |
| |
| if (shouldLoop) { |
| mFlags |= LOOPING; |
| } |
| |
| return OK; |
| } |
| |
| void PreviewPlayer::setDuration_l(int64_t durationUs) { |
| if (mDurationUs < 0 || durationUs > mDurationUs) { |
| mDurationUs = durationUs; |
| } |
| } |
| |
| status_t PreviewPlayer::getDuration(int64_t *durationUs) { |
| Mutex::Autolock autoLock(mLock); |
| if (mDurationUs < 0) { |
| return UNKNOWN_ERROR; |
| } |
| |
| *durationUs = mDurationUs; |
| return OK; |
| } |
| |
| status_t PreviewPlayer::getPosition(int64_t *positionUs) { |
| Mutex::Autolock autoLock(mLock); |
| |
| if (mSeeking != NO_SEEK) { |
| *positionUs = mSeekTimeUs; |
| } else if (mVideoSource != NULL |
| && (mAudioPlayer == NULL || !(mFlags & VIDEO_AT_EOS))) { |
| *positionUs = mVideoTimeUs; |
| } else if (mAudioPlayer != NULL) { |
| *positionUs = mAudioPlayer->getMediaTimeUs(); |
| } else { |
| *positionUs = 0; |
| } |
| |
| return OK; |
| } |
| |
| void PreviewPlayer::setPosition_l(int64_t timeUs) { |
| mVideoTimeUs = timeUs; |
| } |
| |
| status_t PreviewPlayer::seekTo_l(int64_t timeUs) { |
| ALOGV("seekTo_l"); |
| if (mFlags & CACHE_UNDERRUN) { |
| mFlags &= ~CACHE_UNDERRUN; |
| play_l(); |
| } |
| |
| if ((mFlags & PLAYING) && mVideoSource != NULL && (mFlags & VIDEO_AT_EOS)) { |
| // Video playback completed before, there's no pending |
| // video event right now. In order for this new seek |
| // to be honored, we need to post one. |
| |
| postVideoEvent_l(); |
| } |
| |
| mSeeking = SEEK; |
| mSeekNotificationSent = false; |
| mSeekTimeUs = timeUs; |
| mFlags &= ~(AT_EOS | AUDIO_AT_EOS | VIDEO_AT_EOS); |
| |
| seekAudioIfNecessary_l(); |
| |
| if (!(mFlags & PLAYING)) { |
| ALOGV("seeking while paused, sending SEEK_COMPLETE notification" |
| " immediately."); |
| |
| notifyListener_l(MEDIA_SEEK_COMPLETE); |
| mSeekNotificationSent = true; |
| |
| if ((mFlags & PREPARED) && mVideoSource != NULL) { |
| mFlags |= SEEK_PREVIEW; |
| postVideoEvent_l(); |
| } |
| } |
| |
| return OK; |
| } |
| |
| void PreviewPlayer::seekAudioIfNecessary_l() { |
| if (mSeeking != NO_SEEK && mVideoSource == NULL && mAudioPlayer != NULL) { |
| mAudioPlayer->seekTo(mSeekTimeUs); |
| |
| mWatchForAudioSeekComplete = true; |
| mWatchForAudioEOS = true; |
| } |
| } |
| |
| void PreviewPlayer::setAudioSource(const sp<MediaSource>& source) { |
| CHECK(source != NULL); |
| mAudioTrack = source; |
| } |
| |
| void PreviewPlayer::setVideoSource(const sp<MediaSource>& source) { |
| CHECK(source != NULL); |
| mVideoTrack = source; |
| } |
| |
| void PreviewPlayer::finishSeekIfNecessary(int64_t videoTimeUs) { |
| if (mSeeking == SEEK_VIDEO_ONLY) { |
| mSeeking = NO_SEEK; |
| return; |
| } |
| |
| if (mSeeking == NO_SEEK || (mFlags & SEEK_PREVIEW)) { |
| return; |
| } |
| |
| if (mAudioPlayer != NULL) { |
| ALOGV("seeking audio to %lld us (%.2f secs).", videoTimeUs, videoTimeUs / 1E6); |
| |
| // If we don't have a video time, seek audio to the originally |
| // requested seek time instead. |
| |
| mAudioPlayer->seekTo(videoTimeUs < 0 ? mSeekTimeUs : videoTimeUs); |
| mWatchForAudioSeekComplete = true; |
| mWatchForAudioEOS = true; |
| } else if (!mSeekNotificationSent) { |
| // If we're playing video only, report seek complete now, |
| // otherwise audio player will notify us later. |
| notifyListener_l(MEDIA_SEEK_COMPLETE); |
| mSeekNotificationSent = true; |
| } |
| |
| mFlags |= FIRST_FRAME; |
| mSeeking = NO_SEEK; |
| } |
| |
| void PreviewPlayer::onCheckAudioStatus() { |
| Mutex::Autolock autoLock(mLock); |
| if (!mAudioStatusEventPending) { |
| // Event was dispatched and while we were blocking on the mutex, |
| // has already been cancelled. |
| return; |
| } |
| |
| mAudioStatusEventPending = false; |
| |
| if (mWatchForAudioSeekComplete && !mAudioPlayer->isSeeking()) { |
| mWatchForAudioSeekComplete = false; |
| |
| if (!mSeekNotificationSent) { |
| notifyListener_l(MEDIA_SEEK_COMPLETE); |
| mSeekNotificationSent = true; |
| } |
| |
| mSeeking = NO_SEEK; |
| } |
| |
| status_t finalStatus; |
| if (mWatchForAudioEOS && mAudioPlayer->reachedEOS(&finalStatus)) { |
| mWatchForAudioEOS = false; |
| mFlags |= AUDIO_AT_EOS; |
| mFlags |= FIRST_FRAME; |
| postStreamDoneEvent_l(finalStatus); |
| } |
| } |
| |
| void PreviewPlayer::postVideoEvent_l(int64_t delayUs) { |
| if (mVideoEventPending) { |
| return; |
| } |
| |
| mVideoEventPending = true; |
| mQueue.postEventWithDelay(mVideoEvent, delayUs < 0 ? 10000 : delayUs); |
| } |
| |
| void PreviewPlayer::postStreamDoneEvent_l(status_t status) { |
| if (mStreamDoneEventPending) { |
| return; |
| } |
| mStreamDoneEventPending = true; |
| |
| mStreamDoneStatus = status; |
| mQueue.postEvent(mStreamDoneEvent); |
| } |
| |
| void PreviewPlayer::postVideoLagEvent_l() { |
| if (mVideoLagEventPending) { |
| return; |
| } |
| mVideoLagEventPending = true; |
| mQueue.postEventWithDelay(mVideoLagEvent, 1000000ll); |
| } |
| |
| void PreviewPlayer::postCheckAudioStatusEvent_l(int64_t delayUs) { |
| if (mAudioStatusEventPending) { |
| return; |
| } |
| mAudioStatusEventPending = true; |
| mQueue.postEventWithDelay(mCheckAudioStatusEvent, delayUs); |
| } |
| |
| void PreviewPlayer::abortPrepare(status_t err) { |
| CHECK(err != OK); |
| |
| if (mIsAsyncPrepare) { |
| notifyListener_l(MEDIA_ERROR, MEDIA_ERROR_UNKNOWN, err); |
| } |
| |
| mPrepareResult = err; |
| mFlags &= ~(PREPARING|PREPARE_CANCELLED|PREPARING_CONNECTED); |
| mAsyncPrepareEvent = NULL; |
| mPreparedCondition.broadcast(); |
| } |
| |
| uint32_t PreviewPlayer::getSourceSeekFlags() const { |
| Mutex::Autolock lock(mLock); |
| return mExtractorFlags; |
| } |
| |
| void PreviewPlayer::postAudioEOS(int64_t delayUs) { |
| Mutex::Autolock autoLock(mLock); |
| postCheckAudioStatusEvent_l(delayUs); |
| } |
| |
| void PreviewPlayer::postAudioSeekComplete() { |
| Mutex::Autolock autoLock(mLock); |
| postCheckAudioStatusEvent_l(0 /* delayUs */); |
| } |
| |
| void PreviewPlayer::updateBatteryUsage_l() { |
| uint32_t params = IMediaPlayerService::kBatteryDataTrackDecoder; |
| if ((mAudioSource != NULL) && (mAudioSource != mAudioTrack)) { |
| params |= IMediaPlayerService::kBatteryDataTrackAudio; |
| } |
| if (mVideoSource != NULL) { |
| params |= IMediaPlayerService::kBatteryDataTrackVideo; |
| } |
| addBatteryData(params); |
| } |
| |
| } // namespace android |