blob: 044e51ad05797ea5686ff1523c2c719783b608b1 [file] [log] [blame]
/*
* Copyright (C) 2010 The Android Open Source Project
*
* Licensed under the Apache License, Version 2.0 (the "License");
* you may not use this file except in compliance with the License.
* You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
#include <assert.h>
#include <jni.h>
#include <pthread.h>
#include <string.h>
//#define LOG_NDEBUG 0
#define LOG_TAG "NativeMedia"
#include <utils/Log.h>
#include <OMXAL/OpenMAXAL.h>
#include <OMXAL/OpenMAXAL_Android.h>
#include <android/native_window_jni.h>
// engine interfaces
static XAObjectItf engineObject = NULL;
static XAEngineItf engineEngine = NULL;
// output mix interfaces
static XAObjectItf outputMixObject = NULL;
// streaming media player interfaces
static XAObjectItf playerObj = NULL;
static XAPlayItf playerPlayItf = NULL;
static XAAndroidBufferQueueItf playerBQItf = NULL;
static XAStreamInformationItf playerStreamInfoItf = NULL;
static XAVolumeItf playerVolItf = NULL;
// number of required interfaces for the MediaPlayer creation
#define NB_MAXAL_INTERFACES 3 // XAAndroidBufferQueueItf, XAStreamInformationItf and XAPlayItf
// video sink for the player
static ANativeWindow* theNativeWindow;
// number of buffers in our buffer queue, an arbitrary number
#define NB_BUFFERS 16
// we're streaming MPEG-2 transport stream data, operate on transport stream block size
#define MPEG2_TS_BLOCK_SIZE 188
// number of MPEG-2 transport stream blocks per buffer, an arbitrary number
#define BLOCKS_PER_BUFFER 20
// determines how much memory we're dedicating to memory caching
#define BUFFER_SIZE (BLOCKS_PER_BUFFER*MPEG2_TS_BLOCK_SIZE)
// where we cache in memory the data to play
// note this memory is re-used by the buffer queue callback
char dataCache[BUFFER_SIZE * NB_BUFFERS];
// handle of the file to play
FILE *file;
// has the app reached the end of the file
jboolean reachedEof = JNI_FALSE;
// constant to identify a buffer context which is the end of the stream to decode
static const int kEosBufferCntxt = 1980; // a magic value we can compare against
// for mutual exclusion between callback thread and application thread(s)
pthread_mutex_t mutex = PTHREAD_MUTEX_INITIALIZER;
pthread_cond_t cond = PTHREAD_COND_INITIALIZER;
// whether a discontinuity is in progress
jboolean discontinuity = JNI_FALSE;
static jboolean enqueueInitialBuffers(jboolean discontinuity);
// Callback for XAPlayItf through which we receive the XA_PLAYEVENT_HEADATEND event */
void PlayCallback(XAPlayItf caller, void *pContext, XAuint32 event) {
if (event & XA_PLAYEVENT_HEADATEND) {
ALOGV("XA_PLAYEVENT_HEADATEND received, all MP2TS data has been decoded\n");
}
}
// AndroidBufferQueueItf callback for an audio player
XAresult AndroidBufferQueueCallback(
XAAndroidBufferQueueItf caller,
void *pCallbackContext, /* input */
void *pBufferContext, /* input */
void *pBufferData, /* input */
XAuint32 dataSize, /* input */
XAuint32 dataUsed, /* input */
const XAAndroidBufferItem *pItems,/* input */
XAuint32 itemsLength /* input */)
{
XAresult res;
int ok;
// pCallbackContext was specified as NULL at RegisterCallback and is unused here
assert(NULL == pCallbackContext);
// note there is never any contention on this mutex unless a discontinuity request is active
ok = pthread_mutex_lock(&mutex);
assert(0 == ok);
// was a discontinuity requested?
if (discontinuity) {
// FIXME sorry, can't rewind after EOS
if (!reachedEof) {
// clear the buffer queue
res = (*playerBQItf)->Clear(playerBQItf);
assert(XA_RESULT_SUCCESS == res);
// rewind the data source so we are guaranteed to be at an appropriate point
rewind(file);
// Enqueue the initial buffers, with a discontinuity indicator on first buffer
(void) enqueueInitialBuffers(JNI_TRUE);
}
// acknowledge the discontinuity request
discontinuity = JNI_FALSE;
ok = pthread_cond_signal(&cond);
assert(0 == ok);
goto exit;
}
if ((pBufferData == NULL) && (pBufferContext != NULL)) {
const int processedCommand = *(int *)pBufferContext;
if (kEosBufferCntxt == processedCommand) {
ALOGV("EOS was processed\n");
// our buffer with the EOS message has been consumed
assert(0 == dataSize);
goto exit;
}
}
// pBufferData is a pointer to a buffer that we previously Enqueued
assert(BUFFER_SIZE == dataSize);
assert(dataCache <= (char *) pBufferData && (char *) pBufferData <
&dataCache[BUFFER_SIZE * NB_BUFFERS]);
assert(0 == (((char *) pBufferData - dataCache) % BUFFER_SIZE));
#if 0
// sample code to use the XAVolumeItf
XAAndroidBufferQueueState state;
(*caller)->GetState(caller, &state);
switch (state.index) {
case 300:
(*playerVolItf)->SetVolumeLevel(playerVolItf, -600); // -6dB
ALOGV("setting volume to -6dB");
break;
case 400:
(*playerVolItf)->SetVolumeLevel(playerVolItf, -1200); // -12dB
ALOGV("setting volume to -12dB");
break;
case 500:
(*playerVolItf)->SetVolumeLevel(playerVolItf, 0); // full volume
ALOGV("setting volume to 0dB (full volume)");
break;
case 600:
(*playerVolItf)->SetMute(playerVolItf, XA_BOOLEAN_TRUE); // mute
ALOGV("muting player");
break;
case 700:
(*playerVolItf)->SetMute(playerVolItf, XA_BOOLEAN_FALSE); // unmute
ALOGV("unmuting player");
break;
case 800:
(*playerVolItf)->SetStereoPosition(playerVolItf, -1000);
(*playerVolItf)->EnableStereoPosition(playerVolItf, XA_BOOLEAN_TRUE);
ALOGV("pan sound to the left (hard-left)");
break;
case 900:
(*playerVolItf)->EnableStereoPosition(playerVolItf, XA_BOOLEAN_FALSE);
ALOGV("disabling stereo position");
break;
default:
break;
}
#endif
// don't bother trying to read more data once we've hit EOF
if (reachedEof) {
goto exit;
}
size_t nbRead;
// note we do call fread from multiple threads, but never concurrently
nbRead = fread(pBufferData, BUFFER_SIZE, 1, file);
if (nbRead > 0) {
assert(1 == nbRead);
res = (*caller)->Enqueue(caller, NULL /*pBufferContext*/,
pBufferData /*pData*/,
nbRead * BUFFER_SIZE /*dataLength*/,
NULL /*pMsg*/,
0 /*msgLength*/);
assert(XA_RESULT_SUCCESS == res);
} else {
// signal EOS
XAAndroidBufferItem msgEos[1];
msgEos[0].itemKey = XA_ANDROID_ITEMKEY_EOS;
msgEos[0].itemSize = 0;
// EOS message has no parameters, so the total size of the message is the size of the key
// plus the size if itemSize, both XAuint32
res = (*caller)->Enqueue(caller, (void *)&kEosBufferCntxt /*pBufferContext*/,
NULL /*pData*/, 0 /*dataLength*/,
msgEos /*pMsg*/,
// FIXME == sizeof(BufferItem)? */
sizeof(XAuint32)*2 /*msgLength*/);
assert(XA_RESULT_SUCCESS == res);
reachedEof = JNI_TRUE;
}
exit:
ok = pthread_mutex_unlock(&mutex);
assert(0 == ok);
return XA_RESULT_SUCCESS;
}
void StreamChangeCallback (XAStreamInformationItf caller,
XAuint32 eventId,
XAuint32 streamIndex,
void * pEventData,
void * pContext )
{
ALOGV("StreamChangeCallback called for stream %u", streamIndex);
// pContext was specified as NULL at RegisterStreamChangeCallback and is unused here
assert(NULL == pContext);
switch (eventId) {
case XA_STREAMCBEVENT_PROPERTYCHANGE: {
/** From spec 1.0.1:
"This event indicates that stream property change has occurred.
The streamIndex parameter identifies the stream with the property change.
The pEventData parameter for this event is not used and shall be ignored."
*/
XAresult res;
XAuint32 domain;
res = (*caller)->QueryStreamType(caller, streamIndex, &domain);
assert(XA_RESULT_SUCCESS == res);
switch (domain) {
case XA_DOMAINTYPE_VIDEO: {
XAVideoStreamInformation videoInfo;
res = (*caller)->QueryStreamInformation(caller, streamIndex, &videoInfo);
assert(XA_RESULT_SUCCESS == res);
ALOGI("Found video size %u x %u, codec ID=%u, frameRate=%u, bitRate=%u, duration=%u ms",
videoInfo.width, videoInfo.height, videoInfo.codecId, videoInfo.frameRate,
videoInfo.bitRate, videoInfo.duration);
} break;
default:
fprintf(stderr, "Unexpected domain %u\n", domain);
break;
}
} break;
default:
fprintf(stderr, "Unexpected stream event ID %u\n", eventId);
break;
}
}
// create the engine and output mix objects
void Java_com_example_nativemedia_NativeMedia_createEngine(JNIEnv* env, jclass clazz)
{
XAresult res;
// create engine
res = xaCreateEngine(&engineObject, 0, NULL, 0, NULL, NULL);
assert(XA_RESULT_SUCCESS == res);
// realize the engine
res = (*engineObject)->Realize(engineObject, XA_BOOLEAN_FALSE);
assert(XA_RESULT_SUCCESS == res);
// get the engine interface, which is needed in order to create other objects
res = (*engineObject)->GetInterface(engineObject, XA_IID_ENGINE, &engineEngine);
assert(XA_RESULT_SUCCESS == res);
// create output mix
res = (*engineEngine)->CreateOutputMix(engineEngine, &outputMixObject, 0, NULL, NULL);
assert(XA_RESULT_SUCCESS == res);
// realize the output mix
res = (*outputMixObject)->Realize(outputMixObject, XA_BOOLEAN_FALSE);
assert(XA_RESULT_SUCCESS == res);
}
// Enqueue the initial buffers, and optionally signal a discontinuity in the first buffer
static jboolean enqueueInitialBuffers(jboolean discontinuity)
{
/* Fill our cache */
size_t nbRead;
nbRead = fread(dataCache, BUFFER_SIZE, NB_BUFFERS, file);
if (nbRead <= 0) {
// could be premature EOF or I/O error
ALOGE("Error filling cache, exiting\n");
return JNI_FALSE;
}
assert(1 <= nbRead && nbRead <= NB_BUFFERS);
ALOGV("Initially queueing %u buffers of %u bytes each", nbRead, BUFFER_SIZE);
/* Enqueue the content of our cache before starting to play,
we don't want to starve the player */
size_t i;
for (i = 0; i < nbRead; i++) {
XAresult res;
if (discontinuity) {
// signal discontinuity
XAAndroidBufferItem items[1];
items[0].itemKey = XA_ANDROID_ITEMKEY_DISCONTINUITY;
items[0].itemSize = 0;
// DISCONTINUITY message has no parameters,
// so the total size of the message is the size of the key
// plus the size if itemSize, both XAuint32
res = (*playerBQItf)->Enqueue(playerBQItf, NULL /*pBufferContext*/,
dataCache + i*BUFFER_SIZE, BUFFER_SIZE, items /*pMsg*/,
// FIXME == sizeof(BufferItem)? */
sizeof(XAuint32)*2 /*msgLength*/);
discontinuity = JNI_FALSE;
} else {
res = (*playerBQItf)->Enqueue(playerBQItf, NULL /*pBufferContext*/,
dataCache + i*BUFFER_SIZE, BUFFER_SIZE, NULL, 0);
}
assert(XA_RESULT_SUCCESS == res);
}
return JNI_TRUE;
}
// create streaming media player
jboolean Java_com_example_nativemedia_NativeMedia_createStreamingMediaPlayer(JNIEnv* env,
jclass clazz, jstring filename)
{
XAresult res;
// convert Java string to UTF-8
const char *utf8 = (*env)->GetStringUTFChars(env, filename, NULL);
assert(NULL != utf8);
// open the file to play
file = fopen(utf8, "rb");
if (file == NULL) {
ALOGE("Failed to open %s", utf8);
return JNI_FALSE;
}
// configure data source
XADataLocator_AndroidBufferQueue loc_abq = { XA_DATALOCATOR_ANDROIDBUFFERQUEUE, NB_BUFFERS };
XADataFormat_MIME format_mime = {
XA_DATAFORMAT_MIME, XA_ANDROID_MIME_MP2TS, XA_CONTAINERTYPE_MPEG_TS };
XADataSource dataSrc = {&loc_abq, &format_mime};
// configure audio sink
XADataLocator_OutputMix loc_outmix = { XA_DATALOCATOR_OUTPUTMIX, outputMixObject };
XADataSink audioSnk = { &loc_outmix, NULL };
// configure image video sink
XADataLocator_NativeDisplay loc_nd = {
XA_DATALOCATOR_NATIVEDISPLAY, // locatorType
// the video sink must be an ANativeWindow created from a Surface or SurfaceTextureClient
(void*)theNativeWindow, // hWindow
// must be NULL
NULL // hDisplay
};
XADataSink imageVideoSink = {&loc_nd, NULL};
// declare interfaces to use
XAboolean required[NB_MAXAL_INTERFACES]
= {XA_BOOLEAN_TRUE, XA_BOOLEAN_TRUE, XA_BOOLEAN_TRUE};
XAInterfaceID iidArray[NB_MAXAL_INTERFACES]
= {XA_IID_PLAY, XA_IID_ANDROIDBUFFERQUEUESOURCE,
XA_IID_STREAMINFORMATION};
// create media player
res = (*engineEngine)->CreateMediaPlayer(engineEngine, &playerObj, &dataSrc,
NULL, &audioSnk, &imageVideoSink, NULL, NULL,
NB_MAXAL_INTERFACES /*XAuint32 numInterfaces*/,
iidArray /*const XAInterfaceID *pInterfaceIds*/,
required /*const XAboolean *pInterfaceRequired*/);
assert(XA_RESULT_SUCCESS == res);
// release the Java string and UTF-8
(*env)->ReleaseStringUTFChars(env, filename, utf8);
// realize the player
res = (*playerObj)->Realize(playerObj, XA_BOOLEAN_FALSE);
assert(XA_RESULT_SUCCESS == res);
// get the play interface
res = (*playerObj)->GetInterface(playerObj, XA_IID_PLAY, &playerPlayItf);
assert(XA_RESULT_SUCCESS == res);
// get the stream information interface (for video size)
res = (*playerObj)->GetInterface(playerObj, XA_IID_STREAMINFORMATION, &playerStreamInfoItf);
assert(XA_RESULT_SUCCESS == res);
// get the volume interface
res = (*playerObj)->GetInterface(playerObj, XA_IID_VOLUME, &playerVolItf);
assert(XA_RESULT_SUCCESS == res);
// get the Android buffer queue interface
res = (*playerObj)->GetInterface(playerObj, XA_IID_ANDROIDBUFFERQUEUESOURCE, &playerBQItf);
assert(XA_RESULT_SUCCESS == res);
// specify which events we want to be notified of
res = (*playerBQItf)->SetCallbackEventsMask(playerBQItf, XA_ANDROIDBUFFERQUEUEEVENT_PROCESSED);
// use the play interface to set up a callback for the XA_PLAYEVENT_HEADATEND event */
res = (*playerPlayItf)->SetCallbackEventsMask(playerPlayItf, XA_PLAYEVENT_HEADATEND);
assert(XA_RESULT_SUCCESS == res);
res = (*playerPlayItf)->RegisterCallback(playerPlayItf,
PlayCallback /*callback*/, NULL /*pContext*/);
assert(XA_RESULT_SUCCESS == res);
// register the callback from which OpenMAX AL can retrieve the data to play
res = (*playerBQItf)->RegisterCallback(playerBQItf, AndroidBufferQueueCallback, NULL);
assert(XA_RESULT_SUCCESS == res);
// we want to be notified of the video size once it's found, so we register a callback for that
res = (*playerStreamInfoItf)->RegisterStreamChangeCallback(playerStreamInfoItf,
StreamChangeCallback, NULL);
// enqueue the initial buffers
if (!enqueueInitialBuffers(JNI_FALSE)) {
return JNI_FALSE;
}
// prepare the player
res = (*playerPlayItf)->SetPlayState(playerPlayItf, XA_PLAYSTATE_PAUSED);
assert(XA_RESULT_SUCCESS == res);
// set the volume
res = (*playerVolItf)->SetVolumeLevel(playerVolItf, 0);//-300);
assert(XA_RESULT_SUCCESS == res);
// start the playback
res = (*playerPlayItf)->SetPlayState(playerPlayItf, XA_PLAYSTATE_PLAYING);
assert(XA_RESULT_SUCCESS == res);
return JNI_TRUE;
}
// set the playing state for the streaming media player
void Java_com_example_nativemedia_NativeMedia_setPlayingStreamingMediaPlayer(JNIEnv* env,
jclass clazz, jboolean isPlaying)
{
XAresult res;
// make sure the streaming media player was created
if (NULL != playerPlayItf) {
// set the player's state
res = (*playerPlayItf)->SetPlayState(playerPlayItf, isPlaying ?
XA_PLAYSTATE_PLAYING : XA_PLAYSTATE_PAUSED);
assert(XA_RESULT_SUCCESS == res);
}
}
// shut down the native media system
void Java_com_example_nativemedia_NativeMedia_shutdown(JNIEnv* env, jclass clazz)
{
// destroy streaming media player object, and invalidate all associated interfaces
if (playerObj != NULL) {
(*playerObj)->Destroy(playerObj);
playerObj = NULL;
playerPlayItf = NULL;
playerBQItf = NULL;
playerStreamInfoItf = NULL;
playerVolItf = NULL;
}
// destroy output mix object, and invalidate all associated interfaces
if (outputMixObject != NULL) {
(*outputMixObject)->Destroy(outputMixObject);
outputMixObject = NULL;
}
// destroy engine object, and invalidate all associated interfaces
if (engineObject != NULL) {
(*engineObject)->Destroy(engineObject);
engineObject = NULL;
engineEngine = NULL;
}
// close the file
if (file != NULL) {
fclose(file);
file = NULL;
}
// make sure we don't leak native windows
if (theNativeWindow != NULL) {
ANativeWindow_release(theNativeWindow);
theNativeWindow = NULL;
}
}
// set the surface
void Java_com_example_nativemedia_NativeMedia_setSurface(JNIEnv *env, jclass clazz, jobject surface)
{
// obtain a native window from a Java surface
theNativeWindow = ANativeWindow_fromSurface(env, surface);
}
// rewind the streaming media player
void Java_com_example_nativemedia_NativeMedia_rewindStreamingMediaPlayer(JNIEnv *env, jclass clazz)
{
XAresult res;
// make sure the streaming media player was created
if (NULL != playerBQItf && NULL != file) {
// first wait for buffers currently in queue to be drained
int ok;
ok = pthread_mutex_lock(&mutex);
assert(0 == ok);
discontinuity = JNI_TRUE;
// wait for discontinuity request to be observed by buffer queue callback
// FIXME sorry, can't rewind after EOS
while (discontinuity && !reachedEof) {
ok = pthread_cond_wait(&cond, &mutex);
assert(0 == ok);
}
ok = pthread_mutex_unlock(&mutex);
assert(0 == ok);
}
}