| /* |
| * Copyright (C) 2011 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| /* |
| * |
| */ |
| #include <string.h> |
| #include <jni.h> |
| #include <stdio.h> |
| #include <stdlib.h> |
| #include <time.h> |
| #include <db_utilities_camera.h> |
| |
| #include "mosaic/AlignFeatures.h" |
| #include "mosaic/Blend.h" |
| #include "mosaic/Mosaic.h" |
| #include "mosaic/Log.h" |
| #define LOG_TAG "FEATURE_MOS_JNI" |
| |
| #ifdef __cplusplus |
| extern "C" { |
| #endif |
| |
| #include "mosaic_renderer_jni.h" |
| |
| char buffer[1024]; |
| |
| const int MAX_FRAMES = 100; |
| |
| static double mTx; |
| |
| int tWidth[NR]; |
| int tHeight[NR]; |
| |
| ImageType tImage[NR][MAX_FRAMES];// = {{ImageUtils::IMAGE_TYPE_NOIMAGE}}; // YVU24 format image |
| Mosaic *mosaic[NR] = {NULL,NULL}; |
| ImageType resultYVU = ImageUtils::IMAGE_TYPE_NOIMAGE; |
| ImageType resultBGR = ImageUtils::IMAGE_TYPE_NOIMAGE; |
| float gTRS[11]; // 9 elements of the transformation, 1 for frame-number, 1 for alignment error code. |
| // Variables to keep track of the mosaic computation progress for both LR & HR. |
| float gProgress[NR]; |
| // Variables to be able to cancel the mosaic computation when the GUI says so. |
| bool gCancelComputation[NR]; |
| |
| int c; |
| int width=0, height=0; |
| int mosaicWidth=0, mosaicHeight=0; |
| |
| //int blendingType = Blend::BLEND_TYPE_FULL; |
| //int blendingType = Blend::BLEND_TYPE_CYLPAN; |
| int blendingType = Blend::BLEND_TYPE_HORZ; |
| int stripType = Blend::STRIP_TYPE_THIN; |
| bool high_res = false; |
| bool quarter_res[NR] = {false,false}; |
| float thresh_still[NR] = {5.0f,0.0f}; |
| |
| /* return current time in milliseconds*/ |
| |
| #ifndef now_ms |
| static double |
| now_ms(void) |
| { |
| //struct timespec res; |
| struct timeval res; |
| //clock_gettime(CLOCK_REALTIME, &res); |
| gettimeofday(&res, NULL); |
| return 1000.0*res.tv_sec + (double)res.tv_usec/1e3; |
| } |
| #endif |
| |
| |
| static int frame_number_HR = 0; |
| static int frame_number_LR = 0; |
| |
| int Init(int mID, int nmax) |
| { |
| double t0, t1, time_c; |
| |
| if(mosaic[mID]!=NULL) |
| { |
| delete mosaic[mID]; |
| mosaic[mID] = NULL; |
| } |
| |
| mosaic[mID] = new Mosaic(); |
| |
| t0 = now_ms(); |
| |
| // When processing higher than 720x480 video, process low-res at |
| // quarter resolution |
| if(tWidth[LR]>180) |
| quarter_res[LR] = true; |
| |
| |
| // Check for initialization and if not, initialize |
| if (!mosaic[mID]->isInitialized()) |
| { |
| mosaic[mID]->initialize(blendingType, stripType, tWidth[mID], tHeight[mID], |
| nmax, quarter_res[mID], thresh_still[mID]); |
| } |
| |
| t1 = now_ms(); |
| time_c = t1 - t0; |
| LOGV("Init[%d]: %g ms [%d frames]",mID,time_c,nmax); |
| return 1; |
| } |
| |
| void GenerateQuarterResImagePlanar(ImageType im, int input_w, int input_h, |
| ImageType &out) |
| { |
| ImageType imp; |
| ImageType outp; |
| |
| int count = 0; |
| |
| for (int j = 0; j < input_h; j += H2L_FACTOR) |
| { |
| imp = im + j * input_w; |
| outp = out + (j / H2L_FACTOR) * (input_w / H2L_FACTOR); |
| |
| for (int i = 0; i < input_w; i += H2L_FACTOR) |
| { |
| *outp++ = *(imp + i); |
| count++; |
| } |
| } |
| |
| for (int j = input_h; j < 2 * input_h; j += H2L_FACTOR) |
| { |
| imp = im + j * input_w; |
| outp = out + (j / H2L_FACTOR) * (input_w / H2L_FACTOR); |
| |
| for (int i = 0; i < input_w; i += H2L_FACTOR) |
| { |
| *outp++ = *(imp + i); |
| count++; |
| } |
| } |
| |
| for (int j = 2 * input_h; j < 3 * input_h; j += H2L_FACTOR) |
| { |
| imp = im + j * input_w; |
| outp = out + (j / H2L_FACTOR) * (input_w / H2L_FACTOR); |
| |
| for (int i = 0; i < input_w; i += H2L_FACTOR) |
| { |
| *outp++ = *(imp + i); |
| count++; |
| } |
| } |
| } |
| |
| int AddFrame(int mID, int k, float* trs1d) |
| { |
| double t0, t1, time_c; |
| double trs[3][3]; |
| |
| int ret_code = mosaic[mID]->addFrame(tImage[mID][k]); |
| |
| mosaic[mID]->getAligner()->getLastTRS(trs); |
| |
| if(trs1d!=NULL) |
| { |
| |
| trs1d[0] = trs[0][0]; |
| trs1d[1] = trs[0][1]; |
| trs1d[2] = trs[0][2]; |
| trs1d[3] = trs[1][0]; |
| trs1d[4] = trs[1][1]; |
| trs1d[5] = trs[1][2]; |
| trs1d[6] = trs[2][0]; |
| trs1d[7] = trs[2][1]; |
| trs1d[8] = trs[2][2]; |
| } |
| |
| return ret_code; |
| } |
| |
| int Finalize(int mID) |
| { |
| double t0, t1, time_c; |
| |
| t0 = now_ms(); |
| // Create the mosaic |
| int ret = mosaic[mID]->createMosaic(gProgress[mID], gCancelComputation[mID]); |
| t1 = now_ms(); |
| time_c = t1 - t0; |
| LOGV("CreateMosaic: %g ms",time_c); |
| |
| // Get back the result |
| resultYVU = mosaic[mID]->getMosaic(mosaicWidth, mosaicHeight); |
| |
| return ret; |
| } |
| |
| void YUV420toYVU24(ImageType yvu24, ImageType yuv420sp, int width, int height) |
| { |
| int frameSize = width * height; |
| |
| ImageType oyp = yvu24; |
| ImageType ovp = yvu24+frameSize; |
| ImageType oup = yvu24+frameSize+frameSize; |
| |
| for (int j = 0, yp = 0; j < height; j++) |
| { |
| unsigned char u = 0, v = 0; |
| int uvp = frameSize + (j >> 1) * width; |
| for (int i = 0; i < width; i++, yp++) |
| { |
| *oyp++ = yuv420sp[yp]; |
| //int y = (0xff & (int)yuv420sp[yp]) -16; |
| //yvu24p[yp] = (y<0)?0:y; |
| |
| if ((i & 1) == 0) |
| { |
| v = yuv420sp[uvp++]; |
| u = yuv420sp[uvp++]; |
| } |
| |
| *ovp++ = v; |
| *oup++ = u; |
| } |
| } |
| } |
| |
| void YUV420toYVU24_NEW(ImageType yvu24, ImageType yuv420sp, int width, |
| int height) |
| { |
| int frameSize = width * height; |
| |
| ImageType oyp = yvu24; |
| ImageType ovp = yvu24 + frameSize; |
| ImageType oup = yvu24 + frameSize + frameSize; |
| |
| memcpy(yvu24, yuv420sp, frameSize * sizeof(unsigned char)); |
| |
| for (int j = 0; j < height; j += 2) |
| { |
| unsigned char u = 0, v = 0; |
| int uvp = frameSize + (j >> 1) * width; |
| ovp = yvu24 + frameSize + j * width; |
| oup = ovp + frameSize; |
| |
| ImageType iuvp = yuv420sp + uvp; |
| |
| for (int i = 0; i < width; i += 2) |
| { |
| v = *iuvp++; |
| u = *iuvp++; |
| |
| *ovp++ = v; |
| *oup++ = u; |
| |
| *ovp++ = v; |
| *oup++ = u; |
| |
| } |
| memcpy(ovp, ovp - width, width * sizeof(unsigned char)); |
| memcpy(oup, oup - width, width * sizeof(unsigned char)); |
| } |
| } |
| |
| |
| JNIEXPORT void JNICALL Java_com_android_camera_Mosaic_allocateMosaicMemory( |
| JNIEnv* env, jobject thiz, jint width, jint height) |
| { |
| tWidth[HR] = width; |
| tHeight[HR] = height; |
| tWidth[LR] = int(width / H2L_FACTOR); |
| tHeight[LR] = int(height / H2L_FACTOR); |
| |
| for(int i=0; i<MAX_FRAMES; i++) |
| { |
| tImage[LR][i] = ImageUtils::allocateImage(tWidth[LR], tHeight[LR], |
| ImageUtils::IMAGE_TYPE_NUM_CHANNELS); |
| tImage[HR][i] = ImageUtils::allocateImage(tWidth[HR], tHeight[HR], |
| ImageUtils::IMAGE_TYPE_NUM_CHANNELS); |
| } |
| |
| AllocateTextureMemory(tWidth[HR], tHeight[HR], tWidth[LR], tHeight[LR]); |
| } |
| |
| JNIEXPORT void JNICALL Java_com_android_camera_Mosaic_freeMosaicMemory( |
| JNIEnv* env, jobject thiz) |
| { |
| for(int i = 0; i < MAX_FRAMES; i++) |
| { |
| ImageUtils::freeImage(tImage[LR][i]); |
| ImageUtils::freeImage(tImage[HR][i]); |
| } |
| |
| FreeTextureMemory(); |
| } |
| |
| |
| void decodeYUV444SP(unsigned char* rgb, unsigned char* yuv420sp, int width, |
| int height) |
| { |
| int frameSize = width * height; |
| |
| for (int j = 0, yp = 0; j < height; j++) |
| { |
| int vp = frameSize + j * width, u = 0, v = 0; |
| int up = vp + frameSize; |
| |
| for (int i = 0; i < width; i++, yp++, vp++, up++) |
| { |
| int y = (0xff & ((int) yuv420sp[yp])) - 16; |
| if (y < 0) y = 0; |
| |
| v = (0xff & yuv420sp[vp]) - 128; |
| u = (0xff & yuv420sp[up]) - 128; |
| |
| int y1192 = 1192 * y; |
| int r = (y1192 + 1634 * v); |
| int g = (y1192 - 833 * v - 400 * u); |
| int b = (y1192 + 2066 * u); |
| |
| if (r < 0) r = 0; else if (r > 262143) r = 262143; |
| if (g < 0) g = 0; else if (g > 262143) g = 262143; |
| if (b < 0) b = 0; else if (b > 262143) b = 262143; |
| |
| //rgb[yp] = 0xff000000 | ((r << 6) & 0xff0000) | ((g >> 2) & 0xff00) | ((b >> 10) & 0xff); |
| int p = j*width*3+i*3; |
| rgb[p+0] = (r<<6 & 0xFF0000)>>16; |
| rgb[p+1] = (g>>2 & 0xFF00)>>8; |
| rgb[p+2] = b>>10 & 0xFF; |
| } |
| } |
| } |
| |
| static int count = 0; |
| |
| void ConvertYVUAiToPlanarYVU(unsigned char *planar, unsigned char *in, int width, |
| int height) |
| { |
| int planeSize = width * height; |
| unsigned char* Yptr = planar; |
| unsigned char* Vptr = planar + planeSize; |
| unsigned char* Uptr = Vptr + planeSize; |
| |
| for (int i = 0; i < planeSize; i++) |
| { |
| *Yptr++ = *in++; |
| *Vptr++ = *in++; |
| *Uptr++ = *in++; |
| in++; // Alpha |
| } |
| } |
| |
| JNIEXPORT jfloatArray JNICALL Java_com_android_camera_Mosaic_setSourceImageFromGPU( |
| JNIEnv* env, jobject thiz) |
| { |
| double t0, t1, time_c; |
| t0 = now_ms(); |
| int ret_code = Mosaic::MOSAIC_RET_OK; |
| |
| if(frame_number_HR<MAX_FRAMES && frame_number_LR<MAX_FRAMES) |
| { |
| double last_tx = mTx; |
| |
| sem_wait(&gPreviewImage_semaphore); |
| ConvertYVUAiToPlanarYVU(tImage[LR][frame_number_LR], gPreviewImage[LR], |
| tWidth[LR], tHeight[LR]); |
| |
| sem_post(&gPreviewImage_semaphore); |
| |
| ret_code = AddFrame(LR, frame_number_LR, gTRS); |
| |
| if(ret_code == Mosaic::MOSAIC_RET_OK || ret_code == Mosaic::MOSAIC_RET_FEW_INLIERS) |
| { |
| // Copy into HR buffer only if this is a valid frame |
| sem_wait(&gPreviewImage_semaphore); |
| ConvertYVUAiToPlanarYVU(tImage[HR][frame_number_HR], gPreviewImage[HR], |
| tWidth[HR], tHeight[HR]); |
| sem_post(&gPreviewImage_semaphore); |
| |
| frame_number_LR++; |
| frame_number_HR++; |
| } |
| } |
| else |
| { |
| gTRS[1] = gTRS[2] = gTRS[3] = gTRS[5] = gTRS[6] = gTRS[7] = 0.0f; |
| gTRS[0] = gTRS[4] = gTRS[8] = 1.0f; |
| } |
| |
| UpdateWarpTransformation(gTRS); |
| |
| gTRS[9] = frame_number_HR; |
| gTRS[10] = ret_code; |
| |
| jfloatArray bytes = env->NewFloatArray(11); |
| if(bytes != 0) |
| { |
| env->SetFloatArrayRegion(bytes, 0, 11, (jfloat*) gTRS); |
| } |
| return bytes; |
| } |
| |
| |
| |
| JNIEXPORT jfloatArray JNICALL Java_com_android_camera_Mosaic_setSourceImage( |
| JNIEnv* env, jobject thiz, jbyteArray photo_data) |
| { |
| double t0, t1, time_c; |
| t0 = now_ms(); |
| |
| int ret_code = Mosaic::MOSAIC_RET_OK; |
| |
| if(frame_number_HR<MAX_FRAMES && frame_number_LR<MAX_FRAMES) |
| { |
| jbyte *pixels = env->GetByteArrayElements(photo_data, 0); |
| |
| YUV420toYVU24_NEW(tImage[HR][frame_number_HR], (ImageType)pixels, |
| tWidth[HR], tHeight[HR]); |
| |
| env->ReleaseByteArrayElements(photo_data, pixels, 0); |
| |
| double last_tx = mTx; |
| |
| t0 = now_ms(); |
| GenerateQuarterResImagePlanar(tImage[HR][frame_number_HR], tWidth[HR], |
| tHeight[HR], tImage[LR][frame_number_LR]); |
| |
| |
| sem_wait(&gPreviewImage_semaphore); |
| decodeYUV444SP(gPreviewImage[LR], tImage[LR][frame_number_LR], |
| gPreviewImageWidth[LR], gPreviewImageHeight[LR]); |
| sem_post(&gPreviewImage_semaphore); |
| |
| ret_code = AddFrame(LR, frame_number_LR, gTRS); |
| |
| if(ret_code == Mosaic::MOSAIC_RET_OK || ret_code == Mosaic::MOSAIC_RET_FEW_INLIERS) |
| { |
| frame_number_LR++; |
| frame_number_HR++; |
| } |
| |
| } |
| else |
| { |
| gTRS[1] = gTRS[2] = gTRS[3] = gTRS[5] = gTRS[6] = gTRS[7] = 0.0f; |
| gTRS[0] = gTRS[4] = gTRS[8] = 1.0f; |
| } |
| |
| UpdateWarpTransformation(gTRS); |
| |
| gTRS[9] = frame_number_HR; |
| gTRS[10] = ret_code; |
| |
| jfloatArray bytes = env->NewFloatArray(11); |
| if(bytes != 0) |
| { |
| env->SetFloatArrayRegion(bytes, 0, 11, (jfloat*) gTRS); |
| } |
| return bytes; |
| } |
| |
| JNIEXPORT void JNICALL Java_com_android_camera_Mosaic_setBlendingType( |
| JNIEnv* env, jobject thiz, jint type) |
| { |
| blendingType = int(type); |
| } |
| |
| JNIEXPORT void JNICALL Java_com_android_camera_Mosaic_setStripType( |
| JNIEnv* env, jobject thiz, jint type) |
| { |
| stripType = int(type); |
| } |
| |
| JNIEXPORT void JNICALL Java_com_android_camera_Mosaic_reset( |
| JNIEnv* env, jobject thiz) |
| { |
| frame_number_HR = 0; |
| frame_number_LR = 0; |
| |
| gProgress[LR] = 0.0; |
| gProgress[HR] = 0.0; |
| |
| gCancelComputation[LR] = false; |
| gCancelComputation[HR] = false; |
| |
| Init(LR,MAX_FRAMES); |
| } |
| |
| JNIEXPORT jint JNICALL Java_com_android_camera_Mosaic_reportProgress( |
| JNIEnv* env, jobject thiz, jboolean hires, jboolean cancel_computation) |
| { |
| if(bool(hires)) |
| gCancelComputation[HR] = cancel_computation; |
| else |
| gCancelComputation[LR] = cancel_computation; |
| |
| if(bool(hires)) |
| return (jint) gProgress[HR]; |
| else |
| return (jint) gProgress[LR]; |
| } |
| |
| JNIEXPORT jint JNICALL Java_com_android_camera_Mosaic_createMosaic( |
| JNIEnv* env, jobject thiz, jboolean value) |
| { |
| high_res = bool(value); |
| |
| int ret; |
| |
| if(high_res) |
| { |
| LOGV("createMosaic() - High-Res Mode"); |
| double t0, t1, time_c; |
| |
| gProgress[HR] = 0.0; |
| t0 = now_ms(); |
| |
| Init(HR, frame_number_HR); |
| |
| for(int k = 0; k < frame_number_HR; k++) |
| { |
| if (gCancelComputation[HR]) |
| break; |
| AddFrame(HR, k, NULL); |
| gProgress[HR] += TIME_PERCENT_ALIGN/frame_number_HR; |
| } |
| |
| if (gCancelComputation[HR]) |
| { |
| ret = Mosaic::MOSAIC_RET_CANCELLED; |
| } |
| else |
| { |
| gProgress[HR] = TIME_PERCENT_ALIGN; |
| |
| t1 = now_ms(); |
| time_c = t1 - t0; |
| LOGV("AlignAll - %d frames [HR]: %g ms", frame_number_HR, time_c); |
| |
| ret = Finalize(HR); |
| |
| gProgress[HR] = 100.0; |
| } |
| |
| high_res = false; |
| } |
| else |
| { |
| LOGV("createMosaic() - Low-Res Mode"); |
| gProgress[LR] = TIME_PERCENT_ALIGN; |
| |
| ret = Finalize(LR); |
| |
| gProgress[LR] = 100.0; |
| } |
| |
| return (jint) ret; |
| } |
| |
| JNIEXPORT jintArray JNICALL Java_com_android_camera_Mosaic_getFinalMosaic( |
| JNIEnv* env, jobject thiz) |
| { |
| int y,x; |
| int width = mosaicWidth; |
| int height = mosaicHeight; |
| int imageSize = width * height; |
| |
| // Convert back to RGB24 |
| resultBGR = ImageUtils::allocateImage(mosaicWidth, mosaicHeight, |
| ImageUtils::IMAGE_TYPE_NUM_CHANNELS); |
| ImageUtils::yvu2bgr(resultBGR, resultYVU, mosaicWidth, mosaicHeight); |
| |
| LOGV("MosBytes: %d, W = %d, H = %d", imageSize, width, height); |
| |
| int* image = new int[imageSize]; |
| int* dims = new int[2]; |
| |
| for(y=0; y<height; y++) |
| { |
| for(x=0; x<width; x++) |
| { |
| image[y*width+x] = (0xFF<<24) | (resultBGR[y*width*3+x*3+2]<<16)| |
| (resultBGR[y*width*3+x*3+1]<<8)| (resultBGR[y*width*3+x*3]); |
| } |
| } |
| |
| dims[0] = width; |
| dims[1] = height; |
| |
| ImageUtils::freeImage(resultBGR); |
| |
| jintArray bytes = env->NewIntArray(imageSize+2); |
| if (bytes == 0) { |
| LOGE("Error in creating the image."); |
| delete[] image; |
| return 0; |
| } |
| env->SetIntArrayRegion(bytes, 0, imageSize, (jint*) image); |
| env->SetIntArrayRegion(bytes, imageSize, 2, (jint*) dims); |
| delete[] image; |
| delete[] dims; |
| return bytes; |
| } |
| |
| JNIEXPORT jbyteArray JNICALL Java_com_android_camera_Mosaic_getFinalMosaicNV21( |
| JNIEnv* env, jobject thiz) |
| { |
| int y,x; |
| int width; |
| int height; |
| |
| width = mosaicWidth; |
| height = mosaicHeight; |
| |
| int imageSize = 1.5*width * height; |
| |
| // Convert YVU to NV21 format in-place |
| ImageType V = resultYVU+mosaicWidth*mosaicHeight; |
| ImageType U = V+mosaicWidth*mosaicHeight; |
| for(int j=0; j<mosaicHeight/2; j++) |
| { |
| for(int i=0; i<mosaicWidth; i+=2) |
| { |
| V[j*mosaicWidth+i] = V[(2*j)*mosaicWidth+i]; // V |
| V[j*mosaicWidth+i+1] = U[(2*j)*mosaicWidth+i]; // U |
| } |
| } |
| |
| LOGV("MosBytes: %d, W = %d, H = %d", imageSize, width, height); |
| |
| unsigned char* dims = new unsigned char[8]; |
| |
| dims[0] = (unsigned char)(width >> 24); |
| dims[1] = (unsigned char)(width >> 16); |
| dims[2] = (unsigned char)(width >> 8); |
| dims[3] = (unsigned char)width; |
| |
| dims[4] = (unsigned char)(height >> 24); |
| dims[5] = (unsigned char)(height >> 16); |
| dims[6] = (unsigned char)(height >> 8); |
| dims[7] = (unsigned char)height; |
| |
| jbyteArray bytes = env->NewByteArray(imageSize+8); |
| if (bytes == 0) { |
| LOGE("Error in creating the image."); |
| ImageUtils::freeImage(resultYVU); |
| return 0; |
| } |
| env->SetByteArrayRegion(bytes, 0, imageSize, (jbyte*) resultYVU); |
| env->SetByteArrayRegion(bytes, imageSize, 8, (jbyte*) dims); |
| delete[] dims; |
| ImageUtils::freeImage(resultYVU); |
| return bytes; |
| } |
| |
| #ifdef __cplusplus |
| } |
| #endif |