| /* |
| * Copyright (C) 2012 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| #define LOG_TAG "Camera2" |
| #define ATRACE_TAG ATRACE_TAG_CAMERA |
| //#define LOG_NDEBUG 0 |
| |
| #include <utils/Log.h> |
| #include <utils/Trace.h> |
| |
| #include <cutils/properties.h> |
| #include <gui/Surface.h> |
| #include "camera2/Parameters.h" |
| #include "Camera2Client.h" |
| #include "Camera2Device.h" |
| #include "Camera3Device.h" |
| |
| #include "camera2/ZslProcessor.h" |
| #include "camera2/ZslProcessor3.h" |
| |
| #define ALOG1(...) ALOGD_IF(gLogLevel >= 1, __VA_ARGS__); |
| #define ALOG2(...) ALOGD_IF(gLogLevel >= 2, __VA_ARGS__); |
| |
| namespace android { |
| using namespace camera2; |
| |
| static int getCallingPid() { |
| return IPCThreadState::self()->getCallingPid(); |
| } |
| |
| // Interface used by CameraService |
| |
| Camera2Client::Camera2Client(const sp<CameraService>& cameraService, |
| const sp<ICameraClient>& cameraClient, |
| const String16& clientPackageName, |
| int cameraId, |
| int cameraFacing, |
| int clientPid, |
| uid_t clientUid, |
| int servicePid, |
| int deviceVersion): |
| Camera2ClientBase(cameraService, cameraClient, clientPackageName, |
| cameraId, cameraFacing, clientPid, clientUid, servicePid), |
| mParameters(cameraId, cameraFacing), |
| mDeviceVersion(deviceVersion) |
| { |
| ATRACE_CALL(); |
| ALOGI("Camera %d: Opened", cameraId); |
| |
| switch (mDeviceVersion) { |
| case CAMERA_DEVICE_API_VERSION_2_0: |
| mDevice = new Camera2Device(cameraId); |
| break; |
| case CAMERA_DEVICE_API_VERSION_3_0: |
| mDevice = new Camera3Device(cameraId); |
| break; |
| default: |
| ALOGE("Camera %d: Unknown HAL device version %d", |
| cameraId, mDeviceVersion); |
| mDevice = NULL; |
| break; |
| } |
| |
| |
| SharedParameters::Lock l(mParameters); |
| l.mParameters.state = Parameters::DISCONNECTED; |
| } |
| |
| status_t Camera2Client::initialize(camera_module_t *module) |
| { |
| ATRACE_CALL(); |
| ALOGV("%s: Initializing client for camera %d", __FUNCTION__, mCameraId); |
| status_t res; |
| |
| res = Camera2ClientBase::initialize(module); |
| if (res != OK) { |
| return res; |
| } |
| |
| SharedParameters::Lock l(mParameters); |
| |
| res = l.mParameters.initialize(&(mDevice->info())); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: unable to build defaults: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return NO_INIT; |
| } |
| |
| String8 threadName; |
| |
| mStreamingProcessor = new StreamingProcessor(this); |
| |
| mFrameProcessor = new FrameProcessor(mDevice, this); |
| threadName = String8::format("C2-%d-FrameProc", |
| mCameraId); |
| mFrameProcessor->run(threadName.string()); |
| |
| mCaptureSequencer = new CaptureSequencer(this); |
| threadName = String8::format("C2-%d-CaptureSeq", |
| mCameraId); |
| mCaptureSequencer->run(threadName.string()); |
| |
| mJpegProcessor = new JpegProcessor(this, mCaptureSequencer); |
| threadName = String8::format("C2-%d-JpegProc", |
| mCameraId); |
| mJpegProcessor->run(threadName.string()); |
| |
| switch (mDeviceVersion) { |
| case CAMERA_DEVICE_API_VERSION_2_0: { |
| sp<ZslProcessor> zslProc = |
| new ZslProcessor(this, mCaptureSequencer); |
| mZslProcessor = zslProc; |
| mZslProcessorThread = zslProc; |
| break; |
| } |
| case CAMERA_DEVICE_API_VERSION_3_0:{ |
| sp<ZslProcessor3> zslProc = |
| new ZslProcessor3(this, mCaptureSequencer); |
| mZslProcessor = zslProc; |
| mZslProcessorThread = zslProc; |
| break; |
| } |
| default: |
| break; |
| } |
| threadName = String8::format("C2-%d-ZslProc", |
| mCameraId); |
| mZslProcessorThread->run(threadName.string()); |
| |
| mCallbackProcessor = new CallbackProcessor(this); |
| threadName = String8::format("C2-%d-CallbkProc", |
| mCameraId); |
| mCallbackProcessor->run(threadName.string()); |
| |
| if (gLogLevel >= 1) { |
| ALOGD("%s: Default parameters converted from camera %d:", __FUNCTION__, |
| mCameraId); |
| ALOGD("%s", l.mParameters.paramsFlattened.string()); |
| } |
| |
| return OK; |
| } |
| |
| Camera2Client::~Camera2Client() { |
| ATRACE_CALL(); |
| ALOGV("~Camera2Client"); |
| |
| mDestructionStarted = true; |
| |
| disconnect(); |
| |
| ALOGI("Camera %d: Closed", mCameraId); |
| } |
| |
| status_t Camera2Client::dump(int fd, const Vector<String16>& args) { |
| String8 result; |
| result.appendFormat("Client2[%d] (%p) Client: %s PID: %d, dump:\n", |
| mCameraId, |
| getRemoteCallback()->asBinder().get(), |
| String8(mClientPackageName).string(), |
| mClientPid); |
| result.append(" State: "); |
| #define CASE_APPEND_ENUM(x) case x: result.append(#x "\n"); break; |
| |
| const Parameters& p = mParameters.unsafeAccess(); |
| |
| result.append(Parameters::getStateName(p.state)); |
| |
| result.append("\n Current parameters:\n"); |
| result.appendFormat(" Preview size: %d x %d\n", |
| p.previewWidth, p.previewHeight); |
| result.appendFormat(" Preview FPS range: %d - %d\n", |
| p.previewFpsRange[0], p.previewFpsRange[1]); |
| result.appendFormat(" Preview HAL pixel format: 0x%x\n", |
| p.previewFormat); |
| result.appendFormat(" Preview transform: %x\n", |
| p.previewTransform); |
| result.appendFormat(" Picture size: %d x %d\n", |
| p.pictureWidth, p.pictureHeight); |
| result.appendFormat(" Jpeg thumbnail size: %d x %d\n", |
| p.jpegThumbSize[0], p.jpegThumbSize[1]); |
| result.appendFormat(" Jpeg quality: %d, thumbnail quality: %d\n", |
| p.jpegQuality, p.jpegThumbQuality); |
| result.appendFormat(" Jpeg rotation: %d\n", p.jpegRotation); |
| result.appendFormat(" GPS tags %s\n", |
| p.gpsEnabled ? "enabled" : "disabled"); |
| if (p.gpsEnabled) { |
| result.appendFormat(" GPS lat x long x alt: %f x %f x %f\n", |
| p.gpsCoordinates[0], p.gpsCoordinates[1], |
| p.gpsCoordinates[2]); |
| result.appendFormat(" GPS timestamp: %lld\n", |
| p.gpsTimestamp); |
| result.appendFormat(" GPS processing method: %s\n", |
| p.gpsProcessingMethod.string()); |
| } |
| |
| result.append(" White balance mode: "); |
| switch (p.wbMode) { |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_AUTO) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_INCANDESCENT) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_FLUORESCENT) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_WARM_FLUORESCENT) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_DAYLIGHT) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_CLOUDY_DAYLIGHT) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_TWILIGHT) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_SHADE) |
| default: result.append("UNKNOWN\n"); |
| } |
| |
| result.append(" Effect mode: "); |
| switch (p.effectMode) { |
| CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_OFF) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_MONO) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_NEGATIVE) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_SOLARIZE) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_SEPIA) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_POSTERIZE) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_WHITEBOARD) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_BLACKBOARD) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_AQUA) |
| default: result.append("UNKNOWN\n"); |
| } |
| |
| result.append(" Antibanding mode: "); |
| switch (p.antibandingMode) { |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AE_ANTIBANDING_MODE_AUTO) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AE_ANTIBANDING_MODE_OFF) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AE_ANTIBANDING_MODE_50HZ) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AE_ANTIBANDING_MODE_60HZ) |
| default: result.append("UNKNOWN\n"); |
| } |
| |
| result.append(" Scene mode: "); |
| switch (p.sceneMode) { |
| case ANDROID_CONTROL_SCENE_MODE_UNSUPPORTED: |
| result.append("AUTO\n"); break; |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_ACTION) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_PORTRAIT) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_LANDSCAPE) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_NIGHT) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_NIGHT_PORTRAIT) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_THEATRE) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_BEACH) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_SNOW) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_SUNSET) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_STEADYPHOTO) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_FIREWORKS) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_SPORTS) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_PARTY) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_CANDLELIGHT) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_BARCODE) |
| default: result.append("UNKNOWN\n"); |
| } |
| |
| result.append(" Flash mode: "); |
| switch (p.flashMode) { |
| CASE_APPEND_ENUM(Parameters::FLASH_MODE_OFF) |
| CASE_APPEND_ENUM(Parameters::FLASH_MODE_AUTO) |
| CASE_APPEND_ENUM(Parameters::FLASH_MODE_ON) |
| CASE_APPEND_ENUM(Parameters::FLASH_MODE_TORCH) |
| CASE_APPEND_ENUM(Parameters::FLASH_MODE_RED_EYE) |
| CASE_APPEND_ENUM(Parameters::FLASH_MODE_INVALID) |
| default: result.append("UNKNOWN\n"); |
| } |
| |
| result.append(" Focus mode: "); |
| switch (p.focusMode) { |
| CASE_APPEND_ENUM(Parameters::FOCUS_MODE_AUTO) |
| CASE_APPEND_ENUM(Parameters::FOCUS_MODE_MACRO) |
| CASE_APPEND_ENUM(Parameters::FOCUS_MODE_CONTINUOUS_VIDEO) |
| CASE_APPEND_ENUM(Parameters::FOCUS_MODE_CONTINUOUS_PICTURE) |
| CASE_APPEND_ENUM(Parameters::FOCUS_MODE_EDOF) |
| CASE_APPEND_ENUM(Parameters::FOCUS_MODE_INFINITY) |
| CASE_APPEND_ENUM(Parameters::FOCUS_MODE_FIXED) |
| CASE_APPEND_ENUM(Parameters::FOCUS_MODE_INVALID) |
| default: result.append("UNKNOWN\n"); |
| } |
| |
| result.append(" Focus state: "); |
| switch (p.focusState) { |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AF_STATE_INACTIVE) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AF_STATE_PASSIVE_SCAN) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AF_STATE_PASSIVE_FOCUSED) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AF_STATE_ACTIVE_SCAN) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AF_STATE_FOCUSED_LOCKED) |
| CASE_APPEND_ENUM(ANDROID_CONTROL_AF_STATE_NOT_FOCUSED_LOCKED) |
| default: result.append("UNKNOWN\n"); |
| } |
| |
| result.append(" Focusing areas:\n"); |
| for (size_t i = 0; i < p.focusingAreas.size(); i++) { |
| result.appendFormat(" [ (%d, %d, %d, %d), weight %d ]\n", |
| p.focusingAreas[i].left, |
| p.focusingAreas[i].top, |
| p.focusingAreas[i].right, |
| p.focusingAreas[i].bottom, |
| p.focusingAreas[i].weight); |
| } |
| |
| result.appendFormat(" Exposure compensation index: %d\n", |
| p.exposureCompensation); |
| |
| result.appendFormat(" AE lock %s, AWB lock %s\n", |
| p.autoExposureLock ? "enabled" : "disabled", |
| p.autoWhiteBalanceLock ? "enabled" : "disabled" ); |
| |
| result.appendFormat(" Metering areas:\n"); |
| for (size_t i = 0; i < p.meteringAreas.size(); i++) { |
| result.appendFormat(" [ (%d, %d, %d, %d), weight %d ]\n", |
| p.meteringAreas[i].left, |
| p.meteringAreas[i].top, |
| p.meteringAreas[i].right, |
| p.meteringAreas[i].bottom, |
| p.meteringAreas[i].weight); |
| } |
| |
| result.appendFormat(" Zoom index: %d\n", p.zoom); |
| result.appendFormat(" Video size: %d x %d\n", p.videoWidth, |
| p.videoHeight); |
| |
| result.appendFormat(" Recording hint is %s\n", |
| p.recordingHint ? "set" : "not set"); |
| |
| result.appendFormat(" Video stabilization is %s\n", |
| p.videoStabilization ? "enabled" : "disabled"); |
| |
| result.append(" Current streams:\n"); |
| result.appendFormat(" Preview stream ID: %d\n", |
| getPreviewStreamId()); |
| result.appendFormat(" Capture stream ID: %d\n", |
| getCaptureStreamId()); |
| result.appendFormat(" Recording stream ID: %d\n", |
| getRecordingStreamId()); |
| |
| result.append(" Quirks for this camera:\n"); |
| bool haveQuirk = false; |
| if (p.quirks.triggerAfWithAuto) { |
| result.appendFormat(" triggerAfWithAuto\n"); |
| haveQuirk = true; |
| } |
| if (p.quirks.useZslFormat) { |
| result.appendFormat(" useZslFormat\n"); |
| haveQuirk = true; |
| } |
| if (p.quirks.meteringCropRegion) { |
| result.appendFormat(" meteringCropRegion\n"); |
| haveQuirk = true; |
| } |
| if (!haveQuirk) { |
| result.appendFormat(" none\n"); |
| } |
| |
| write(fd, result.string(), result.size()); |
| |
| mStreamingProcessor->dump(fd, args); |
| |
| mCaptureSequencer->dump(fd, args); |
| |
| mFrameProcessor->dump(fd, args); |
| |
| mZslProcessor->dump(fd, args); |
| |
| return dumpDevice(fd, args); |
| #undef CASE_APPEND_ENUM |
| } |
| |
| // ICamera interface |
| |
| void Camera2Client::disconnect() { |
| ATRACE_CALL(); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| |
| // Allow both client and the media server to disconnect at all times |
| int callingPid = getCallingPid(); |
| if (callingPid != mClientPid && callingPid != mServicePid) return; |
| |
| if (mDevice == 0) return; |
| |
| ALOGV("Camera %d: Shutting down", mCameraId); |
| |
| /** |
| * disconnect() cannot call any methods that might need to promote a |
| * wp<Camera2Client>, since disconnect can be called from the destructor, at |
| * which point all such promotions will fail. |
| */ |
| |
| stopPreviewL(); |
| |
| { |
| SharedParameters::Lock l(mParameters); |
| if (l.mParameters.state == Parameters::DISCONNECTED) return; |
| l.mParameters.state = Parameters::DISCONNECTED; |
| } |
| |
| mStreamingProcessor->deletePreviewStream(); |
| mStreamingProcessor->deleteRecordingStream(); |
| mJpegProcessor->deleteStream(); |
| mCallbackProcessor->deleteStream(); |
| mZslProcessor->deleteStream(); |
| |
| mFrameProcessor->requestExit(); |
| mCaptureSequencer->requestExit(); |
| mJpegProcessor->requestExit(); |
| mZslProcessorThread->requestExit(); |
| mCallbackProcessor->requestExit(); |
| |
| ALOGV("Camera %d: Waiting for threads", mCameraId); |
| |
| mFrameProcessor->join(); |
| mCaptureSequencer->join(); |
| mJpegProcessor->join(); |
| mZslProcessorThread->join(); |
| mCallbackProcessor->join(); |
| |
| ALOGV("Camera %d: Disconnecting device", mCameraId); |
| |
| mDevice->disconnect(); |
| |
| mDevice.clear(); |
| |
| CameraService::Client::disconnect(); |
| } |
| |
| status_t Camera2Client::connect(const sp<ICameraClient>& client) { |
| ATRACE_CALL(); |
| ALOGV("%s: E", __FUNCTION__); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| |
| if (mClientPid != 0 && getCallingPid() != mClientPid) { |
| ALOGE("%s: Camera %d: Connection attempt from pid %d; " |
| "current locked to pid %d", __FUNCTION__, |
| mCameraId, getCallingPid(), mClientPid); |
| return BAD_VALUE; |
| } |
| |
| mClientPid = getCallingPid(); |
| |
| mRemoteCallback = client; |
| mSharedCameraCallbacks = client; |
| |
| return OK; |
| } |
| |
| status_t Camera2Client::lock() { |
| ATRACE_CALL(); |
| ALOGV("%s: E", __FUNCTION__); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| ALOGV("%s: Camera %d: Lock call from pid %d; current client pid %d", |
| __FUNCTION__, mCameraId, getCallingPid(), mClientPid); |
| |
| if (mClientPid == 0) { |
| mClientPid = getCallingPid(); |
| return OK; |
| } |
| |
| if (mClientPid != getCallingPid()) { |
| ALOGE("%s: Camera %d: Lock call from pid %d; currently locked to pid %d", |
| __FUNCTION__, mCameraId, getCallingPid(), mClientPid); |
| return EBUSY; |
| } |
| |
| return OK; |
| } |
| |
| status_t Camera2Client::unlock() { |
| ATRACE_CALL(); |
| ALOGV("%s: E", __FUNCTION__); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| ALOGV("%s: Camera %d: Unlock call from pid %d; current client pid %d", |
| __FUNCTION__, mCameraId, getCallingPid(), mClientPid); |
| |
| if (mClientPid == getCallingPid()) { |
| SharedParameters::Lock l(mParameters); |
| if (l.mParameters.state == Parameters::RECORD || |
| l.mParameters.state == Parameters::VIDEO_SNAPSHOT) { |
| ALOGD("Not allowed to unlock camera during recording."); |
| return INVALID_OPERATION; |
| } |
| mClientPid = 0; |
| mRemoteCallback.clear(); |
| mSharedCameraCallbacks.clear(); |
| return OK; |
| } |
| |
| ALOGE("%s: Camera %d: Unlock call from pid %d; currently locked to pid %d", |
| __FUNCTION__, mCameraId, getCallingPid(), mClientPid); |
| return EBUSY; |
| } |
| |
| status_t Camera2Client::setPreviewDisplay( |
| const sp<Surface>& surface) { |
| ATRACE_CALL(); |
| ALOGV("%s: E", __FUNCTION__); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| status_t res; |
| if ( (res = checkPid(__FUNCTION__) ) != OK) return res; |
| |
| sp<IBinder> binder; |
| sp<ANativeWindow> window; |
| if (surface != 0) { |
| binder = surface->getIGraphicBufferProducer()->asBinder(); |
| window = surface; |
| } |
| |
| return setPreviewWindowL(binder,window); |
| } |
| |
| status_t Camera2Client::setPreviewTexture( |
| const sp<IGraphicBufferProducer>& bufferProducer) { |
| ATRACE_CALL(); |
| ALOGV("%s: E", __FUNCTION__); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| status_t res; |
| if ( (res = checkPid(__FUNCTION__) ) != OK) return res; |
| |
| sp<IBinder> binder; |
| sp<ANativeWindow> window; |
| if (bufferProducer != 0) { |
| binder = bufferProducer->asBinder(); |
| window = new Surface(bufferProducer); |
| } |
| return setPreviewWindowL(binder, window); |
| } |
| |
| status_t Camera2Client::setPreviewWindowL(const sp<IBinder>& binder, |
| sp<ANativeWindow> window) { |
| ATRACE_CALL(); |
| status_t res; |
| |
| if (binder == mPreviewSurface) { |
| ALOGV("%s: Camera %d: New window is same as old window", |
| __FUNCTION__, mCameraId); |
| return NO_ERROR; |
| } |
| |
| Parameters::State state; |
| { |
| SharedParameters::Lock l(mParameters); |
| state = l.mParameters.state; |
| } |
| switch (state) { |
| case Parameters::DISCONNECTED: |
| case Parameters::RECORD: |
| case Parameters::STILL_CAPTURE: |
| case Parameters::VIDEO_SNAPSHOT: |
| ALOGE("%s: Camera %d: Cannot set preview display while in state %s", |
| __FUNCTION__, mCameraId, |
| Parameters::getStateName(state)); |
| return INVALID_OPERATION; |
| case Parameters::STOPPED: |
| case Parameters::WAITING_FOR_PREVIEW_WINDOW: |
| // OK |
| break; |
| case Parameters::PREVIEW: |
| // Already running preview - need to stop and create a new stream |
| res = stopStream(); |
| if (res != OK) { |
| ALOGE("%s: Unable to stop preview to swap windows: %s (%d)", |
| __FUNCTION__, strerror(-res), res); |
| return res; |
| } |
| state = Parameters::WAITING_FOR_PREVIEW_WINDOW; |
| break; |
| } |
| |
| mPreviewSurface = binder; |
| res = mStreamingProcessor->setPreviewWindow(window); |
| if (res != OK) { |
| ALOGE("%s: Unable to set new preview window: %s (%d)", |
| __FUNCTION__, strerror(-res), res); |
| return res; |
| } |
| |
| if (state == Parameters::WAITING_FOR_PREVIEW_WINDOW) { |
| SharedParameters::Lock l(mParameters); |
| l.mParameters.state = state; |
| return startPreviewL(l.mParameters, false); |
| } |
| |
| return OK; |
| } |
| |
| void Camera2Client::setPreviewCallbackFlag(int flag) { |
| ATRACE_CALL(); |
| ALOGV("%s: Camera %d: Flag 0x%x", __FUNCTION__, mCameraId, flag); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| |
| if ( checkPid(__FUNCTION__) != OK) return; |
| |
| SharedParameters::Lock l(mParameters); |
| setPreviewCallbackFlagL(l.mParameters, flag); |
| } |
| |
| void Camera2Client::setPreviewCallbackFlagL(Parameters ¶ms, int flag) { |
| status_t res = OK; |
| if (flag & CAMERA_FRAME_CALLBACK_FLAG_ONE_SHOT_MASK) { |
| ALOGV("%s: setting oneshot", __FUNCTION__); |
| params.previewCallbackOneShot = true; |
| } |
| if (params.previewCallbackFlags != (uint32_t)flag) { |
| params.previewCallbackFlags = flag; |
| switch(params.state) { |
| case Parameters::PREVIEW: |
| res = startPreviewL(params, true); |
| break; |
| case Parameters::RECORD: |
| case Parameters::VIDEO_SNAPSHOT: |
| res = startRecordingL(params, true); |
| break; |
| default: |
| break; |
| } |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to refresh request in state %s", |
| __FUNCTION__, mCameraId, |
| Parameters::getStateName(params.state)); |
| } |
| } |
| |
| } |
| |
| status_t Camera2Client::startPreview() { |
| ATRACE_CALL(); |
| ALOGV("%s: E", __FUNCTION__); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| status_t res; |
| if ( (res = checkPid(__FUNCTION__) ) != OK) return res; |
| SharedParameters::Lock l(mParameters); |
| return startPreviewL(l.mParameters, false); |
| } |
| |
| status_t Camera2Client::startPreviewL(Parameters ¶ms, bool restart) { |
| ATRACE_CALL(); |
| status_t res; |
| |
| ALOGV("%s: state == %d, restart = %d", __FUNCTION__, params.state, restart); |
| |
| if ( (params.state == Parameters::PREVIEW || |
| params.state == Parameters::RECORD || |
| params.state == Parameters::VIDEO_SNAPSHOT) |
| && !restart) { |
| // Succeed attempt to re-enter a streaming state |
| ALOGI("%s: Camera %d: Preview already active, ignoring restart", |
| __FUNCTION__, mCameraId); |
| return OK; |
| } |
| if (params.state > Parameters::PREVIEW && !restart) { |
| ALOGE("%s: Can't start preview in state %s", |
| __FUNCTION__, |
| Parameters::getStateName(params.state)); |
| return INVALID_OPERATION; |
| } |
| |
| if (!mStreamingProcessor->haveValidPreviewWindow()) { |
| params.state = Parameters::WAITING_FOR_PREVIEW_WINDOW; |
| return OK; |
| } |
| params.state = Parameters::STOPPED; |
| |
| res = mStreamingProcessor->updatePreviewStream(params); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to update preview stream: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| |
| Vector<uint8_t> outputStreams; |
| bool callbacksEnabled = params.previewCallbackFlags & |
| CAMERA_FRAME_CALLBACK_FLAG_ENABLE_MASK; |
| if (callbacksEnabled) { |
| res = mCallbackProcessor->updateStream(params); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to update callback stream: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| outputStreams.push(getCallbackStreamId()); |
| } |
| if (params.zslMode && !params.recordingHint) { |
| res = mZslProcessor->updateStream(params); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to update ZSL stream: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| outputStreams.push(getZslStreamId()); |
| } |
| |
| outputStreams.push(getPreviewStreamId()); |
| |
| if (!params.recordingHint) { |
| if (!restart) { |
| res = mStreamingProcessor->updatePreviewRequest(params); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Can't set up preview request: " |
| "%s (%d)", __FUNCTION__, mCameraId, |
| strerror(-res), res); |
| return res; |
| } |
| } |
| res = mStreamingProcessor->startStream(StreamingProcessor::PREVIEW, |
| outputStreams); |
| } else { |
| // With recording hint set, we're going to be operating under the |
| // assumption that the user will record video. To optimize recording |
| // startup time, create the necessary output streams for recording and |
| // video snapshot now if they don't already exist. |
| res = mJpegProcessor->updateStream(params); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Can't pre-configure still image " |
| "stream: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| |
| if (!restart) { |
| res = mStreamingProcessor->updateRecordingRequest(params); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Can't set up preview request with " |
| "record hint: %s (%d)", __FUNCTION__, mCameraId, |
| strerror(-res), res); |
| return res; |
| } |
| } |
| res = mStreamingProcessor->startStream(StreamingProcessor::RECORD, |
| outputStreams); |
| } |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to start streaming preview: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| |
| params.state = Parameters::PREVIEW; |
| return OK; |
| } |
| |
| void Camera2Client::stopPreview() { |
| ATRACE_CALL(); |
| ALOGV("%s: E", __FUNCTION__); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| status_t res; |
| if ( (res = checkPid(__FUNCTION__) ) != OK) return; |
| stopPreviewL(); |
| } |
| |
| void Camera2Client::stopPreviewL() { |
| ATRACE_CALL(); |
| status_t res; |
| const nsecs_t kStopCaptureTimeout = 3000000000LL; // 3 seconds |
| Parameters::State state; |
| { |
| SharedParameters::Lock l(mParameters); |
| state = l.mParameters.state; |
| } |
| |
| switch (state) { |
| case Parameters::DISCONNECTED: |
| // Nothing to do. |
| break; |
| case Parameters::STOPPED: |
| case Parameters::VIDEO_SNAPSHOT: |
| case Parameters::STILL_CAPTURE: |
| mCaptureSequencer->waitUntilIdle(kStopCaptureTimeout); |
| // no break |
| case Parameters::RECORD: |
| case Parameters::PREVIEW: |
| res = stopStream(); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Can't stop streaming: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| } |
| res = mDevice->waitUntilDrained(); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Waiting to stop streaming failed: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| } |
| // no break |
| case Parameters::WAITING_FOR_PREVIEW_WINDOW: { |
| SharedParameters::Lock l(mParameters); |
| l.mParameters.state = Parameters::STOPPED; |
| commandStopFaceDetectionL(l.mParameters); |
| break; |
| } |
| default: |
| ALOGE("%s: Camera %d: Unknown state %d", __FUNCTION__, mCameraId, |
| state); |
| } |
| } |
| |
| bool Camera2Client::previewEnabled() { |
| ATRACE_CALL(); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| status_t res; |
| if ( (res = checkPid(__FUNCTION__) ) != OK) return false; |
| |
| SharedParameters::Lock l(mParameters); |
| return l.mParameters.state == Parameters::PREVIEW; |
| } |
| |
| status_t Camera2Client::storeMetaDataInBuffers(bool enabled) { |
| ATRACE_CALL(); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| status_t res; |
| if ( (res = checkPid(__FUNCTION__) ) != OK) return res; |
| |
| SharedParameters::Lock l(mParameters); |
| switch (l.mParameters.state) { |
| case Parameters::RECORD: |
| case Parameters::VIDEO_SNAPSHOT: |
| ALOGE("%s: Camera %d: Can't be called in state %s", |
| __FUNCTION__, mCameraId, |
| Parameters::getStateName(l.mParameters.state)); |
| return INVALID_OPERATION; |
| default: |
| // OK |
| break; |
| } |
| |
| l.mParameters.storeMetadataInBuffers = enabled; |
| |
| return OK; |
| } |
| |
| status_t Camera2Client::startRecording() { |
| ATRACE_CALL(); |
| ALOGV("%s: E", __FUNCTION__); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| status_t res; |
| if ( (res = checkPid(__FUNCTION__) ) != OK) return res; |
| SharedParameters::Lock l(mParameters); |
| |
| return startRecordingL(l.mParameters, false); |
| } |
| |
| status_t Camera2Client::startRecordingL(Parameters ¶ms, bool restart) { |
| status_t res; |
| |
| ALOGV("%s: state == %d, restart = %d", __FUNCTION__, params.state, restart); |
| |
| switch (params.state) { |
| case Parameters::STOPPED: |
| res = startPreviewL(params, false); |
| if (res != OK) return res; |
| break; |
| case Parameters::PREVIEW: |
| // Ready to go |
| break; |
| case Parameters::RECORD: |
| case Parameters::VIDEO_SNAPSHOT: |
| // OK to call this when recording is already on, just skip unless |
| // we're looking to restart |
| if (!restart) return OK; |
| break; |
| default: |
| ALOGE("%s: Camera %d: Can't start recording in state %s", |
| __FUNCTION__, mCameraId, |
| Parameters::getStateName(params.state)); |
| return INVALID_OPERATION; |
| }; |
| |
| if (!params.storeMetadataInBuffers) { |
| ALOGE("%s: Camera %d: Recording only supported in metadata mode, but " |
| "non-metadata recording mode requested!", __FUNCTION__, |
| mCameraId); |
| return INVALID_OPERATION; |
| } |
| |
| if (!restart) { |
| mCameraService->playSound(CameraService::SOUND_RECORDING); |
| mStreamingProcessor->updateRecordingRequest(params); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to update recording request: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| } |
| |
| res = mStreamingProcessor->updateRecordingStream(params); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to update recording stream: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| |
| Vector<uint8_t> outputStreams; |
| bool callbacksEnabled = params.previewCallbackFlags & |
| CAMERA_FRAME_CALLBACK_FLAG_ENABLE_MASK; |
| if (callbacksEnabled) { |
| res = mCallbackProcessor->updateStream(params); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to update callback stream: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| outputStreams.push(getCallbackStreamId()); |
| } |
| outputStreams.push(getPreviewStreamId()); |
| outputStreams.push(getRecordingStreamId()); |
| |
| res = mStreamingProcessor->startStream(StreamingProcessor::RECORD, |
| outputStreams); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to start recording stream: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| |
| if (params.state < Parameters::RECORD) { |
| params.state = Parameters::RECORD; |
| } |
| |
| return OK; |
| } |
| |
| void Camera2Client::stopRecording() { |
| ATRACE_CALL(); |
| ALOGV("%s: E", __FUNCTION__); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| SharedParameters::Lock l(mParameters); |
| |
| status_t res; |
| if ( (res = checkPid(__FUNCTION__) ) != OK) return; |
| |
| switch (l.mParameters.state) { |
| case Parameters::RECORD: |
| // OK to stop |
| break; |
| case Parameters::STOPPED: |
| case Parameters::PREVIEW: |
| case Parameters::STILL_CAPTURE: |
| case Parameters::VIDEO_SNAPSHOT: |
| default: |
| ALOGE("%s: Camera %d: Can't stop recording in state %s", |
| __FUNCTION__, mCameraId, |
| Parameters::getStateName(l.mParameters.state)); |
| return; |
| }; |
| |
| mCameraService->playSound(CameraService::SOUND_RECORDING); |
| |
| res = startPreviewL(l.mParameters, true); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to return to preview", |
| __FUNCTION__, mCameraId); |
| } |
| } |
| |
| bool Camera2Client::recordingEnabled() { |
| ATRACE_CALL(); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| |
| if ( checkPid(__FUNCTION__) != OK) return false; |
| |
| return recordingEnabledL(); |
| } |
| |
| bool Camera2Client::recordingEnabledL() { |
| ATRACE_CALL(); |
| SharedParameters::Lock l(mParameters); |
| |
| return (l.mParameters.state == Parameters::RECORD |
| || l.mParameters.state == Parameters::VIDEO_SNAPSHOT); |
| } |
| |
| void Camera2Client::releaseRecordingFrame(const sp<IMemory>& mem) { |
| ATRACE_CALL(); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| if ( checkPid(__FUNCTION__) != OK) return; |
| |
| mStreamingProcessor->releaseRecordingFrame(mem); |
| } |
| |
| status_t Camera2Client::autoFocus() { |
| ATRACE_CALL(); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| ALOGV("%s: Camera %d", __FUNCTION__, mCameraId); |
| status_t res; |
| if ( (res = checkPid(__FUNCTION__) ) != OK) return res; |
| |
| int triggerId; |
| bool notifyImmediately = false; |
| bool notifySuccess = false; |
| { |
| SharedParameters::Lock l(mParameters); |
| if (l.mParameters.state < Parameters::PREVIEW) { |
| return INVALID_OPERATION; |
| } |
| |
| /** |
| * If the camera does not support auto-focus, it is a no-op and |
| * onAutoFocus(boolean, Camera) callback will be called immediately |
| * with a fake value of success set to true. |
| */ |
| if (l.mParameters.focusMode == Parameters::FOCUS_MODE_FIXED) { |
| notifyImmediately = true; |
| notifySuccess = true; |
| } |
| /** |
| * If we're in CAF mode, and AF has already been locked, just fire back |
| * the callback right away; the HAL would not send a notification since |
| * no state change would happen on a AF trigger. |
| */ |
| if ( (l.mParameters.focusMode == Parameters::FOCUS_MODE_CONTINUOUS_PICTURE || |
| l.mParameters.focusMode == Parameters::FOCUS_MODE_CONTINUOUS_VIDEO) && |
| l.mParameters.focusState == ANDROID_CONTROL_AF_STATE_FOCUSED_LOCKED ) { |
| notifyImmediately = true; |
| notifySuccess = true; |
| } |
| /** |
| * Send immediate notification back to client |
| */ |
| if (notifyImmediately) { |
| SharedCameraCallbacks::Lock l(mSharedCameraCallbacks); |
| if (l.mRemoteCallback != 0) { |
| l.mRemoteCallback->notifyCallback(CAMERA_MSG_FOCUS, |
| notifySuccess ? 1 : 0, 0); |
| } |
| return OK; |
| } |
| /** |
| * Handle quirk mode for AF in scene modes |
| */ |
| if (l.mParameters.quirks.triggerAfWithAuto && |
| l.mParameters.sceneMode != ANDROID_CONTROL_SCENE_MODE_UNSUPPORTED && |
| l.mParameters.focusMode != Parameters::FOCUS_MODE_AUTO && |
| !l.mParameters.focusingAreas[0].isEmpty()) { |
| ALOGV("%s: Quirk: Switching from focusMode %d to AUTO", |
| __FUNCTION__, l.mParameters.focusMode); |
| l.mParameters.shadowFocusMode = l.mParameters.focusMode; |
| l.mParameters.focusMode = Parameters::FOCUS_MODE_AUTO; |
| updateRequests(l.mParameters); |
| } |
| |
| l.mParameters.currentAfTriggerId = ++l.mParameters.afTriggerCounter; |
| triggerId = l.mParameters.currentAfTriggerId; |
| } |
| syncWithDevice(); |
| |
| mDevice->triggerAutofocus(triggerId); |
| |
| return OK; |
| } |
| |
| status_t Camera2Client::cancelAutoFocus() { |
| ATRACE_CALL(); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| ALOGV("%s: Camera %d", __FUNCTION__, mCameraId); |
| status_t res; |
| if ( (res = checkPid(__FUNCTION__) ) != OK) return res; |
| |
| int triggerId; |
| { |
| SharedParameters::Lock l(mParameters); |
| triggerId = ++l.mParameters.afTriggerCounter; |
| |
| // When using triggerAfWithAuto quirk, may need to reset focus mode to |
| // the real state at this point. No need to cancel explicitly if |
| // changing the AF mode. |
| if (l.mParameters.shadowFocusMode != Parameters::FOCUS_MODE_INVALID) { |
| ALOGV("%s: Quirk: Restoring focus mode to %d", __FUNCTION__, |
| l.mParameters.shadowFocusMode); |
| l.mParameters.focusMode = l.mParameters.shadowFocusMode; |
| l.mParameters.shadowFocusMode = Parameters::FOCUS_MODE_INVALID; |
| updateRequests(l.mParameters); |
| |
| return OK; |
| } |
| } |
| syncWithDevice(); |
| |
| mDevice->triggerCancelAutofocus(triggerId); |
| |
| return OK; |
| } |
| |
| status_t Camera2Client::takePicture(int msgType) { |
| ATRACE_CALL(); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| status_t res; |
| if ( (res = checkPid(__FUNCTION__) ) != OK) return res; |
| |
| { |
| SharedParameters::Lock l(mParameters); |
| switch (l.mParameters.state) { |
| case Parameters::DISCONNECTED: |
| case Parameters::STOPPED: |
| case Parameters::WAITING_FOR_PREVIEW_WINDOW: |
| ALOGE("%s: Camera %d: Cannot take picture without preview enabled", |
| __FUNCTION__, mCameraId); |
| return INVALID_OPERATION; |
| case Parameters::PREVIEW: |
| // Good to go for takePicture |
| res = commandStopFaceDetectionL(l.mParameters); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to stop face detection for still capture", |
| __FUNCTION__, mCameraId); |
| return res; |
| } |
| l.mParameters.state = Parameters::STILL_CAPTURE; |
| break; |
| case Parameters::RECORD: |
| // Good to go for video snapshot |
| l.mParameters.state = Parameters::VIDEO_SNAPSHOT; |
| break; |
| case Parameters::STILL_CAPTURE: |
| case Parameters::VIDEO_SNAPSHOT: |
| ALOGE("%s: Camera %d: Already taking a picture", |
| __FUNCTION__, mCameraId); |
| return INVALID_OPERATION; |
| } |
| |
| ALOGV("%s: Camera %d: Starting picture capture", __FUNCTION__, mCameraId); |
| |
| res = mJpegProcessor->updateStream(l.mParameters); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Can't set up still image stream: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| } |
| |
| // Need HAL to have correct settings before (possibly) triggering precapture |
| syncWithDevice(); |
| |
| res = mCaptureSequencer->startCapture(msgType); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to start capture: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| } |
| |
| return res; |
| } |
| |
| status_t Camera2Client::setParameters(const String8& params) { |
| ATRACE_CALL(); |
| ALOGV("%s: Camera %d", __FUNCTION__, mCameraId); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| status_t res; |
| if ( (res = checkPid(__FUNCTION__) ) != OK) return res; |
| |
| SharedParameters::Lock l(mParameters); |
| |
| res = l.mParameters.set(params); |
| if (res != OK) return res; |
| |
| res = updateRequests(l.mParameters); |
| |
| return res; |
| } |
| |
| String8 Camera2Client::getParameters() const { |
| ATRACE_CALL(); |
| ALOGV("%s: Camera %d", __FUNCTION__, mCameraId); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| if ( checkPid(__FUNCTION__) != OK) return String8(); |
| |
| SharedParameters::ReadLock l(mParameters); |
| |
| return l.mParameters.get(); |
| } |
| |
| status_t Camera2Client::sendCommand(int32_t cmd, int32_t arg1, int32_t arg2) { |
| ATRACE_CALL(); |
| Mutex::Autolock icl(mBinderSerializationLock); |
| status_t res; |
| if ( (res = checkPid(__FUNCTION__) ) != OK) return res; |
| |
| ALOGV("%s: Camera %d: Command %d (%d, %d)", __FUNCTION__, mCameraId, |
| cmd, arg1, arg2); |
| |
| switch (cmd) { |
| case CAMERA_CMD_START_SMOOTH_ZOOM: |
| return commandStartSmoothZoomL(); |
| case CAMERA_CMD_STOP_SMOOTH_ZOOM: |
| return commandStopSmoothZoomL(); |
| case CAMERA_CMD_SET_DISPLAY_ORIENTATION: |
| return commandSetDisplayOrientationL(arg1); |
| case CAMERA_CMD_ENABLE_SHUTTER_SOUND: |
| return commandEnableShutterSoundL(arg1 == 1); |
| case CAMERA_CMD_PLAY_RECORDING_SOUND: |
| return commandPlayRecordingSoundL(); |
| case CAMERA_CMD_START_FACE_DETECTION: |
| return commandStartFaceDetectionL(arg1); |
| case CAMERA_CMD_STOP_FACE_DETECTION: { |
| SharedParameters::Lock l(mParameters); |
| return commandStopFaceDetectionL(l.mParameters); |
| } |
| case CAMERA_CMD_ENABLE_FOCUS_MOVE_MSG: |
| return commandEnableFocusMoveMsgL(arg1 == 1); |
| case CAMERA_CMD_PING: |
| return commandPingL(); |
| case CAMERA_CMD_SET_VIDEO_BUFFER_COUNT: |
| return commandSetVideoBufferCountL(arg1); |
| default: |
| ALOGE("%s: Unknown command %d (arguments %d, %d)", |
| __FUNCTION__, cmd, arg1, arg2); |
| return BAD_VALUE; |
| } |
| } |
| |
| status_t Camera2Client::commandStartSmoothZoomL() { |
| ALOGE("%s: Unimplemented!", __FUNCTION__); |
| return OK; |
| } |
| |
| status_t Camera2Client::commandStopSmoothZoomL() { |
| ALOGE("%s: Unimplemented!", __FUNCTION__); |
| return OK; |
| } |
| |
| status_t Camera2Client::commandSetDisplayOrientationL(int degrees) { |
| int transform = Parameters::degToTransform(degrees, |
| mCameraFacing == CAMERA_FACING_FRONT); |
| if (transform == -1) { |
| ALOGE("%s: Camera %d: Error setting %d as display orientation value", |
| __FUNCTION__, mCameraId, degrees); |
| return BAD_VALUE; |
| } |
| SharedParameters::Lock l(mParameters); |
| if (transform != l.mParameters.previewTransform && |
| getPreviewStreamId() != NO_STREAM) { |
| mDevice->setStreamTransform(getPreviewStreamId(), transform); |
| } |
| l.mParameters.previewTransform = transform; |
| return OK; |
| } |
| |
| status_t Camera2Client::commandEnableShutterSoundL(bool enable) { |
| SharedParameters::Lock l(mParameters); |
| if (enable) { |
| l.mParameters.playShutterSound = true; |
| return OK; |
| } |
| |
| // Disabling shutter sound may not be allowed. In that case only |
| // allow the mediaserver process to disable the sound. |
| char value[PROPERTY_VALUE_MAX]; |
| property_get("ro.camera.sound.forced", value, "0"); |
| if (strncmp(value, "0", 2) != 0) { |
| // Disabling shutter sound is not allowed. Deny if the current |
| // process is not mediaserver. |
| if (getCallingPid() != getpid()) { |
| ALOGE("Failed to disable shutter sound. Permission denied (pid %d)", |
| getCallingPid()); |
| return PERMISSION_DENIED; |
| } |
| } |
| |
| l.mParameters.playShutterSound = false; |
| return OK; |
| } |
| |
| status_t Camera2Client::commandPlayRecordingSoundL() { |
| mCameraService->playSound(CameraService::SOUND_RECORDING); |
| return OK; |
| } |
| |
| status_t Camera2Client::commandStartFaceDetectionL(int /*type*/) { |
| ALOGV("%s: Camera %d: Starting face detection", |
| __FUNCTION__, mCameraId); |
| status_t res; |
| SharedParameters::Lock l(mParameters); |
| switch (l.mParameters.state) { |
| case Parameters::DISCONNECTED: |
| case Parameters::STOPPED: |
| case Parameters::WAITING_FOR_PREVIEW_WINDOW: |
| case Parameters::STILL_CAPTURE: |
| ALOGE("%s: Camera %d: Cannot start face detection without preview active", |
| __FUNCTION__, mCameraId); |
| return INVALID_OPERATION; |
| case Parameters::PREVIEW: |
| case Parameters::RECORD: |
| case Parameters::VIDEO_SNAPSHOT: |
| // Good to go for starting face detect |
| break; |
| } |
| // Ignoring type |
| if (l.mParameters.fastInfo.bestFaceDetectMode == |
| ANDROID_STATISTICS_FACE_DETECT_MODE_OFF) { |
| ALOGE("%s: Camera %d: Face detection not supported", |
| __FUNCTION__, mCameraId); |
| return BAD_VALUE; |
| } |
| if (l.mParameters.enableFaceDetect) return OK; |
| |
| l.mParameters.enableFaceDetect = true; |
| |
| res = updateRequests(l.mParameters); |
| |
| return res; |
| } |
| |
| status_t Camera2Client::commandStopFaceDetectionL(Parameters ¶ms) { |
| status_t res = OK; |
| ALOGV("%s: Camera %d: Stopping face detection", |
| __FUNCTION__, mCameraId); |
| |
| if (!params.enableFaceDetect) return OK; |
| |
| params.enableFaceDetect = false; |
| |
| if (params.state == Parameters::PREVIEW |
| || params.state == Parameters::RECORD |
| || params.state == Parameters::VIDEO_SNAPSHOT) { |
| res = updateRequests(params); |
| } |
| |
| return res; |
| } |
| |
| status_t Camera2Client::commandEnableFocusMoveMsgL(bool enable) { |
| SharedParameters::Lock l(mParameters); |
| l.mParameters.enableFocusMoveMessages = enable; |
| |
| return OK; |
| } |
| |
| status_t Camera2Client::commandPingL() { |
| // Always ping back if access is proper and device is alive |
| SharedParameters::Lock l(mParameters); |
| if (l.mParameters.state != Parameters::DISCONNECTED) { |
| return OK; |
| } else { |
| return NO_INIT; |
| } |
| } |
| |
| status_t Camera2Client::commandSetVideoBufferCountL(size_t count) { |
| if (recordingEnabledL()) { |
| ALOGE("%s: Camera %d: Error setting video buffer count after " |
| "recording was started", __FUNCTION__, mCameraId); |
| return INVALID_OPERATION; |
| } |
| |
| return mStreamingProcessor->setRecordingBufferCount(count); |
| } |
| |
| /** Device-related methods */ |
| void Camera2Client::notifyAutoFocus(uint8_t newState, int triggerId) { |
| ALOGV("%s: Autofocus state now %d, last trigger %d", |
| __FUNCTION__, newState, triggerId); |
| bool sendCompletedMessage = false; |
| bool sendMovingMessage = false; |
| |
| bool success = false; |
| bool afInMotion = false; |
| { |
| SharedParameters::Lock l(mParameters); |
| l.mParameters.focusState = newState; |
| switch (l.mParameters.focusMode) { |
| case Parameters::FOCUS_MODE_AUTO: |
| case Parameters::FOCUS_MODE_MACRO: |
| // Don't send notifications upstream if they're not for the current AF |
| // trigger. For example, if cancel was called in between, or if we |
| // already sent a notification about this AF call. |
| if (triggerId != l.mParameters.currentAfTriggerId) break; |
| switch (newState) { |
| case ANDROID_CONTROL_AF_STATE_FOCUSED_LOCKED: |
| success = true; |
| // no break |
| case ANDROID_CONTROL_AF_STATE_NOT_FOCUSED_LOCKED: |
| sendCompletedMessage = true; |
| l.mParameters.currentAfTriggerId = -1; |
| break; |
| case ANDROID_CONTROL_AF_STATE_ACTIVE_SCAN: |
| // Just starting focusing, ignore |
| break; |
| case ANDROID_CONTROL_AF_STATE_INACTIVE: |
| case ANDROID_CONTROL_AF_STATE_PASSIVE_SCAN: |
| case ANDROID_CONTROL_AF_STATE_PASSIVE_FOCUSED: |
| default: |
| // Unexpected in AUTO/MACRO mode |
| ALOGE("%s: Unexpected AF state transition in AUTO/MACRO mode: %d", |
| __FUNCTION__, newState); |
| break; |
| } |
| break; |
| case Parameters::FOCUS_MODE_CONTINUOUS_VIDEO: |
| case Parameters::FOCUS_MODE_CONTINUOUS_PICTURE: |
| switch (newState) { |
| case ANDROID_CONTROL_AF_STATE_FOCUSED_LOCKED: |
| success = true; |
| // no break |
| case ANDROID_CONTROL_AF_STATE_NOT_FOCUSED_LOCKED: |
| // Don't send notifications upstream if they're not for |
| // the current AF trigger. For example, if cancel was |
| // called in between, or if we already sent a |
| // notification about this AF call. |
| // Send both a 'AF done' callback and a 'AF move' callback |
| if (triggerId != l.mParameters.currentAfTriggerId) break; |
| sendCompletedMessage = true; |
| afInMotion = false; |
| if (l.mParameters.enableFocusMoveMessages && |
| l.mParameters.afInMotion) { |
| sendMovingMessage = true; |
| } |
| l.mParameters.currentAfTriggerId = -1; |
| break; |
| case ANDROID_CONTROL_AF_STATE_INACTIVE: |
| // Cancel was called, or we switched state; care if |
| // currently moving |
| afInMotion = false; |
| if (l.mParameters.enableFocusMoveMessages && |
| l.mParameters.afInMotion) { |
| sendMovingMessage = true; |
| } |
| break; |
| case ANDROID_CONTROL_AF_STATE_PASSIVE_SCAN: |
| // Start passive scan, inform upstream |
| afInMotion = true; |
| // no break |
| case ANDROID_CONTROL_AF_STATE_PASSIVE_FOCUSED: |
| // Stop passive scan, inform upstream |
| if (l.mParameters.enableFocusMoveMessages) { |
| sendMovingMessage = true; |
| } |
| break; |
| } |
| l.mParameters.afInMotion = afInMotion; |
| break; |
| case Parameters::FOCUS_MODE_EDOF: |
| case Parameters::FOCUS_MODE_INFINITY: |
| case Parameters::FOCUS_MODE_FIXED: |
| default: |
| if (newState != ANDROID_CONTROL_AF_STATE_INACTIVE) { |
| ALOGE("%s: Unexpected AF state change %d " |
| "(ID %d) in focus mode %d", |
| __FUNCTION__, newState, triggerId, |
| l.mParameters.focusMode); |
| } |
| } |
| } |
| if (sendMovingMessage) { |
| SharedCameraCallbacks::Lock l(mSharedCameraCallbacks); |
| if (l.mRemoteCallback != 0) { |
| l.mRemoteCallback->notifyCallback(CAMERA_MSG_FOCUS_MOVE, |
| afInMotion ? 1 : 0, 0); |
| } |
| } |
| if (sendCompletedMessage) { |
| SharedCameraCallbacks::Lock l(mSharedCameraCallbacks); |
| if (l.mRemoteCallback != 0) { |
| l.mRemoteCallback->notifyCallback(CAMERA_MSG_FOCUS, |
| success ? 1 : 0, 0); |
| } |
| } |
| } |
| |
| void Camera2Client::notifyAutoExposure(uint8_t newState, int triggerId) { |
| ALOGV("%s: Autoexposure state now %d, last trigger %d", |
| __FUNCTION__, newState, triggerId); |
| mCaptureSequencer->notifyAutoExposure(newState, triggerId); |
| } |
| |
| camera2::SharedParameters& Camera2Client::getParameters() { |
| return mParameters; |
| } |
| |
| int Camera2Client::getPreviewStreamId() const { |
| return mStreamingProcessor->getPreviewStreamId(); |
| } |
| |
| int Camera2Client::getCaptureStreamId() const { |
| return mJpegProcessor->getStreamId(); |
| } |
| |
| int Camera2Client::getCallbackStreamId() const { |
| return mCallbackProcessor->getStreamId(); |
| } |
| |
| int Camera2Client::getRecordingStreamId() const { |
| return mStreamingProcessor->getRecordingStreamId(); |
| } |
| |
| int Camera2Client::getZslStreamId() const { |
| return mZslProcessor->getStreamId(); |
| } |
| |
| status_t Camera2Client::registerFrameListener(int32_t minId, int32_t maxId, |
| wp<camera2::FrameProcessor::FilteredListener> listener) { |
| return mFrameProcessor->registerListener(minId, maxId, listener); |
| } |
| |
| status_t Camera2Client::removeFrameListener(int32_t minId, int32_t maxId, |
| wp<camera2::FrameProcessor::FilteredListener> listener) { |
| return mFrameProcessor->removeListener(minId, maxId, listener); |
| } |
| |
| status_t Camera2Client::stopStream() { |
| return mStreamingProcessor->stopStream(); |
| } |
| |
| const int32_t Camera2Client::kPreviewRequestIdStart; |
| const int32_t Camera2Client::kPreviewRequestIdEnd; |
| const int32_t Camera2Client::kRecordingRequestIdStart; |
| const int32_t Camera2Client::kRecordingRequestIdEnd; |
| const int32_t Camera2Client::kCaptureRequestIdStart; |
| const int32_t Camera2Client::kCaptureRequestIdEnd; |
| |
| /** Utility methods */ |
| |
| status_t Camera2Client::updateRequests(Parameters ¶ms) { |
| status_t res; |
| |
| ALOGV("%s: Camera %d: state = %d", __FUNCTION__, getCameraId(), params.state); |
| |
| res = mStreamingProcessor->incrementStreamingIds(); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to increment request IDs: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| |
| res = mStreamingProcessor->updatePreviewRequest(params); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to update preview request: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| res = mStreamingProcessor->updateRecordingRequest(params); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Unable to update recording request: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| |
| if (params.state == Parameters::PREVIEW) { |
| res = startPreviewL(params, true); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Error streaming new preview request: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| } else if (params.state == Parameters::RECORD || |
| params.state == Parameters::VIDEO_SNAPSHOT) { |
| res = startRecordingL(params, true); |
| if (res != OK) { |
| ALOGE("%s: Camera %d: Error streaming new record request: %s (%d)", |
| __FUNCTION__, mCameraId, strerror(-res), res); |
| return res; |
| } |
| } |
| return res; |
| } |
| |
| |
| size_t Camera2Client::calculateBufferSize(int width, int height, |
| int format, int stride) { |
| switch (format) { |
| case HAL_PIXEL_FORMAT_YCbCr_422_SP: // NV16 |
| return width * height * 2; |
| case HAL_PIXEL_FORMAT_YCrCb_420_SP: // NV21 |
| return width * height * 3 / 2; |
| case HAL_PIXEL_FORMAT_YCbCr_422_I: // YUY2 |
| return width * height * 2; |
| case HAL_PIXEL_FORMAT_YV12: { // YV12 |
| size_t ySize = stride * height; |
| size_t uvStride = (stride / 2 + 0xF) & ~0xF; |
| size_t uvSize = uvStride * height / 2; |
| return ySize + uvSize * 2; |
| } |
| case HAL_PIXEL_FORMAT_RGB_565: |
| return width * height * 2; |
| case HAL_PIXEL_FORMAT_RGBA_8888: |
| return width * height * 4; |
| case HAL_PIXEL_FORMAT_RAW_SENSOR: |
| return width * height * 2; |
| default: |
| ALOGE("%s: Unknown preview format: %x", |
| __FUNCTION__, format); |
| return 0; |
| } |
| } |
| |
| status_t Camera2Client::syncWithDevice() { |
| ATRACE_CALL(); |
| const nsecs_t kMaxSyncTimeout = 500000000; // 500 ms |
| status_t res; |
| |
| int32_t activeRequestId = mStreamingProcessor->getActiveRequestId(); |
| if (activeRequestId == 0) return OK; |
| |
| res = mDevice->waitUntilRequestReceived(activeRequestId, kMaxSyncTimeout); |
| if (res == TIMED_OUT) { |
| ALOGE("%s: Camera %d: Timed out waiting sync with HAL", |
| __FUNCTION__, mCameraId); |
| } else if (res != OK) { |
| ALOGE("%s: Camera %d: Error while waiting to sync with HAL", |
| __FUNCTION__, mCameraId); |
| } |
| return res; |
| } |
| |
| |
| } // namespace android |