Camera2Client.cpp revision 28d0327bc80e1e424c3db34c62e8f73cd5ffab1e
1/* 2 * Copyright (C) 2012 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17#define LOG_TAG "Camera2Client" 18#define ATRACE_TAG ATRACE_TAG_CAMERA 19//#define LOG_NDEBUG 0 20 21#include <inttypes.h> 22#include <utils/Log.h> 23#include <utils/Trace.h> 24 25#include <cutils/properties.h> 26#include <gui/Surface.h> 27 28#include "api1/Camera2Client.h" 29 30#include "api1/client2/StreamingProcessor.h" 31#include "api1/client2/JpegProcessor.h" 32#include "api1/client2/CaptureSequencer.h" 33#include "api1/client2/CallbackProcessor.h" 34#include "api1/client2/ZslProcessor.h" 35#include "api1/client2/ZslProcessor3.h" 36 37#define ALOG1(...) ALOGD_IF(gLogLevel >= 1, __VA_ARGS__); 38#define ALOG2(...) ALOGD_IF(gLogLevel >= 2, __VA_ARGS__); 39 40namespace android { 41using namespace camera2; 42 43static int getCallingPid() { 44 return IPCThreadState::self()->getCallingPid(); 45} 46 47// Interface used by CameraService 48 49Camera2Client::Camera2Client(const sp<CameraService>& cameraService, 50 const sp<ICameraClient>& cameraClient, 51 const String16& clientPackageName, 52 int cameraId, 53 int cameraFacing, 54 int clientPid, 55 uid_t clientUid, 56 int servicePid): 57 Camera2ClientBase(cameraService, cameraClient, clientPackageName, 58 cameraId, cameraFacing, clientPid, clientUid, servicePid), 59 mParameters(cameraId, cameraFacing) 60{ 61 ATRACE_CALL(); 62 63 SharedParameters::Lock l(mParameters); 64 l.mParameters.state = Parameters::DISCONNECTED; 65} 66 67status_t Camera2Client::initialize(camera_module_t *module) 68{ 69 ATRACE_CALL(); 70 ALOGV("%s: Initializing client for camera %d", __FUNCTION__, mCameraId); 71 status_t res; 72 73 res = Camera2ClientBase::initialize(module); 74 if (res != OK) { 75 return res; 76 } 77 78 { 79 SharedParameters::Lock l(mParameters); 80 81 res = l.mParameters.initialize(&(mDevice->info()), mDeviceVersion); 82 if (res != OK) { 83 ALOGE("%s: Camera %d: unable to build defaults: %s (%d)", 84 __FUNCTION__, mCameraId, strerror(-res), res); 85 return NO_INIT; 86 } 87 } 88 89 String8 threadName; 90 91 mStreamingProcessor = new StreamingProcessor(this); 92 threadName = String8::format("C2-%d-StreamProc", 93 mCameraId); 94 mStreamingProcessor->run(threadName.string()); 95 96 mFrameProcessor = new FrameProcessor(mDevice, this); 97 threadName = String8::format("C2-%d-FrameProc", 98 mCameraId); 99 mFrameProcessor->run(threadName.string()); 100 101 mCaptureSequencer = new CaptureSequencer(this); 102 threadName = String8::format("C2-%d-CaptureSeq", 103 mCameraId); 104 mCaptureSequencer->run(threadName.string()); 105 106 mJpegProcessor = new JpegProcessor(this, mCaptureSequencer); 107 threadName = String8::format("C2-%d-JpegProc", 108 mCameraId); 109 mJpegProcessor->run(threadName.string()); 110 111 switch (mDeviceVersion) { 112 case CAMERA_DEVICE_API_VERSION_2_0: { 113 sp<ZslProcessor> zslProc = 114 new ZslProcessor(this, mCaptureSequencer); 115 mZslProcessor = zslProc; 116 mZslProcessorThread = zslProc; 117 break; 118 } 119 case CAMERA_DEVICE_API_VERSION_3_0: 120 case CAMERA_DEVICE_API_VERSION_3_1: 121 case CAMERA_DEVICE_API_VERSION_3_2: { 122 sp<ZslProcessor3> zslProc = 123 new ZslProcessor3(this, mCaptureSequencer); 124 mZslProcessor = zslProc; 125 mZslProcessorThread = zslProc; 126 break; 127 } 128 default: 129 break; 130 } 131 threadName = String8::format("C2-%d-ZslProc", 132 mCameraId); 133 mZslProcessorThread->run(threadName.string()); 134 135 mCallbackProcessor = new CallbackProcessor(this); 136 threadName = String8::format("C2-%d-CallbkProc", 137 mCameraId); 138 mCallbackProcessor->run(threadName.string()); 139 140 if (gLogLevel >= 1) { 141 SharedParameters::Lock l(mParameters); 142 ALOGD("%s: Default parameters converted from camera %d:", __FUNCTION__, 143 mCameraId); 144 ALOGD("%s", l.mParameters.paramsFlattened.string()); 145 } 146 147 return OK; 148} 149 150Camera2Client::~Camera2Client() { 151 ATRACE_CALL(); 152 ALOGV("~Camera2Client"); 153 154 mDestructionStarted = true; 155 156 disconnect(); 157 158 ALOGI("Camera %d: Closed", mCameraId); 159} 160 161status_t Camera2Client::dump(int fd, const Vector<String16>& args) { 162 String8 result; 163 result.appendFormat("Client2[%d] (%p) Client: %s PID: %d, dump:\n", 164 mCameraId, 165 getRemoteCallback()->asBinder().get(), 166 String8(mClientPackageName).string(), 167 mClientPid); 168 result.append(" State: "); 169#define CASE_APPEND_ENUM(x) case x: result.append(#x "\n"); break; 170 171 const Parameters& p = mParameters.unsafeAccess(); 172 173 result.append(Parameters::getStateName(p.state)); 174 175 result.append("\n Current parameters:\n"); 176 result.appendFormat(" Preview size: %d x %d\n", 177 p.previewWidth, p.previewHeight); 178 result.appendFormat(" Preview FPS range: %d - %d\n", 179 p.previewFpsRange[0], p.previewFpsRange[1]); 180 result.appendFormat(" Preview HAL pixel format: 0x%x\n", 181 p.previewFormat); 182 result.appendFormat(" Preview transform: %x\n", 183 p.previewTransform); 184 result.appendFormat(" Picture size: %d x %d\n", 185 p.pictureWidth, p.pictureHeight); 186 result.appendFormat(" Jpeg thumbnail size: %d x %d\n", 187 p.jpegThumbSize[0], p.jpegThumbSize[1]); 188 result.appendFormat(" Jpeg quality: %d, thumbnail quality: %d\n", 189 p.jpegQuality, p.jpegThumbQuality); 190 result.appendFormat(" Jpeg rotation: %d\n", p.jpegRotation); 191 result.appendFormat(" GPS tags %s\n", 192 p.gpsEnabled ? "enabled" : "disabled"); 193 if (p.gpsEnabled) { 194 result.appendFormat(" GPS lat x long x alt: %f x %f x %f\n", 195 p.gpsCoordinates[0], p.gpsCoordinates[1], 196 p.gpsCoordinates[2]); 197 result.appendFormat(" GPS timestamp: %" PRId64 "\n", 198 p.gpsTimestamp); 199 result.appendFormat(" GPS processing method: %s\n", 200 p.gpsProcessingMethod.string()); 201 } 202 203 result.append(" White balance mode: "); 204 switch (p.wbMode) { 205 CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_AUTO) 206 CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_INCANDESCENT) 207 CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_FLUORESCENT) 208 CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_WARM_FLUORESCENT) 209 CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_DAYLIGHT) 210 CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_CLOUDY_DAYLIGHT) 211 CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_TWILIGHT) 212 CASE_APPEND_ENUM(ANDROID_CONTROL_AWB_MODE_SHADE) 213 default: result.append("UNKNOWN\n"); 214 } 215 216 result.append(" Effect mode: "); 217 switch (p.effectMode) { 218 CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_OFF) 219 CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_MONO) 220 CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_NEGATIVE) 221 CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_SOLARIZE) 222 CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_SEPIA) 223 CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_POSTERIZE) 224 CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_WHITEBOARD) 225 CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_BLACKBOARD) 226 CASE_APPEND_ENUM(ANDROID_CONTROL_EFFECT_MODE_AQUA) 227 default: result.append("UNKNOWN\n"); 228 } 229 230 result.append(" Antibanding mode: "); 231 switch (p.antibandingMode) { 232 CASE_APPEND_ENUM(ANDROID_CONTROL_AE_ANTIBANDING_MODE_AUTO) 233 CASE_APPEND_ENUM(ANDROID_CONTROL_AE_ANTIBANDING_MODE_OFF) 234 CASE_APPEND_ENUM(ANDROID_CONTROL_AE_ANTIBANDING_MODE_50HZ) 235 CASE_APPEND_ENUM(ANDROID_CONTROL_AE_ANTIBANDING_MODE_60HZ) 236 default: result.append("UNKNOWN\n"); 237 } 238 239 result.append(" Scene mode: "); 240 switch (p.sceneMode) { 241 case ANDROID_CONTROL_SCENE_MODE_DISABLED: 242 result.append("AUTO\n"); break; 243 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_ACTION) 244 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_PORTRAIT) 245 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_LANDSCAPE) 246 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_NIGHT) 247 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_NIGHT_PORTRAIT) 248 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_THEATRE) 249 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_BEACH) 250 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_SNOW) 251 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_SUNSET) 252 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_STEADYPHOTO) 253 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_FIREWORKS) 254 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_SPORTS) 255 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_PARTY) 256 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_CANDLELIGHT) 257 CASE_APPEND_ENUM(ANDROID_CONTROL_SCENE_MODE_BARCODE) 258 default: result.append("UNKNOWN\n"); 259 } 260 261 result.append(" Flash mode: "); 262 switch (p.flashMode) { 263 CASE_APPEND_ENUM(Parameters::FLASH_MODE_OFF) 264 CASE_APPEND_ENUM(Parameters::FLASH_MODE_AUTO) 265 CASE_APPEND_ENUM(Parameters::FLASH_MODE_ON) 266 CASE_APPEND_ENUM(Parameters::FLASH_MODE_TORCH) 267 CASE_APPEND_ENUM(Parameters::FLASH_MODE_RED_EYE) 268 CASE_APPEND_ENUM(Parameters::FLASH_MODE_INVALID) 269 default: result.append("UNKNOWN\n"); 270 } 271 272 result.append(" Focus mode: "); 273 switch (p.focusMode) { 274 CASE_APPEND_ENUM(Parameters::FOCUS_MODE_AUTO) 275 CASE_APPEND_ENUM(Parameters::FOCUS_MODE_MACRO) 276 CASE_APPEND_ENUM(Parameters::FOCUS_MODE_CONTINUOUS_VIDEO) 277 CASE_APPEND_ENUM(Parameters::FOCUS_MODE_CONTINUOUS_PICTURE) 278 CASE_APPEND_ENUM(Parameters::FOCUS_MODE_EDOF) 279 CASE_APPEND_ENUM(Parameters::FOCUS_MODE_INFINITY) 280 CASE_APPEND_ENUM(Parameters::FOCUS_MODE_FIXED) 281 CASE_APPEND_ENUM(Parameters::FOCUS_MODE_INVALID) 282 default: result.append("UNKNOWN\n"); 283 } 284 285 result.append(" Focus state: "); 286 switch (p.focusState) { 287 CASE_APPEND_ENUM(ANDROID_CONTROL_AF_STATE_INACTIVE) 288 CASE_APPEND_ENUM(ANDROID_CONTROL_AF_STATE_PASSIVE_SCAN) 289 CASE_APPEND_ENUM(ANDROID_CONTROL_AF_STATE_PASSIVE_FOCUSED) 290 CASE_APPEND_ENUM(ANDROID_CONTROL_AF_STATE_PASSIVE_UNFOCUSED) 291 CASE_APPEND_ENUM(ANDROID_CONTROL_AF_STATE_ACTIVE_SCAN) 292 CASE_APPEND_ENUM(ANDROID_CONTROL_AF_STATE_FOCUSED_LOCKED) 293 CASE_APPEND_ENUM(ANDROID_CONTROL_AF_STATE_NOT_FOCUSED_LOCKED) 294 default: result.append("UNKNOWN\n"); 295 } 296 297 result.append(" Focusing areas:\n"); 298 for (size_t i = 0; i < p.focusingAreas.size(); i++) { 299 result.appendFormat(" [ (%d, %d, %d, %d), weight %d ]\n", 300 p.focusingAreas[i].left, 301 p.focusingAreas[i].top, 302 p.focusingAreas[i].right, 303 p.focusingAreas[i].bottom, 304 p.focusingAreas[i].weight); 305 } 306 307 result.appendFormat(" Exposure compensation index: %d\n", 308 p.exposureCompensation); 309 310 result.appendFormat(" AE lock %s, AWB lock %s\n", 311 p.autoExposureLock ? "enabled" : "disabled", 312 p.autoWhiteBalanceLock ? "enabled" : "disabled" ); 313 314 result.appendFormat(" Metering areas:\n"); 315 for (size_t i = 0; i < p.meteringAreas.size(); i++) { 316 result.appendFormat(" [ (%d, %d, %d, %d), weight %d ]\n", 317 p.meteringAreas[i].left, 318 p.meteringAreas[i].top, 319 p.meteringAreas[i].right, 320 p.meteringAreas[i].bottom, 321 p.meteringAreas[i].weight); 322 } 323 324 result.appendFormat(" Zoom index: %d\n", p.zoom); 325 result.appendFormat(" Video size: %d x %d\n", p.videoWidth, 326 p.videoHeight); 327 328 result.appendFormat(" Recording hint is %s\n", 329 p.recordingHint ? "set" : "not set"); 330 331 result.appendFormat(" Video stabilization is %s\n", 332 p.videoStabilization ? "enabled" : "disabled"); 333 334 result.appendFormat(" Selected still capture FPS range: %d - %d\n", 335 p.fastInfo.bestStillCaptureFpsRange[0], 336 p.fastInfo.bestStillCaptureFpsRange[1]); 337 338 result.append(" Current streams:\n"); 339 result.appendFormat(" Preview stream ID: %d\n", 340 getPreviewStreamId()); 341 result.appendFormat(" Capture stream ID: %d\n", 342 getCaptureStreamId()); 343 result.appendFormat(" Recording stream ID: %d\n", 344 getRecordingStreamId()); 345 346 result.append(" Quirks for this camera:\n"); 347 bool haveQuirk = false; 348 if (p.quirks.triggerAfWithAuto) { 349 result.appendFormat(" triggerAfWithAuto\n"); 350 haveQuirk = true; 351 } 352 if (p.quirks.useZslFormat) { 353 result.appendFormat(" useZslFormat\n"); 354 haveQuirk = true; 355 } 356 if (p.quirks.meteringCropRegion) { 357 result.appendFormat(" meteringCropRegion\n"); 358 haveQuirk = true; 359 } 360 if (p.quirks.partialResults) { 361 result.appendFormat(" usePartialResult\n"); 362 haveQuirk = true; 363 } 364 if (!haveQuirk) { 365 result.appendFormat(" none\n"); 366 } 367 368 write(fd, result.string(), result.size()); 369 370 mStreamingProcessor->dump(fd, args); 371 372 mCaptureSequencer->dump(fd, args); 373 374 mFrameProcessor->dump(fd, args); 375 376 mZslProcessor->dump(fd, args); 377 378 return dumpDevice(fd, args); 379#undef CASE_APPEND_ENUM 380} 381 382// ICamera interface 383 384void Camera2Client::disconnect() { 385 ATRACE_CALL(); 386 Mutex::Autolock icl(mBinderSerializationLock); 387 388 // Allow both client and the media server to disconnect at all times 389 int callingPid = getCallingPid(); 390 if (callingPid != mClientPid && callingPid != mServicePid) return; 391 392 if (mDevice == 0) return; 393 394 ALOGV("Camera %d: Shutting down", mCameraId); 395 396 /** 397 * disconnect() cannot call any methods that might need to promote a 398 * wp<Camera2Client>, since disconnect can be called from the destructor, at 399 * which point all such promotions will fail. 400 */ 401 402 stopPreviewL(); 403 404 { 405 SharedParameters::Lock l(mParameters); 406 if (l.mParameters.state == Parameters::DISCONNECTED) return; 407 l.mParameters.state = Parameters::DISCONNECTED; 408 } 409 410 mStreamingProcessor->requestExit(); 411 mFrameProcessor->requestExit(); 412 mCaptureSequencer->requestExit(); 413 mJpegProcessor->requestExit(); 414 mZslProcessorThread->requestExit(); 415 mCallbackProcessor->requestExit(); 416 417 ALOGV("Camera %d: Waiting for threads", mCameraId); 418 419 mStreamingProcessor->join(); 420 mFrameProcessor->join(); 421 mCaptureSequencer->join(); 422 mJpegProcessor->join(); 423 mZslProcessorThread->join(); 424 mCallbackProcessor->join(); 425 426 ALOGV("Camera %d: Deleting streams", mCameraId); 427 428 mStreamingProcessor->deletePreviewStream(); 429 mStreamingProcessor->deleteRecordingStream(); 430 mJpegProcessor->deleteStream(); 431 mCallbackProcessor->deleteStream(); 432 mZslProcessor->deleteStream(); 433 434 ALOGV("Camera %d: Disconnecting device", mCameraId); 435 436 mDevice->disconnect(); 437 438 mDevice.clear(); 439 440 CameraService::Client::disconnect(); 441} 442 443status_t Camera2Client::connect(const sp<ICameraClient>& client) { 444 ATRACE_CALL(); 445 ALOGV("%s: E", __FUNCTION__); 446 Mutex::Autolock icl(mBinderSerializationLock); 447 448 if (mClientPid != 0 && getCallingPid() != mClientPid) { 449 ALOGE("%s: Camera %d: Connection attempt from pid %d; " 450 "current locked to pid %d", __FUNCTION__, 451 mCameraId, getCallingPid(), mClientPid); 452 return BAD_VALUE; 453 } 454 455 mClientPid = getCallingPid(); 456 457 mRemoteCallback = client; 458 mSharedCameraCallbacks = client; 459 460 return OK; 461} 462 463status_t Camera2Client::lock() { 464 ATRACE_CALL(); 465 ALOGV("%s: E", __FUNCTION__); 466 Mutex::Autolock icl(mBinderSerializationLock); 467 ALOGV("%s: Camera %d: Lock call from pid %d; current client pid %d", 468 __FUNCTION__, mCameraId, getCallingPid(), mClientPid); 469 470 if (mClientPid == 0) { 471 mClientPid = getCallingPid(); 472 return OK; 473 } 474 475 if (mClientPid != getCallingPid()) { 476 ALOGE("%s: Camera %d: Lock call from pid %d; currently locked to pid %d", 477 __FUNCTION__, mCameraId, getCallingPid(), mClientPid); 478 return EBUSY; 479 } 480 481 return OK; 482} 483 484status_t Camera2Client::unlock() { 485 ATRACE_CALL(); 486 ALOGV("%s: E", __FUNCTION__); 487 Mutex::Autolock icl(mBinderSerializationLock); 488 ALOGV("%s: Camera %d: Unlock call from pid %d; current client pid %d", 489 __FUNCTION__, mCameraId, getCallingPid(), mClientPid); 490 491 if (mClientPid == getCallingPid()) { 492 SharedParameters::Lock l(mParameters); 493 if (l.mParameters.state == Parameters::RECORD || 494 l.mParameters.state == Parameters::VIDEO_SNAPSHOT) { 495 ALOGD("Not allowed to unlock camera during recording."); 496 return INVALID_OPERATION; 497 } 498 mClientPid = 0; 499 mRemoteCallback.clear(); 500 mSharedCameraCallbacks.clear(); 501 return OK; 502 } 503 504 ALOGE("%s: Camera %d: Unlock call from pid %d; currently locked to pid %d", 505 __FUNCTION__, mCameraId, getCallingPid(), mClientPid); 506 return EBUSY; 507} 508 509status_t Camera2Client::setPreviewTarget( 510 const sp<IGraphicBufferProducer>& bufferProducer) { 511 ATRACE_CALL(); 512 ALOGV("%s: E", __FUNCTION__); 513 Mutex::Autolock icl(mBinderSerializationLock); 514 status_t res; 515 if ( (res = checkPid(__FUNCTION__) ) != OK) return res; 516 517 sp<IBinder> binder; 518 sp<ANativeWindow> window; 519 if (bufferProducer != 0) { 520 binder = bufferProducer->asBinder(); 521 // Using controlledByApp flag to ensure that the buffer queue remains in 522 // async mode for the old camera API, where many applications depend 523 // on that behavior. 524 window = new Surface(bufferProducer, /*controlledByApp*/ true); 525 } 526 return setPreviewWindowL(binder, window); 527} 528 529status_t Camera2Client::setPreviewWindowL(const sp<IBinder>& binder, 530 sp<ANativeWindow> window) { 531 ATRACE_CALL(); 532 status_t res; 533 534 if (binder == mPreviewSurface) { 535 ALOGV("%s: Camera %d: New window is same as old window", 536 __FUNCTION__, mCameraId); 537 return NO_ERROR; 538 } 539 540 Parameters::State state; 541 { 542 SharedParameters::Lock l(mParameters); 543 state = l.mParameters.state; 544 } 545 switch (state) { 546 case Parameters::DISCONNECTED: 547 case Parameters::RECORD: 548 case Parameters::STILL_CAPTURE: 549 case Parameters::VIDEO_SNAPSHOT: 550 ALOGE("%s: Camera %d: Cannot set preview display while in state %s", 551 __FUNCTION__, mCameraId, 552 Parameters::getStateName(state)); 553 return INVALID_OPERATION; 554 case Parameters::STOPPED: 555 case Parameters::WAITING_FOR_PREVIEW_WINDOW: 556 // OK 557 break; 558 case Parameters::PREVIEW: 559 // Already running preview - need to stop and create a new stream 560 res = stopStream(); 561 if (res != OK) { 562 ALOGE("%s: Unable to stop preview to swap windows: %s (%d)", 563 __FUNCTION__, strerror(-res), res); 564 return res; 565 } 566 state = Parameters::WAITING_FOR_PREVIEW_WINDOW; 567 break; 568 } 569 570 mPreviewSurface = binder; 571 res = mStreamingProcessor->setPreviewWindow(window); 572 if (res != OK) { 573 ALOGE("%s: Unable to set new preview window: %s (%d)", 574 __FUNCTION__, strerror(-res), res); 575 return res; 576 } 577 578 if (state == Parameters::WAITING_FOR_PREVIEW_WINDOW) { 579 SharedParameters::Lock l(mParameters); 580 l.mParameters.state = state; 581 return startPreviewL(l.mParameters, false); 582 } 583 584 return OK; 585} 586 587void Camera2Client::setPreviewCallbackFlag(int flag) { 588 ATRACE_CALL(); 589 ALOGV("%s: Camera %d: Flag 0x%x", __FUNCTION__, mCameraId, flag); 590 Mutex::Autolock icl(mBinderSerializationLock); 591 592 if ( checkPid(__FUNCTION__) != OK) return; 593 594 SharedParameters::Lock l(mParameters); 595 setPreviewCallbackFlagL(l.mParameters, flag); 596} 597 598void Camera2Client::setPreviewCallbackFlagL(Parameters ¶ms, int flag) { 599 status_t res = OK; 600 601 switch(params.state) { 602 case Parameters::STOPPED: 603 case Parameters::WAITING_FOR_PREVIEW_WINDOW: 604 case Parameters::PREVIEW: 605 case Parameters::STILL_CAPTURE: 606 // OK 607 break; 608 default: 609 if (flag & CAMERA_FRAME_CALLBACK_FLAG_ENABLE_MASK) { 610 ALOGE("%s: Camera %d: Can't use preview callbacks " 611 "in state %d", __FUNCTION__, mCameraId, params.state); 612 return; 613 } 614 } 615 616 if (flag & CAMERA_FRAME_CALLBACK_FLAG_ONE_SHOT_MASK) { 617 ALOGV("%s: setting oneshot", __FUNCTION__); 618 params.previewCallbackOneShot = true; 619 } 620 if (params.previewCallbackFlags != (uint32_t)flag) { 621 622 if (params.previewCallbackSurface && flag != CAMERA_FRAME_CALLBACK_FLAG_NOOP) { 623 // Disable any existing preview callback window when enabling 624 // preview callback flags 625 res = mCallbackProcessor->setCallbackWindow(NULL); 626 if (res != OK) { 627 ALOGE("%s: Camera %d: Unable to clear preview callback surface:" 628 " %s (%d)", __FUNCTION__, mCameraId, strerror(-res), res); 629 return; 630 } 631 params.previewCallbackSurface = false; 632 } 633 634 params.previewCallbackFlags = flag; 635 636 if (params.state == Parameters::PREVIEW) { 637 res = startPreviewL(params, true); 638 if (res != OK) { 639 ALOGE("%s: Camera %d: Unable to refresh request in state %s", 640 __FUNCTION__, mCameraId, 641 Parameters::getStateName(params.state)); 642 } 643 } 644 } 645} 646 647status_t Camera2Client::setPreviewCallbackTarget( 648 const sp<IGraphicBufferProducer>& callbackProducer) { 649 ATRACE_CALL(); 650 ALOGV("%s: E", __FUNCTION__); 651 Mutex::Autolock icl(mBinderSerializationLock); 652 status_t res; 653 if ( (res = checkPid(__FUNCTION__) ) != OK) return res; 654 655 sp<ANativeWindow> window; 656 if (callbackProducer != 0) { 657 window = new Surface(callbackProducer); 658 } 659 660 res = mCallbackProcessor->setCallbackWindow(window); 661 if (res != OK) { 662 ALOGE("%s: Camera %d: Unable to set preview callback surface: %s (%d)", 663 __FUNCTION__, mCameraId, strerror(-res), res); 664 return res; 665 } 666 667 SharedParameters::Lock l(mParameters); 668 669 if (window != NULL) { 670 // Disable traditional callbacks when a valid callback target is given 671 l.mParameters.previewCallbackFlags = CAMERA_FRAME_CALLBACK_FLAG_NOOP; 672 l.mParameters.previewCallbackOneShot = false; 673 l.mParameters.previewCallbackSurface = true; 674 } else { 675 // Disable callback target if given a NULL interface. 676 l.mParameters.previewCallbackSurface = false; 677 } 678 679 switch(l.mParameters.state) { 680 case Parameters::PREVIEW: 681 res = startPreviewL(l.mParameters, true); 682 break; 683 case Parameters::RECORD: 684 case Parameters::VIDEO_SNAPSHOT: 685 res = startRecordingL(l.mParameters, true); 686 break; 687 default: 688 break; 689 } 690 if (res != OK) { 691 ALOGE("%s: Camera %d: Unable to refresh request in state %s", 692 __FUNCTION__, mCameraId, 693 Parameters::getStateName(l.mParameters.state)); 694 } 695 696 return OK; 697} 698 699 700status_t Camera2Client::startPreview() { 701 ATRACE_CALL(); 702 ALOGV("%s: E", __FUNCTION__); 703 Mutex::Autolock icl(mBinderSerializationLock); 704 status_t res; 705 if ( (res = checkPid(__FUNCTION__) ) != OK) return res; 706 SharedParameters::Lock l(mParameters); 707 return startPreviewL(l.mParameters, false); 708} 709 710status_t Camera2Client::startPreviewL(Parameters ¶ms, bool restart) { 711 ATRACE_CALL(); 712 status_t res; 713 714 ALOGV("%s: state == %d, restart = %d", __FUNCTION__, params.state, restart); 715 716 if ( (params.state == Parameters::PREVIEW || 717 params.state == Parameters::RECORD || 718 params.state == Parameters::VIDEO_SNAPSHOT) 719 && !restart) { 720 // Succeed attempt to re-enter a streaming state 721 ALOGI("%s: Camera %d: Preview already active, ignoring restart", 722 __FUNCTION__, mCameraId); 723 return OK; 724 } 725 if (params.state > Parameters::PREVIEW && !restart) { 726 ALOGE("%s: Can't start preview in state %s", 727 __FUNCTION__, 728 Parameters::getStateName(params.state)); 729 return INVALID_OPERATION; 730 } 731 732 if (!mStreamingProcessor->haveValidPreviewWindow()) { 733 params.state = Parameters::WAITING_FOR_PREVIEW_WINDOW; 734 return OK; 735 } 736 params.state = Parameters::STOPPED; 737 int lastPreviewStreamId = mStreamingProcessor->getPreviewStreamId(); 738 739 res = mStreamingProcessor->updatePreviewStream(params); 740 if (res != OK) { 741 ALOGE("%s: Camera %d: Unable to update preview stream: %s (%d)", 742 __FUNCTION__, mCameraId, strerror(-res), res); 743 return res; 744 } 745 746 bool previewStreamChanged = mStreamingProcessor->getPreviewStreamId() != lastPreviewStreamId; 747 748 // We could wait to create the JPEG output stream until first actual use 749 // (first takePicture call). However, this would substantially increase the 750 // first capture latency on HAL3 devices, and potentially on some HAL2 751 // devices. So create it unconditionally at preview start. As a drawback, 752 // this increases gralloc memory consumption for applications that don't 753 // ever take a picture. 754 // TODO: Find a better compromise, though this likely would involve HAL 755 // changes. 756 int lastJpegStreamId = mJpegProcessor->getStreamId(); 757 res = updateProcessorStream(mJpegProcessor, params); 758 if (res != OK) { 759 ALOGE("%s: Camera %d: Can't pre-configure still image " 760 "stream: %s (%d)", 761 __FUNCTION__, mCameraId, strerror(-res), res); 762 return res; 763 } 764 bool jpegStreamChanged = mJpegProcessor->getStreamId() != lastJpegStreamId; 765 766 Vector<int32_t> outputStreams; 767 bool callbacksEnabled = (params.previewCallbackFlags & 768 CAMERA_FRAME_CALLBACK_FLAG_ENABLE_MASK) || 769 params.previewCallbackSurface; 770 771 if (callbacksEnabled) { 772 // Can't have recording stream hanging around when enabling callbacks, 773 // since it exceeds the max stream count on some devices. 774 if (mStreamingProcessor->getRecordingStreamId() != NO_STREAM) { 775 ALOGV("%s: Camera %d: Clearing out recording stream before " 776 "creating callback stream", __FUNCTION__, mCameraId); 777 res = mStreamingProcessor->stopStream(); 778 if (res != OK) { 779 ALOGE("%s: Camera %d: Can't stop streaming to delete " 780 "recording stream", __FUNCTION__, mCameraId); 781 return res; 782 } 783 res = mStreamingProcessor->deleteRecordingStream(); 784 if (res != OK) { 785 ALOGE("%s: Camera %d: Unable to delete recording stream before " 786 "enabling callbacks: %s (%d)", __FUNCTION__, mCameraId, 787 strerror(-res), res); 788 return res; 789 } 790 } 791 792 res = mCallbackProcessor->updateStream(params); 793 if (res != OK) { 794 ALOGE("%s: Camera %d: Unable to update callback stream: %s (%d)", 795 __FUNCTION__, mCameraId, strerror(-res), res); 796 return res; 797 } 798 outputStreams.push(getCallbackStreamId()); 799 } else if (previewStreamChanged && mCallbackProcessor->getStreamId() != NO_STREAM) { 800 /** 801 * Delete the unused callback stream when preview stream is changed and 802 * preview is not enabled. Don't need stop preview stream as preview is in 803 * STOPPED state now. 804 */ 805 ALOGV("%s: Camera %d: Delete unused preview callback stream.", __FUNCTION__, mCameraId); 806 res = mCallbackProcessor->deleteStream(); 807 if (res != OK) { 808 ALOGE("%s: Camera %d: Unable to delete callback stream %s (%d)", 809 __FUNCTION__, mCameraId, strerror(-res), res); 810 return res; 811 } 812 } 813 if (params.zslMode && !params.recordingHint) { 814 res = updateProcessorStream(mZslProcessor, params); 815 if (res != OK) { 816 ALOGE("%s: Camera %d: Unable to update ZSL stream: %s (%d)", 817 __FUNCTION__, mCameraId, strerror(-res), res); 818 return res; 819 } 820 821 if (jpegStreamChanged) { 822 ALOGV("%s: Camera %d: Clear ZSL buffer queue when Jpeg size is changed", 823 __FUNCTION__, mCameraId); 824 mZslProcessor->clearZslQueue(); 825 } 826 outputStreams.push(getZslStreamId()); 827 } else { 828 mZslProcessor->deleteStream(); 829 } 830 831 outputStreams.push(getPreviewStreamId()); 832 833 if (!params.recordingHint) { 834 if (!restart) { 835 res = mStreamingProcessor->updatePreviewRequest(params); 836 if (res != OK) { 837 ALOGE("%s: Camera %d: Can't set up preview request: " 838 "%s (%d)", __FUNCTION__, mCameraId, 839 strerror(-res), res); 840 return res; 841 } 842 } 843 res = mStreamingProcessor->startStream(StreamingProcessor::PREVIEW, 844 outputStreams); 845 } else { 846 if (!restart) { 847 res = mStreamingProcessor->updateRecordingRequest(params); 848 if (res != OK) { 849 ALOGE("%s: Camera %d: Can't set up preview request with " 850 "record hint: %s (%d)", __FUNCTION__, mCameraId, 851 strerror(-res), res); 852 return res; 853 } 854 } 855 res = mStreamingProcessor->startStream(StreamingProcessor::RECORD, 856 outputStreams); 857 } 858 if (res != OK) { 859 ALOGE("%s: Camera %d: Unable to start streaming preview: %s (%d)", 860 __FUNCTION__, mCameraId, strerror(-res), res); 861 return res; 862 } 863 864 params.state = Parameters::PREVIEW; 865 return OK; 866} 867 868void Camera2Client::stopPreview() { 869 ATRACE_CALL(); 870 ALOGV("%s: E", __FUNCTION__); 871 Mutex::Autolock icl(mBinderSerializationLock); 872 status_t res; 873 if ( (res = checkPid(__FUNCTION__) ) != OK) return; 874 stopPreviewL(); 875} 876 877void Camera2Client::stopPreviewL() { 878 ATRACE_CALL(); 879 status_t res; 880 const nsecs_t kStopCaptureTimeout = 3000000000LL; // 3 seconds 881 Parameters::State state; 882 { 883 SharedParameters::Lock l(mParameters); 884 state = l.mParameters.state; 885 } 886 887 switch (state) { 888 case Parameters::DISCONNECTED: 889 // Nothing to do. 890 break; 891 case Parameters::STOPPED: 892 case Parameters::VIDEO_SNAPSHOT: 893 case Parameters::STILL_CAPTURE: 894 mCaptureSequencer->waitUntilIdle(kStopCaptureTimeout); 895 // no break 896 case Parameters::RECORD: 897 case Parameters::PREVIEW: 898 syncWithDevice(); 899 res = stopStream(); 900 if (res != OK) { 901 ALOGE("%s: Camera %d: Can't stop streaming: %s (%d)", 902 __FUNCTION__, mCameraId, strerror(-res), res); 903 } 904 res = mDevice->waitUntilDrained(); 905 if (res != OK) { 906 ALOGE("%s: Camera %d: Waiting to stop streaming failed: %s (%d)", 907 __FUNCTION__, mCameraId, strerror(-res), res); 908 } 909 // Clean up recording stream 910 res = mStreamingProcessor->deleteRecordingStream(); 911 if (res != OK) { 912 ALOGE("%s: Camera %d: Unable to delete recording stream before " 913 "stop preview: %s (%d)", 914 __FUNCTION__, mCameraId, strerror(-res), res); 915 } 916 // no break 917 case Parameters::WAITING_FOR_PREVIEW_WINDOW: { 918 SharedParameters::Lock l(mParameters); 919 l.mParameters.state = Parameters::STOPPED; 920 commandStopFaceDetectionL(l.mParameters); 921 break; 922 } 923 default: 924 ALOGE("%s: Camera %d: Unknown state %d", __FUNCTION__, mCameraId, 925 state); 926 } 927} 928 929bool Camera2Client::previewEnabled() { 930 ATRACE_CALL(); 931 Mutex::Autolock icl(mBinderSerializationLock); 932 status_t res; 933 if ( (res = checkPid(__FUNCTION__) ) != OK) return false; 934 935 SharedParameters::Lock l(mParameters); 936 return l.mParameters.state == Parameters::PREVIEW; 937} 938 939status_t Camera2Client::storeMetaDataInBuffers(bool enabled) { 940 ATRACE_CALL(); 941 Mutex::Autolock icl(mBinderSerializationLock); 942 status_t res; 943 if ( (res = checkPid(__FUNCTION__) ) != OK) return res; 944 945 SharedParameters::Lock l(mParameters); 946 switch (l.mParameters.state) { 947 case Parameters::RECORD: 948 case Parameters::VIDEO_SNAPSHOT: 949 ALOGE("%s: Camera %d: Can't be called in state %s", 950 __FUNCTION__, mCameraId, 951 Parameters::getStateName(l.mParameters.state)); 952 return INVALID_OPERATION; 953 default: 954 // OK 955 break; 956 } 957 958 l.mParameters.storeMetadataInBuffers = enabled; 959 960 return OK; 961} 962 963status_t Camera2Client::startRecording() { 964 ATRACE_CALL(); 965 ALOGV("%s: E", __FUNCTION__); 966 Mutex::Autolock icl(mBinderSerializationLock); 967 status_t res; 968 if ( (res = checkPid(__FUNCTION__) ) != OK) return res; 969 SharedParameters::Lock l(mParameters); 970 971 return startRecordingL(l.mParameters, false); 972} 973 974status_t Camera2Client::startRecordingL(Parameters ¶ms, bool restart) { 975 status_t res; 976 977 ALOGV("%s: state == %d, restart = %d", __FUNCTION__, params.state, restart); 978 979 switch (params.state) { 980 case Parameters::STOPPED: 981 res = startPreviewL(params, false); 982 if (res != OK) return res; 983 break; 984 case Parameters::PREVIEW: 985 // Ready to go 986 break; 987 case Parameters::RECORD: 988 case Parameters::VIDEO_SNAPSHOT: 989 // OK to call this when recording is already on, just skip unless 990 // we're looking to restart 991 if (!restart) return OK; 992 break; 993 default: 994 ALOGE("%s: Camera %d: Can't start recording in state %s", 995 __FUNCTION__, mCameraId, 996 Parameters::getStateName(params.state)); 997 return INVALID_OPERATION; 998 }; 999 1000 if (!params.storeMetadataInBuffers) { 1001 ALOGE("%s: Camera %d: Recording only supported in metadata mode, but " 1002 "non-metadata recording mode requested!", __FUNCTION__, 1003 mCameraId); 1004 return INVALID_OPERATION; 1005 } 1006 1007 if (!restart) { 1008 mCameraService->playSound(CameraService::SOUND_RECORDING); 1009 mStreamingProcessor->updateRecordingRequest(params); 1010 if (res != OK) { 1011 ALOGE("%s: Camera %d: Unable to update recording request: %s (%d)", 1012 __FUNCTION__, mCameraId, strerror(-res), res); 1013 return res; 1014 } 1015 } 1016 1017 // Not all devices can support a preview callback stream and a recording 1018 // stream at the same time, so assume none of them can. 1019 if (mCallbackProcessor->getStreamId() != NO_STREAM) { 1020 ALOGV("%s: Camera %d: Clearing out callback stream before " 1021 "creating recording stream", __FUNCTION__, mCameraId); 1022 res = mStreamingProcessor->stopStream(); 1023 if (res != OK) { 1024 ALOGE("%s: Camera %d: Can't stop streaming to delete callback stream", 1025 __FUNCTION__, mCameraId); 1026 return res; 1027 } 1028 res = mCallbackProcessor->deleteStream(); 1029 if (res != OK) { 1030 ALOGE("%s: Camera %d: Unable to delete callback stream before " 1031 "record: %s (%d)", __FUNCTION__, mCameraId, 1032 strerror(-res), res); 1033 return res; 1034 } 1035 } 1036 // Disable callbacks if they're enabled; can't record and use callbacks, 1037 // and we can't fail record start without stagefright asserting. 1038 params.previewCallbackFlags = 0; 1039 1040 res = updateProcessorStream< 1041 StreamingProcessor, 1042 &StreamingProcessor::updateRecordingStream>(mStreamingProcessor, 1043 params); 1044 if (res != OK) { 1045 ALOGE("%s: Camera %d: Unable to update recording stream: %s (%d)", 1046 __FUNCTION__, mCameraId, strerror(-res), res); 1047 return res; 1048 } 1049 1050 Vector<int32_t> outputStreams; 1051 outputStreams.push(getPreviewStreamId()); 1052 outputStreams.push(getRecordingStreamId()); 1053 1054 res = mStreamingProcessor->startStream(StreamingProcessor::RECORD, 1055 outputStreams); 1056 if (res != OK) { 1057 ALOGE("%s: Camera %d: Unable to start recording stream: %s (%d)", 1058 __FUNCTION__, mCameraId, strerror(-res), res); 1059 return res; 1060 } 1061 1062 if (params.state < Parameters::RECORD) { 1063 params.state = Parameters::RECORD; 1064 } 1065 1066 return OK; 1067} 1068 1069void Camera2Client::stopRecording() { 1070 ATRACE_CALL(); 1071 ALOGV("%s: E", __FUNCTION__); 1072 Mutex::Autolock icl(mBinderSerializationLock); 1073 SharedParameters::Lock l(mParameters); 1074 1075 status_t res; 1076 if ( (res = checkPid(__FUNCTION__) ) != OK) return; 1077 1078 switch (l.mParameters.state) { 1079 case Parameters::RECORD: 1080 // OK to stop 1081 break; 1082 case Parameters::STOPPED: 1083 case Parameters::PREVIEW: 1084 case Parameters::STILL_CAPTURE: 1085 case Parameters::VIDEO_SNAPSHOT: 1086 default: 1087 ALOGE("%s: Camera %d: Can't stop recording in state %s", 1088 __FUNCTION__, mCameraId, 1089 Parameters::getStateName(l.mParameters.state)); 1090 return; 1091 }; 1092 1093 mCameraService->playSound(CameraService::SOUND_RECORDING); 1094 1095 res = startPreviewL(l.mParameters, true); 1096 if (res != OK) { 1097 ALOGE("%s: Camera %d: Unable to return to preview", 1098 __FUNCTION__, mCameraId); 1099 } 1100} 1101 1102bool Camera2Client::recordingEnabled() { 1103 ATRACE_CALL(); 1104 Mutex::Autolock icl(mBinderSerializationLock); 1105 1106 if ( checkPid(__FUNCTION__) != OK) return false; 1107 1108 return recordingEnabledL(); 1109} 1110 1111bool Camera2Client::recordingEnabledL() { 1112 ATRACE_CALL(); 1113 SharedParameters::Lock l(mParameters); 1114 1115 return (l.mParameters.state == Parameters::RECORD 1116 || l.mParameters.state == Parameters::VIDEO_SNAPSHOT); 1117} 1118 1119void Camera2Client::releaseRecordingFrame(const sp<IMemory>& mem) { 1120 ATRACE_CALL(); 1121 Mutex::Autolock icl(mBinderSerializationLock); 1122 if ( checkPid(__FUNCTION__) != OK) return; 1123 1124 mStreamingProcessor->releaseRecordingFrame(mem); 1125} 1126 1127status_t Camera2Client::autoFocus() { 1128 ATRACE_CALL(); 1129 Mutex::Autolock icl(mBinderSerializationLock); 1130 ALOGV("%s: Camera %d", __FUNCTION__, mCameraId); 1131 status_t res; 1132 if ( (res = checkPid(__FUNCTION__) ) != OK) return res; 1133 1134 int triggerId; 1135 bool notifyImmediately = false; 1136 bool notifySuccess = false; 1137 { 1138 SharedParameters::Lock l(mParameters); 1139 if (l.mParameters.state < Parameters::PREVIEW) { 1140 return INVALID_OPERATION; 1141 } 1142 1143 /** 1144 * If the camera does not support auto-focus, it is a no-op and 1145 * onAutoFocus(boolean, Camera) callback will be called immediately 1146 * with a fake value of success set to true. 1147 * 1148 * Similarly, if focus mode is set to INFINITY, there's no reason to 1149 * bother the HAL. 1150 */ 1151 if (l.mParameters.focusMode == Parameters::FOCUS_MODE_FIXED || 1152 l.mParameters.focusMode == Parameters::FOCUS_MODE_INFINITY) { 1153 notifyImmediately = true; 1154 notifySuccess = true; 1155 } 1156 /** 1157 * If we're in CAF mode, and AF has already been locked, just fire back 1158 * the callback right away; the HAL would not send a notification since 1159 * no state change would happen on a AF trigger. 1160 */ 1161 if ( (l.mParameters.focusMode == Parameters::FOCUS_MODE_CONTINUOUS_PICTURE || 1162 l.mParameters.focusMode == Parameters::FOCUS_MODE_CONTINUOUS_VIDEO) && 1163 l.mParameters.focusState == ANDROID_CONTROL_AF_STATE_FOCUSED_LOCKED ) { 1164 notifyImmediately = true; 1165 notifySuccess = true; 1166 } 1167 /** 1168 * Send immediate notification back to client 1169 */ 1170 if (notifyImmediately) { 1171 SharedCameraCallbacks::Lock l(mSharedCameraCallbacks); 1172 if (l.mRemoteCallback != 0) { 1173 l.mRemoteCallback->notifyCallback(CAMERA_MSG_FOCUS, 1174 notifySuccess ? 1 : 0, 0); 1175 } 1176 return OK; 1177 } 1178 /** 1179 * Handle quirk mode for AF in scene modes 1180 */ 1181 if (l.mParameters.quirks.triggerAfWithAuto && 1182 l.mParameters.sceneMode != ANDROID_CONTROL_SCENE_MODE_DISABLED && 1183 l.mParameters.focusMode != Parameters::FOCUS_MODE_AUTO && 1184 !l.mParameters.focusingAreas[0].isEmpty()) { 1185 ALOGV("%s: Quirk: Switching from focusMode %d to AUTO", 1186 __FUNCTION__, l.mParameters.focusMode); 1187 l.mParameters.shadowFocusMode = l.mParameters.focusMode; 1188 l.mParameters.focusMode = Parameters::FOCUS_MODE_AUTO; 1189 updateRequests(l.mParameters); 1190 } 1191 1192 l.mParameters.currentAfTriggerId = ++l.mParameters.afTriggerCounter; 1193 triggerId = l.mParameters.currentAfTriggerId; 1194 } 1195 ATRACE_ASYNC_BEGIN(kAutofocusLabel, triggerId); 1196 1197 syncWithDevice(); 1198 1199 mDevice->triggerAutofocus(triggerId); 1200 1201 return OK; 1202} 1203 1204status_t Camera2Client::cancelAutoFocus() { 1205 ATRACE_CALL(); 1206 Mutex::Autolock icl(mBinderSerializationLock); 1207 ALOGV("%s: Camera %d", __FUNCTION__, mCameraId); 1208 status_t res; 1209 if ( (res = checkPid(__FUNCTION__) ) != OK) return res; 1210 1211 int triggerId; 1212 { 1213 SharedParameters::Lock l(mParameters); 1214 // Canceling does nothing in FIXED or INFINITY modes 1215 if (l.mParameters.focusMode == Parameters::FOCUS_MODE_FIXED || 1216 l.mParameters.focusMode == Parameters::FOCUS_MODE_INFINITY) { 1217 return OK; 1218 } 1219 1220 // An active AF trigger is canceled 1221 if (l.mParameters.afTriggerCounter == l.mParameters.currentAfTriggerId) { 1222 ATRACE_ASYNC_END(kAutofocusLabel, l.mParameters.currentAfTriggerId); 1223 } 1224 1225 triggerId = ++l.mParameters.afTriggerCounter; 1226 1227 // When using triggerAfWithAuto quirk, may need to reset focus mode to 1228 // the real state at this point. No need to cancel explicitly if 1229 // changing the AF mode. 1230 if (l.mParameters.shadowFocusMode != Parameters::FOCUS_MODE_INVALID) { 1231 ALOGV("%s: Quirk: Restoring focus mode to %d", __FUNCTION__, 1232 l.mParameters.shadowFocusMode); 1233 l.mParameters.focusMode = l.mParameters.shadowFocusMode; 1234 l.mParameters.shadowFocusMode = Parameters::FOCUS_MODE_INVALID; 1235 updateRequests(l.mParameters); 1236 1237 return OK; 1238 } 1239 } 1240 syncWithDevice(); 1241 1242 mDevice->triggerCancelAutofocus(triggerId); 1243 1244 return OK; 1245} 1246 1247status_t Camera2Client::takePicture(int msgType) { 1248 ATRACE_CALL(); 1249 Mutex::Autolock icl(mBinderSerializationLock); 1250 status_t res; 1251 if ( (res = checkPid(__FUNCTION__) ) != OK) return res; 1252 1253 int takePictureCounter; 1254 { 1255 SharedParameters::Lock l(mParameters); 1256 switch (l.mParameters.state) { 1257 case Parameters::DISCONNECTED: 1258 case Parameters::STOPPED: 1259 case Parameters::WAITING_FOR_PREVIEW_WINDOW: 1260 ALOGE("%s: Camera %d: Cannot take picture without preview enabled", 1261 __FUNCTION__, mCameraId); 1262 return INVALID_OPERATION; 1263 case Parameters::PREVIEW: 1264 // Good to go for takePicture 1265 res = commandStopFaceDetectionL(l.mParameters); 1266 if (res != OK) { 1267 ALOGE("%s: Camera %d: Unable to stop face detection for still capture", 1268 __FUNCTION__, mCameraId); 1269 return res; 1270 } 1271 l.mParameters.state = Parameters::STILL_CAPTURE; 1272 break; 1273 case Parameters::RECORD: 1274 // Good to go for video snapshot 1275 l.mParameters.state = Parameters::VIDEO_SNAPSHOT; 1276 break; 1277 case Parameters::STILL_CAPTURE: 1278 case Parameters::VIDEO_SNAPSHOT: 1279 ALOGE("%s: Camera %d: Already taking a picture", 1280 __FUNCTION__, mCameraId); 1281 return INVALID_OPERATION; 1282 } 1283 1284 ALOGV("%s: Camera %d: Starting picture capture", __FUNCTION__, mCameraId); 1285 1286 int lastJpegStreamId = mJpegProcessor->getStreamId(); 1287 res = updateProcessorStream(mJpegProcessor, l.mParameters); 1288 if (res != OK) { 1289 ALOGE("%s: Camera %d: Can't set up still image stream: %s (%d)", 1290 __FUNCTION__, mCameraId, strerror(-res), res); 1291 return res; 1292 } 1293 takePictureCounter = ++l.mParameters.takePictureCounter; 1294 1295 // Clear ZSL buffer queue when Jpeg size is changed. 1296 bool jpegStreamChanged = mJpegProcessor->getStreamId() != lastJpegStreamId; 1297 if (l.mParameters.zslMode && jpegStreamChanged) { 1298 ALOGV("%s: Camera %d: Clear ZSL buffer queue when Jpeg size is changed", 1299 __FUNCTION__, mCameraId); 1300 mZslProcessor->clearZslQueue(); 1301 } 1302 } 1303 1304 ATRACE_ASYNC_BEGIN(kTakepictureLabel, takePictureCounter); 1305 1306 // Need HAL to have correct settings before (possibly) triggering precapture 1307 syncWithDevice(); 1308 1309 res = mCaptureSequencer->startCapture(msgType); 1310 if (res != OK) { 1311 ALOGE("%s: Camera %d: Unable to start capture: %s (%d)", 1312 __FUNCTION__, mCameraId, strerror(-res), res); 1313 } 1314 1315 return res; 1316} 1317 1318status_t Camera2Client::setParameters(const String8& params) { 1319 ATRACE_CALL(); 1320 ALOGV("%s: Camera %d", __FUNCTION__, mCameraId); 1321 Mutex::Autolock icl(mBinderSerializationLock); 1322 status_t res; 1323 if ( (res = checkPid(__FUNCTION__) ) != OK) return res; 1324 1325 SharedParameters::Lock l(mParameters); 1326 1327 res = l.mParameters.set(params); 1328 if (res != OK) return res; 1329 1330 res = updateRequests(l.mParameters); 1331 1332 return res; 1333} 1334 1335String8 Camera2Client::getParameters() const { 1336 ATRACE_CALL(); 1337 ALOGV("%s: Camera %d", __FUNCTION__, mCameraId); 1338 Mutex::Autolock icl(mBinderSerializationLock); 1339 if ( checkPid(__FUNCTION__) != OK) return String8(); 1340 1341 SharedParameters::ReadLock l(mParameters); 1342 1343 return l.mParameters.get(); 1344} 1345 1346status_t Camera2Client::sendCommand(int32_t cmd, int32_t arg1, int32_t arg2) { 1347 ATRACE_CALL(); 1348 Mutex::Autolock icl(mBinderSerializationLock); 1349 status_t res; 1350 if ( (res = checkPid(__FUNCTION__) ) != OK) return res; 1351 1352 ALOGV("%s: Camera %d: Command %d (%d, %d)", __FUNCTION__, mCameraId, 1353 cmd, arg1, arg2); 1354 1355 switch (cmd) { 1356 case CAMERA_CMD_START_SMOOTH_ZOOM: 1357 return commandStartSmoothZoomL(); 1358 case CAMERA_CMD_STOP_SMOOTH_ZOOM: 1359 return commandStopSmoothZoomL(); 1360 case CAMERA_CMD_SET_DISPLAY_ORIENTATION: 1361 return commandSetDisplayOrientationL(arg1); 1362 case CAMERA_CMD_ENABLE_SHUTTER_SOUND: 1363 return commandEnableShutterSoundL(arg1 == 1); 1364 case CAMERA_CMD_PLAY_RECORDING_SOUND: 1365 return commandPlayRecordingSoundL(); 1366 case CAMERA_CMD_START_FACE_DETECTION: 1367 return commandStartFaceDetectionL(arg1); 1368 case CAMERA_CMD_STOP_FACE_DETECTION: { 1369 SharedParameters::Lock l(mParameters); 1370 return commandStopFaceDetectionL(l.mParameters); 1371 } 1372 case CAMERA_CMD_ENABLE_FOCUS_MOVE_MSG: 1373 return commandEnableFocusMoveMsgL(arg1 == 1); 1374 case CAMERA_CMD_PING: 1375 return commandPingL(); 1376 case CAMERA_CMD_SET_VIDEO_BUFFER_COUNT: 1377 return commandSetVideoBufferCountL(arg1); 1378 default: 1379 ALOGE("%s: Unknown command %d (arguments %d, %d)", 1380 __FUNCTION__, cmd, arg1, arg2); 1381 return BAD_VALUE; 1382 } 1383} 1384 1385status_t Camera2Client::commandStartSmoothZoomL() { 1386 ALOGE("%s: Unimplemented!", __FUNCTION__); 1387 return OK; 1388} 1389 1390status_t Camera2Client::commandStopSmoothZoomL() { 1391 ALOGE("%s: Unimplemented!", __FUNCTION__); 1392 return OK; 1393} 1394 1395status_t Camera2Client::commandSetDisplayOrientationL(int degrees) { 1396 int transform = Parameters::degToTransform(degrees, 1397 mCameraFacing == CAMERA_FACING_FRONT); 1398 if (transform == -1) { 1399 ALOGE("%s: Camera %d: Error setting %d as display orientation value", 1400 __FUNCTION__, mCameraId, degrees); 1401 return BAD_VALUE; 1402 } 1403 SharedParameters::Lock l(mParameters); 1404 if (transform != l.mParameters.previewTransform && 1405 getPreviewStreamId() != NO_STREAM) { 1406 mDevice->setStreamTransform(getPreviewStreamId(), transform); 1407 } 1408 l.mParameters.previewTransform = transform; 1409 return OK; 1410} 1411 1412status_t Camera2Client::commandEnableShutterSoundL(bool enable) { 1413 SharedParameters::Lock l(mParameters); 1414 if (enable) { 1415 l.mParameters.playShutterSound = true; 1416 return OK; 1417 } 1418 1419 // Disabling shutter sound may not be allowed. In that case only 1420 // allow the mediaserver process to disable the sound. 1421 char value[PROPERTY_VALUE_MAX]; 1422 property_get("ro.camera.sound.forced", value, "0"); 1423 if (strncmp(value, "0", 2) != 0) { 1424 // Disabling shutter sound is not allowed. Deny if the current 1425 // process is not mediaserver. 1426 if (getCallingPid() != getpid()) { 1427 ALOGE("Failed to disable shutter sound. Permission denied (pid %d)", 1428 getCallingPid()); 1429 return PERMISSION_DENIED; 1430 } 1431 } 1432 1433 l.mParameters.playShutterSound = false; 1434 return OK; 1435} 1436 1437status_t Camera2Client::commandPlayRecordingSoundL() { 1438 mCameraService->playSound(CameraService::SOUND_RECORDING); 1439 return OK; 1440} 1441 1442status_t Camera2Client::commandStartFaceDetectionL(int /*type*/) { 1443 ALOGV("%s: Camera %d: Starting face detection", 1444 __FUNCTION__, mCameraId); 1445 status_t res; 1446 SharedParameters::Lock l(mParameters); 1447 switch (l.mParameters.state) { 1448 case Parameters::DISCONNECTED: 1449 case Parameters::STOPPED: 1450 case Parameters::WAITING_FOR_PREVIEW_WINDOW: 1451 case Parameters::STILL_CAPTURE: 1452 ALOGE("%s: Camera %d: Cannot start face detection without preview active", 1453 __FUNCTION__, mCameraId); 1454 return INVALID_OPERATION; 1455 case Parameters::PREVIEW: 1456 case Parameters::RECORD: 1457 case Parameters::VIDEO_SNAPSHOT: 1458 // Good to go for starting face detect 1459 break; 1460 } 1461 // Ignoring type 1462 if (l.mParameters.fastInfo.bestFaceDetectMode == 1463 ANDROID_STATISTICS_FACE_DETECT_MODE_OFF) { 1464 ALOGE("%s: Camera %d: Face detection not supported", 1465 __FUNCTION__, mCameraId); 1466 return BAD_VALUE; 1467 } 1468 if (l.mParameters.enableFaceDetect) return OK; 1469 1470 l.mParameters.enableFaceDetect = true; 1471 1472 res = updateRequests(l.mParameters); 1473 1474 return res; 1475} 1476 1477status_t Camera2Client::commandStopFaceDetectionL(Parameters ¶ms) { 1478 status_t res = OK; 1479 ALOGV("%s: Camera %d: Stopping face detection", 1480 __FUNCTION__, mCameraId); 1481 1482 if (!params.enableFaceDetect) return OK; 1483 1484 params.enableFaceDetect = false; 1485 1486 if (params.state == Parameters::PREVIEW 1487 || params.state == Parameters::RECORD 1488 || params.state == Parameters::VIDEO_SNAPSHOT) { 1489 res = updateRequests(params); 1490 } 1491 1492 return res; 1493} 1494 1495status_t Camera2Client::commandEnableFocusMoveMsgL(bool enable) { 1496 SharedParameters::Lock l(mParameters); 1497 l.mParameters.enableFocusMoveMessages = enable; 1498 1499 return OK; 1500} 1501 1502status_t Camera2Client::commandPingL() { 1503 // Always ping back if access is proper and device is alive 1504 SharedParameters::Lock l(mParameters); 1505 if (l.mParameters.state != Parameters::DISCONNECTED) { 1506 return OK; 1507 } else { 1508 return NO_INIT; 1509 } 1510} 1511 1512status_t Camera2Client::commandSetVideoBufferCountL(size_t count) { 1513 if (recordingEnabledL()) { 1514 ALOGE("%s: Camera %d: Error setting video buffer count after " 1515 "recording was started", __FUNCTION__, mCameraId); 1516 return INVALID_OPERATION; 1517 } 1518 1519 return mStreamingProcessor->setRecordingBufferCount(count); 1520} 1521 1522/** Device-related methods */ 1523void Camera2Client::notifyAutoFocus(uint8_t newState, int triggerId) { 1524 ALOGV("%s: Autofocus state now %d, last trigger %d", 1525 __FUNCTION__, newState, triggerId); 1526 bool sendCompletedMessage = false; 1527 bool sendMovingMessage = false; 1528 1529 bool success = false; 1530 bool afInMotion = false; 1531 { 1532 SharedParameters::Lock l(mParameters); 1533 // Trace end of AF state 1534 char tmp[32]; 1535 if (l.mParameters.afStateCounter > 0) { 1536 camera_metadata_enum_snprint( 1537 ANDROID_CONTROL_AF_STATE, l.mParameters.focusState, tmp, sizeof(tmp)); 1538 ATRACE_ASYNC_END(tmp, l.mParameters.afStateCounter); 1539 } 1540 1541 // Update state 1542 l.mParameters.focusState = newState; 1543 l.mParameters.afStateCounter++; 1544 1545 // Trace start of AF state 1546 1547 camera_metadata_enum_snprint( 1548 ANDROID_CONTROL_AF_STATE, l.mParameters.focusState, tmp, sizeof(tmp)); 1549 ATRACE_ASYNC_BEGIN(tmp, l.mParameters.afStateCounter); 1550 1551 switch (l.mParameters.focusMode) { 1552 case Parameters::FOCUS_MODE_AUTO: 1553 case Parameters::FOCUS_MODE_MACRO: 1554 // Don't send notifications upstream if they're not for the current AF 1555 // trigger. For example, if cancel was called in between, or if we 1556 // already sent a notification about this AF call. 1557 if (triggerId != l.mParameters.currentAfTriggerId) break; 1558 switch (newState) { 1559 case ANDROID_CONTROL_AF_STATE_FOCUSED_LOCKED: 1560 success = true; 1561 // no break 1562 case ANDROID_CONTROL_AF_STATE_NOT_FOCUSED_LOCKED: 1563 sendCompletedMessage = true; 1564 l.mParameters.currentAfTriggerId = -1; 1565 break; 1566 case ANDROID_CONTROL_AF_STATE_ACTIVE_SCAN: 1567 // Just starting focusing, ignore 1568 break; 1569 case ANDROID_CONTROL_AF_STATE_INACTIVE: 1570 case ANDROID_CONTROL_AF_STATE_PASSIVE_SCAN: 1571 case ANDROID_CONTROL_AF_STATE_PASSIVE_FOCUSED: 1572 case ANDROID_CONTROL_AF_STATE_PASSIVE_UNFOCUSED: 1573 default: 1574 // Unexpected in AUTO/MACRO mode 1575 ALOGE("%s: Unexpected AF state transition in AUTO/MACRO mode: %d", 1576 __FUNCTION__, newState); 1577 break; 1578 } 1579 break; 1580 case Parameters::FOCUS_MODE_CONTINUOUS_VIDEO: 1581 case Parameters::FOCUS_MODE_CONTINUOUS_PICTURE: 1582 switch (newState) { 1583 case ANDROID_CONTROL_AF_STATE_FOCUSED_LOCKED: 1584 success = true; 1585 // no break 1586 case ANDROID_CONTROL_AF_STATE_NOT_FOCUSED_LOCKED: 1587 // Don't send notifications upstream if they're not for 1588 // the current AF trigger. For example, if cancel was 1589 // called in between, or if we already sent a 1590 // notification about this AF call. 1591 // Send both a 'AF done' callback and a 'AF move' callback 1592 if (triggerId != l.mParameters.currentAfTriggerId) break; 1593 sendCompletedMessage = true; 1594 afInMotion = false; 1595 if (l.mParameters.enableFocusMoveMessages && 1596 l.mParameters.afInMotion) { 1597 sendMovingMessage = true; 1598 } 1599 l.mParameters.currentAfTriggerId = -1; 1600 break; 1601 case ANDROID_CONTROL_AF_STATE_INACTIVE: 1602 // Cancel was called, or we switched state; care if 1603 // currently moving 1604 afInMotion = false; 1605 if (l.mParameters.enableFocusMoveMessages && 1606 l.mParameters.afInMotion) { 1607 sendMovingMessage = true; 1608 } 1609 break; 1610 case ANDROID_CONTROL_AF_STATE_PASSIVE_SCAN: 1611 // Start passive scan, inform upstream 1612 afInMotion = true; 1613 // no break 1614 case ANDROID_CONTROL_AF_STATE_PASSIVE_FOCUSED: 1615 case ANDROID_CONTROL_AF_STATE_PASSIVE_UNFOCUSED: 1616 // Stop passive scan, inform upstream 1617 if (l.mParameters.enableFocusMoveMessages) { 1618 sendMovingMessage = true; 1619 } 1620 break; 1621 } 1622 l.mParameters.afInMotion = afInMotion; 1623 break; 1624 case Parameters::FOCUS_MODE_EDOF: 1625 case Parameters::FOCUS_MODE_INFINITY: 1626 case Parameters::FOCUS_MODE_FIXED: 1627 default: 1628 if (newState != ANDROID_CONTROL_AF_STATE_INACTIVE) { 1629 ALOGE("%s: Unexpected AF state change %d " 1630 "(ID %d) in focus mode %d", 1631 __FUNCTION__, newState, triggerId, 1632 l.mParameters.focusMode); 1633 } 1634 } 1635 } 1636 if (sendMovingMessage) { 1637 SharedCameraCallbacks::Lock l(mSharedCameraCallbacks); 1638 if (l.mRemoteCallback != 0) { 1639 l.mRemoteCallback->notifyCallback(CAMERA_MSG_FOCUS_MOVE, 1640 afInMotion ? 1 : 0, 0); 1641 } 1642 } 1643 if (sendCompletedMessage) { 1644 ATRACE_ASYNC_END(kAutofocusLabel, triggerId); 1645 SharedCameraCallbacks::Lock l(mSharedCameraCallbacks); 1646 if (l.mRemoteCallback != 0) { 1647 l.mRemoteCallback->notifyCallback(CAMERA_MSG_FOCUS, 1648 success ? 1 : 0, 0); 1649 } 1650 } 1651} 1652 1653void Camera2Client::notifyAutoExposure(uint8_t newState, int triggerId) { 1654 ALOGV("%s: Autoexposure state now %d, last trigger %d", 1655 __FUNCTION__, newState, triggerId); 1656 mCaptureSequencer->notifyAutoExposure(newState, triggerId); 1657} 1658 1659camera2::SharedParameters& Camera2Client::getParameters() { 1660 return mParameters; 1661} 1662 1663int Camera2Client::getPreviewStreamId() const { 1664 return mStreamingProcessor->getPreviewStreamId(); 1665} 1666 1667int Camera2Client::getCaptureStreamId() const { 1668 return mJpegProcessor->getStreamId(); 1669} 1670 1671int Camera2Client::getCallbackStreamId() const { 1672 return mCallbackProcessor->getStreamId(); 1673} 1674 1675int Camera2Client::getRecordingStreamId() const { 1676 return mStreamingProcessor->getRecordingStreamId(); 1677} 1678 1679int Camera2Client::getZslStreamId() const { 1680 return mZslProcessor->getStreamId(); 1681} 1682 1683status_t Camera2Client::registerFrameListener(int32_t minId, int32_t maxId, 1684 wp<camera2::FrameProcessor::FilteredListener> listener, bool sendPartials) { 1685 return mFrameProcessor->registerListener(minId, maxId, listener, sendPartials); 1686} 1687 1688status_t Camera2Client::removeFrameListener(int32_t minId, int32_t maxId, 1689 wp<camera2::FrameProcessor::FilteredListener> listener) { 1690 return mFrameProcessor->removeListener(minId, maxId, listener); 1691} 1692 1693status_t Camera2Client::stopStream() { 1694 return mStreamingProcessor->stopStream(); 1695} 1696 1697const int32_t Camera2Client::kPreviewRequestIdStart; 1698const int32_t Camera2Client::kPreviewRequestIdEnd; 1699const int32_t Camera2Client::kRecordingRequestIdStart; 1700const int32_t Camera2Client::kRecordingRequestIdEnd; 1701const int32_t Camera2Client::kCaptureRequestIdStart; 1702const int32_t Camera2Client::kCaptureRequestIdEnd; 1703 1704/** Utility methods */ 1705 1706status_t Camera2Client::updateRequests(Parameters ¶ms) { 1707 status_t res; 1708 1709 ALOGV("%s: Camera %d: state = %d", __FUNCTION__, getCameraId(), params.state); 1710 1711 res = mStreamingProcessor->incrementStreamingIds(); 1712 if (res != OK) { 1713 ALOGE("%s: Camera %d: Unable to increment request IDs: %s (%d)", 1714 __FUNCTION__, mCameraId, strerror(-res), res); 1715 return res; 1716 } 1717 1718 res = mStreamingProcessor->updatePreviewRequest(params); 1719 if (res != OK) { 1720 ALOGE("%s: Camera %d: Unable to update preview request: %s (%d)", 1721 __FUNCTION__, mCameraId, strerror(-res), res); 1722 return res; 1723 } 1724 res = mStreamingProcessor->updateRecordingRequest(params); 1725 if (res != OK) { 1726 ALOGE("%s: Camera %d: Unable to update recording request: %s (%d)", 1727 __FUNCTION__, mCameraId, strerror(-res), res); 1728 return res; 1729 } 1730 1731 if (params.state == Parameters::PREVIEW) { 1732 res = startPreviewL(params, true); 1733 if (res != OK) { 1734 ALOGE("%s: Camera %d: Error streaming new preview request: %s (%d)", 1735 __FUNCTION__, mCameraId, strerror(-res), res); 1736 return res; 1737 } 1738 } else if (params.state == Parameters::RECORD || 1739 params.state == Parameters::VIDEO_SNAPSHOT) { 1740 res = startRecordingL(params, true); 1741 if (res != OK) { 1742 ALOGE("%s: Camera %d: Error streaming new record request: %s (%d)", 1743 __FUNCTION__, mCameraId, strerror(-res), res); 1744 return res; 1745 } 1746 } 1747 return res; 1748} 1749 1750 1751size_t Camera2Client::calculateBufferSize(int width, int height, 1752 int format, int stride) { 1753 switch (format) { 1754 case HAL_PIXEL_FORMAT_YCbCr_422_SP: // NV16 1755 return width * height * 2; 1756 case HAL_PIXEL_FORMAT_YCrCb_420_SP: // NV21 1757 return width * height * 3 / 2; 1758 case HAL_PIXEL_FORMAT_YCbCr_422_I: // YUY2 1759 return width * height * 2; 1760 case HAL_PIXEL_FORMAT_YV12: { // YV12 1761 size_t ySize = stride * height; 1762 size_t uvStride = (stride / 2 + 0xF) & ~0xF; 1763 size_t uvSize = uvStride * height / 2; 1764 return ySize + uvSize * 2; 1765 } 1766 case HAL_PIXEL_FORMAT_RGB_565: 1767 return width * height * 2; 1768 case HAL_PIXEL_FORMAT_RGBA_8888: 1769 return width * height * 4; 1770 case HAL_PIXEL_FORMAT_RAW_SENSOR: 1771 return width * height * 2; 1772 default: 1773 ALOGE("%s: Unknown preview format: %x", 1774 __FUNCTION__, format); 1775 return 0; 1776 } 1777} 1778 1779status_t Camera2Client::syncWithDevice() { 1780 ATRACE_CALL(); 1781 const nsecs_t kMaxSyncTimeout = 500000000; // 500 ms 1782 status_t res; 1783 1784 int32_t activeRequestId = mStreamingProcessor->getActiveRequestId(); 1785 if (activeRequestId == 0) return OK; 1786 1787 res = mDevice->waitUntilRequestReceived(activeRequestId, kMaxSyncTimeout); 1788 if (res == TIMED_OUT) { 1789 ALOGE("%s: Camera %d: Timed out waiting sync with HAL", 1790 __FUNCTION__, mCameraId); 1791 } else if (res != OK) { 1792 ALOGE("%s: Camera %d: Error while waiting to sync with HAL", 1793 __FUNCTION__, mCameraId); 1794 } 1795 return res; 1796} 1797 1798template <typename ProcessorT> 1799status_t Camera2Client::updateProcessorStream(sp<ProcessorT> processor, 1800 camera2::Parameters params) { 1801 // No default template arguments until C++11, so we need this overload 1802 return updateProcessorStream<ProcessorT, &ProcessorT::updateStream>( 1803 processor, params); 1804} 1805 1806template <typename ProcessorT, 1807 status_t (ProcessorT::*updateStreamF)(const Parameters &)> 1808status_t Camera2Client::updateProcessorStream(sp<ProcessorT> processor, 1809 Parameters params) { 1810 status_t res; 1811 1812 // Get raw pointer since sp<T> doesn't have operator->* 1813 ProcessorT *processorPtr = processor.get(); 1814 res = (processorPtr->*updateStreamF)(params); 1815 1816 /** 1817 * Can't update the stream if it's busy? 1818 * 1819 * Then we need to stop the device (by temporarily clearing the request 1820 * queue) and then try again. Resume streaming once we're done. 1821 */ 1822 if (res == -EBUSY) { 1823 ALOGV("%s: Camera %d: Pausing to update stream", __FUNCTION__, 1824 mCameraId); 1825 res = mStreamingProcessor->togglePauseStream(/*pause*/true); 1826 if (res != OK) { 1827 ALOGE("%s: Camera %d: Can't pause streaming: %s (%d)", 1828 __FUNCTION__, mCameraId, strerror(-res), res); 1829 } 1830 1831 res = mDevice->waitUntilDrained(); 1832 if (res != OK) { 1833 ALOGE("%s: Camera %d: Waiting to stop streaming failed: %s (%d)", 1834 __FUNCTION__, mCameraId, strerror(-res), res); 1835 } 1836 1837 res = (processorPtr->*updateStreamF)(params); 1838 if (res != OK) { 1839 ALOGE("%s: Camera %d: Failed to update processing stream " 1840 " despite having halted streaming first: %s (%d)", 1841 __FUNCTION__, mCameraId, strerror(-res), res); 1842 } 1843 1844 res = mStreamingProcessor->togglePauseStream(/*pause*/false); 1845 if (res != OK) { 1846 ALOGE("%s: Camera %d: Can't unpause streaming: %s (%d)", 1847 __FUNCTION__, mCameraId, strerror(-res), res); 1848 } 1849 } 1850 1851 return res; 1852} 1853 1854const char* Camera2Client::kAutofocusLabel = "autofocus"; 1855const char* Camera2Client::kTakepictureLabel = "take_picture"; 1856 1857} // namespace android 1858