/* * Copyright (C) 2013 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package android.hardware.camera2; import android.hardware.camera2.impl.CameraMetadataNative; import java.util.Collections; import java.util.List; /** *

The properties describing a * {@link CameraDevice CameraDevice}.

* *

These properties are fixed for a given CameraDevice, and can be queried * through the {@link CameraManager CameraManager} * interface in addition to through the CameraDevice interface.

* * @see CameraDevice * @see CameraManager */ public final class CameraCharacteristics extends CameraMetadata { private final CameraMetadataNative mProperties; private List> mAvailableRequestKeys; private List> mAvailableResultKeys; /** * Takes ownership of the passed-in properties object * @hide */ public CameraCharacteristics(CameraMetadataNative properties) { mProperties = properties; } @Override public T get(Key key) { return mProperties.get(key); } /** * Returns the list of keys supported by this {@link CameraDevice} for querying * with a {@link CaptureRequest}. * *

The list returned is not modifiable, so any attempts to modify it will throw * a {@code UnsupportedOperationException}.

* *

Each key is only listed once in the list. The order of the keys is undefined.

* *

Note that there is no {@code getAvailableCameraCharacteristicsKeys()} -- use * {@link #getKeys()} instead.

* * @return List of keys supported by this CameraDevice for CaptureRequests. */ public List> getAvailableCaptureRequestKeys() { if (mAvailableRequestKeys == null) { mAvailableRequestKeys = getAvailableKeyList(CaptureRequest.class); } return mAvailableRequestKeys; } /** * Returns the list of keys supported by this {@link CameraDevice} for querying * with a {@link CaptureResult}. * *

The list returned is not modifiable, so any attempts to modify it will throw * a {@code UnsupportedOperationException}.

* *

Each key is only listed once in the list. The order of the keys is undefined.

* *

Note that there is no {@code getAvailableCameraCharacteristicsKeys()} -- use * {@link #getKeys()} instead.

* * @return List of keys supported by this CameraDevice for CaptureResults. */ public List> getAvailableCaptureResultKeys() { if (mAvailableResultKeys == null) { mAvailableResultKeys = getAvailableKeyList(CaptureResult.class); } return mAvailableResultKeys; } /** * Returns the list of keys supported by this {@link CameraDevice} by metadataClass. * *

The list returned is not modifiable, so any attempts to modify it will throw * a {@code UnsupportedOperationException}.

* *

Each key is only listed once in the list. The order of the keys is undefined.

* * @param metadataClass The subclass of CameraMetadata that you want to get the keys for. * * @return List of keys supported by this CameraDevice for metadataClass. * * @throws IllegalArgumentException if metadataClass is not a subclass of CameraMetadata */ private List> getAvailableKeyList(Class metadataClass) { if (metadataClass.equals(CameraMetadata.class)) { throw new AssertionError( "metadataClass must be a strict subclass of CameraMetadata"); } else if (!CameraMetadata.class.isAssignableFrom(metadataClass)) { throw new AssertionError( "metadataClass must be a subclass of CameraMetadata"); } return Collections.unmodifiableList(getKeysStatic(metadataClass, /*instance*/null)); } /*@O~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~ * The key entries below this point are generated from metadata * definitions in /system/media/camera/docs. Do not modify by hand or * modify the comment blocks at the start or end. *~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~*/ /** *

The set of auto-exposure antibanding modes that are * supported by this camera device.

*

Not all of the auto-exposure anti-banding modes may be * supported by a given camera device. This field lists the * valid anti-banding modes that the application may request * for this camera device; they must include AUTO.

*/ public static final Key CONTROL_AE_AVAILABLE_ANTIBANDING_MODES = new Key("android.control.aeAvailableAntibandingModes", byte[].class); /** *

The set of auto-exposure modes that are supported by this * camera device.

*

Not all the auto-exposure modes may be supported by a * given camera device, especially if no flash unit is * available. This entry lists the valid modes for * {@link CaptureRequest#CONTROL_AE_MODE android.control.aeMode} for this camera device.

*

All camera devices support ON, and all camera devices with * flash units support ON_AUTO_FLASH and * ON_ALWAYS_FLASH.

*

Full-capability camera devices always support OFF mode, * which enables application control of camera exposure time, * sensitivity, and frame duration.

* * @see CaptureRequest#CONTROL_AE_MODE */ public static final Key CONTROL_AE_AVAILABLE_MODES = new Key("android.control.aeAvailableModes", byte[].class); /** *

List of frame rate ranges supported by the * AE algorithm/hardware

*/ public static final Key CONTROL_AE_AVAILABLE_TARGET_FPS_RANGES = new Key("android.control.aeAvailableTargetFpsRanges", int[].class); /** *

Maximum and minimum exposure compensation * setting, in counts of * android.control.aeCompensationStepSize

*/ public static final Key CONTROL_AE_COMPENSATION_RANGE = new Key("android.control.aeCompensationRange", int[].class); /** *

Smallest step by which exposure compensation * can be changed

*/ public static final Key CONTROL_AE_COMPENSATION_STEP = new Key("android.control.aeCompensationStep", Rational.class); /** *

List of AF modes that can be * selected with {@link CaptureRequest#CONTROL_AF_MODE android.control.afMode}.

*

Not all the auto-focus modes may be supported by a * given camera device. This entry lists the valid modes for * {@link CaptureRequest#CONTROL_AF_MODE android.control.afMode} for this camera device.

*

All camera devices will support OFF mode, and all camera devices with * adjustable focuser units ({@link CameraCharacteristics#LENS_INFO_MINIMUM_FOCUS_DISTANCE android.lens.info.minimumFocusDistance} > 0) * will support AUTO mode.

* * @see CaptureRequest#CONTROL_AF_MODE * @see CameraCharacteristics#LENS_INFO_MINIMUM_FOCUS_DISTANCE */ public static final Key CONTROL_AF_AVAILABLE_MODES = new Key("android.control.afAvailableModes", byte[].class); /** *

List containing the subset of color effects * specified in {@link CaptureRequest#CONTROL_EFFECT_MODE android.control.effectMode} that is supported by * this device.

*

This list contains the color effect modes that can be applied to * images produced by the camera device. Only modes that have * been fully implemented for the current device may be included here. * Implementations are not expected to be consistent across all devices. * If no color effect modes are available for a device, this should * simply be set to OFF.

*

A color effect will only be applied if * {@link CaptureRequest#CONTROL_MODE android.control.mode} != OFF.

* * @see CaptureRequest#CONTROL_EFFECT_MODE * @see CaptureRequest#CONTROL_MODE */ public static final Key CONTROL_AVAILABLE_EFFECTS = new Key("android.control.availableEffects", byte[].class); /** *

List containing a subset of scene modes * specified in {@link CaptureRequest#CONTROL_SCENE_MODE android.control.sceneMode}.

*

This list contains scene modes that can be set for the camera device. * Only scene modes that have been fully implemented for the * camera device may be included here. Implementations are not expected * to be consistent across all devices. If no scene modes are supported * by the camera device, this will be set to [DISABLED].

* * @see CaptureRequest#CONTROL_SCENE_MODE */ public static final Key CONTROL_AVAILABLE_SCENE_MODES = new Key("android.control.availableSceneModes", byte[].class); /** *

List of video stabilization modes that can * be supported

*/ public static final Key CONTROL_AVAILABLE_VIDEO_STABILIZATION_MODES = new Key("android.control.availableVideoStabilizationModes", byte[].class); /** *

The set of auto-white-balance modes ({@link CaptureRequest#CONTROL_AWB_MODE android.control.awbMode}) * that are supported by this camera device.

*

Not all the auto-white-balance modes may be supported by a * given camera device. This entry lists the valid modes for * {@link CaptureRequest#CONTROL_AWB_MODE android.control.awbMode} for this camera device.

*

All camera devices will support ON mode.

*

Full-capability camera devices will always support OFF mode, * which enables application control of white balance, by using * {@link CaptureRequest#COLOR_CORRECTION_TRANSFORM android.colorCorrection.transform} and {@link CaptureRequest#COLOR_CORRECTION_GAINS android.colorCorrection.gains}({@link CaptureRequest#COLOR_CORRECTION_MODE android.colorCorrection.mode} must be set to TRANSFORM_MATRIX).

* * @see CaptureRequest#COLOR_CORRECTION_GAINS * @see CaptureRequest#COLOR_CORRECTION_MODE * @see CaptureRequest#COLOR_CORRECTION_TRANSFORM * @see CaptureRequest#CONTROL_AWB_MODE */ public static final Key CONTROL_AWB_AVAILABLE_MODES = new Key("android.control.awbAvailableModes", byte[].class); /** *

For AE, AWB, and AF, how many individual * regions can be listed for metering?

*/ public static final Key CONTROL_MAX_REGIONS = new Key("android.control.maxRegions", int.class); /** *

Whether this camera device has a * flash.

*

If no flash, none of the flash controls do * anything. All other metadata should return 0.

*/ public static final Key FLASH_INFO_AVAILABLE = new Key("android.flash.info.available", boolean.class); /** *

Supported resolutions for the JPEG thumbnail

*

Below condiditions will be satisfied for this size list:

*
    *
  • The sizes will be sorted by increasing pixel area (width x height). * If several resolutions have the same area, they will be sorted by increasing width.
  • *
  • The aspect ratio of the largest thumbnail size will be same as the * aspect ratio of largest size in {@link CameraCharacteristics#SCALER_AVAILABLE_JPEG_SIZES android.scaler.availableJpegSizes}. * The largest size is defined as the size that has the largest pixel area * in a given size list.
  • *
  • Each size in {@link CameraCharacteristics#SCALER_AVAILABLE_JPEG_SIZES android.scaler.availableJpegSizes} will have at least * one corresponding size that has the same aspect ratio in availableThumbnailSizes, * and vice versa.
  • *
  • All non (0, 0) sizes will have non-zero widths and heights.
  • *
* * @see CameraCharacteristics#SCALER_AVAILABLE_JPEG_SIZES */ public static final Key JPEG_AVAILABLE_THUMBNAIL_SIZES = new Key("android.jpeg.availableThumbnailSizes", android.hardware.camera2.Size[].class); /** *

List of supported aperture * values.

*

If the camera device doesn't support variable apertures, * listed value will be the fixed aperture.

*

If the camera device supports variable apertures, the aperture value * in this list will be sorted in ascending order.

*/ public static final Key LENS_INFO_AVAILABLE_APERTURES = new Key("android.lens.info.availableApertures", float[].class); /** *

List of supported neutral density filter values for * {@link CaptureRequest#LENS_FILTER_DENSITY android.lens.filterDensity}.

*

If changing {@link CaptureRequest#LENS_FILTER_DENSITY android.lens.filterDensity} is not supported, * availableFilterDensities must contain only 0. Otherwise, this * list contains only the exact filter density values available on * this camera device.

* * @see CaptureRequest#LENS_FILTER_DENSITY */ public static final Key LENS_INFO_AVAILABLE_FILTER_DENSITIES = new Key("android.lens.info.availableFilterDensities", float[].class); /** *

The available focal lengths for this device for use with * {@link CaptureRequest#LENS_FOCAL_LENGTH android.lens.focalLength}.

*

If optical zoom is not supported, this will only report * a single value corresponding to the static focal length of the * device. Otherwise, this will report every focal length supported * by the device.

* * @see CaptureRequest#LENS_FOCAL_LENGTH */ public static final Key LENS_INFO_AVAILABLE_FOCAL_LENGTHS = new Key("android.lens.info.availableFocalLengths", float[].class); /** *

List containing a subset of the optical image * stabilization (OIS) modes specified in * {@link CaptureRequest#LENS_OPTICAL_STABILIZATION_MODE android.lens.opticalStabilizationMode}.

*

If OIS is not implemented for a given camera device, this should * contain only OFF.

* * @see CaptureRequest#LENS_OPTICAL_STABILIZATION_MODE */ public static final Key LENS_INFO_AVAILABLE_OPTICAL_STABILIZATION = new Key("android.lens.info.availableOpticalStabilization", byte[].class); /** *

Hyperfocal distance for this lens; set to * 0 if fixed focus

*

The hyperfocal distance is used for the old * API's 'fixed' setting

*/ public static final Key LENS_INFO_HYPERFOCAL_DISTANCE = new Key("android.lens.info.hyperfocalDistance", float.class); /** *

Shortest distance from frontmost surface * of the lens that can be focused correctly

*

If the lens is fixed-focus, this should be * 0

*/ public static final Key LENS_INFO_MINIMUM_FOCUS_DISTANCE = new Key("android.lens.info.minimumFocusDistance", float.class); /** *

Dimensions of lens shading map.

*

The map should be on the order of 30-40 rows and columns, and * must be smaller than 64x64.

*/ public static final Key LENS_INFO_SHADING_MAP_SIZE = new Key("android.lens.info.shadingMapSize", android.hardware.camera2.Size.class); /** *

Direction the camera faces relative to * device screen

* @see #LENS_FACING_FRONT * @see #LENS_FACING_BACK */ public static final Key LENS_FACING = new Key("android.lens.facing", int.class); /** *

If set to 1, the HAL will always split result * metadata for a single capture into multiple buffers, * returned using multiple process_capture_result calls.

*

Does not need to be listed in static * metadata. Support for partial results will be reworked in * future versions of camera service. This quirk will stop * working at that point; DO NOT USE without careful * consideration of future support.

*

Optional - This value may be {@code null} on some devices.

* @hide */ public static final Key QUIRKS_USE_PARTIAL_RESULT = new Key("android.quirks.usePartialResult", byte.class); /** *

How many output streams can be allocated at * the same time for each type of stream

*

Video snapshot with preview callbacks requires 3 * processed streams (preview, record, app callbacks) and * one JPEG stream (snapshot)

*/ public static final Key REQUEST_MAX_NUM_OUTPUT_STREAMS = new Key("android.request.maxNumOutputStreams", int[].class); /** *

Specifies the number of maximum pipeline stages a frame * has to go through from when it's exposed to when it's available * to the framework.

*

A typical minimum value for this is 2 (one stage to expose, * one stage to readout) from the sensor. The ISP then usually adds * its own stages to do custom HW processing. Further stages may be * added by SW processing.

*

Depending on what settings are used (e.g. YUV, JPEG) and what * processing is enabled (e.g. face detection), the actual pipeline * depth (specified by {@link CaptureResult#REQUEST_PIPELINE_DEPTH android.request.pipelineDepth}) may be less than * the max pipeline depth.

*

A pipeline depth of X stages is equivalent to a pipeline latency of * X frame intervals.

*

This value will be 8 or less.

* * @see CaptureResult#REQUEST_PIPELINE_DEPTH */ public static final Key REQUEST_PIPELINE_MAX_DEPTH = new Key("android.request.pipelineMaxDepth", byte.class); /** *

Optional. Defaults to 1. Defines how many sub-components * a result will be composed of.

*

In order to combat the pipeline latency, partial results * may be delivered to the application layer from the camera device as * soon as they are available.

*

A value of 1 means that partial results are not supported.

*

A typical use case for this might be: after requesting an AF lock the * new AF state might be available 50% of the way through the pipeline. * The camera device could then immediately dispatch this state via a * partial result to the framework/application layer, and the rest of * the metadata via later partial results.

*/ public static final Key REQUEST_PARTIAL_RESULT_COUNT = new Key("android.request.partialResultCount", int.class); /** *

The list of image formats that are supported by this * camera device.

*

All camera devices will support JPEG and YUV_420_888 formats.

*

When set to YUV_420_888, application can access the YUV420 data directly.

*/ public static final Key SCALER_AVAILABLE_FORMATS = new Key("android.scaler.availableFormats", int[].class); /** *

The minimum frame duration that is supported * for each resolution in {@link CameraCharacteristics#SCALER_AVAILABLE_JPEG_SIZES android.scaler.availableJpegSizes}.

*

This corresponds to the minimum steady-state frame duration when only * that JPEG stream is active and captured in a burst, with all * processing (typically in android.*.mode) set to FAST.

*

When multiple streams are configured, the minimum * frame duration will be >= max(individual stream min * durations)

* * @see CameraCharacteristics#SCALER_AVAILABLE_JPEG_SIZES */ public static final Key SCALER_AVAILABLE_JPEG_MIN_DURATIONS = new Key("android.scaler.availableJpegMinDurations", long[].class); /** *

The JPEG resolutions that are supported by this camera device.

*

The resolutions are listed as (width, height) pairs. All camera devices will support * sensor maximum resolution (defined by {@link CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE android.sensor.info.activeArraySize}).

* * @see CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE */ public static final Key SCALER_AVAILABLE_JPEG_SIZES = new Key("android.scaler.availableJpegSizes", android.hardware.camera2.Size[].class); /** *

The maximum ratio between active area width * and crop region width, or between active area height and * crop region height, if the crop region height is larger * than width

*/ public static final Key SCALER_AVAILABLE_MAX_DIGITAL_ZOOM = new Key("android.scaler.availableMaxDigitalZoom", float.class); /** *

For each available processed output size (defined in * {@link CameraCharacteristics#SCALER_AVAILABLE_PROCESSED_SIZES android.scaler.availableProcessedSizes}), this property lists the * minimum supportable frame duration for that size.

*

This should correspond to the frame duration when only that processed * stream is active, with all processing (typically in android.*.mode) * set to FAST.

*

When multiple streams are configured, the minimum frame duration will * be >= max(individual stream min durations).

* * @see CameraCharacteristics#SCALER_AVAILABLE_PROCESSED_SIZES */ public static final Key SCALER_AVAILABLE_PROCESSED_MIN_DURATIONS = new Key("android.scaler.availableProcessedMinDurations", long[].class); /** *

The resolutions available for use with * processed output streams, such as YV12, NV12, and * platform opaque YUV/RGB streams to the GPU or video * encoders.

*

The resolutions are listed as (width, height) pairs.

*

For a given use case, the actual maximum supported resolution * may be lower than what is listed here, depending on the destination * Surface for the image data. For example, for recording video, * the video encoder chosen may have a maximum size limit (e.g. 1080p) * smaller than what the camera (e.g. maximum resolution is 3264x2448) * can provide.

*

Please reference the documentation for the image data destination to * check if it limits the maximum size for image data.

*/ public static final Key SCALER_AVAILABLE_PROCESSED_SIZES = new Key("android.scaler.availableProcessedSizes", android.hardware.camera2.Size[].class); /** *

Area of raw data which corresponds to only * active pixels.

*

It is smaller or equal to * sensor full pixel array, which could include the black calibration pixels.

*/ public static final Key SENSOR_INFO_ACTIVE_ARRAY_SIZE = new Key("android.sensor.info.activeArraySize", android.graphics.Rect.class); /** *

Range of valid sensitivities

*/ public static final Key SENSOR_INFO_SENSITIVITY_RANGE = new Key("android.sensor.info.sensitivityRange", int[].class); /** *

Range of valid exposure * times

*/ public static final Key SENSOR_INFO_EXPOSURE_TIME_RANGE = new Key("android.sensor.info.exposureTimeRange", long[].class); /** *

Maximum possible frame duration (minimum frame * rate)

*

Minimum duration is a function of resolution, * processing settings. See * android.scaler.availableProcessedMinDurations * android.scaler.availableJpegMinDurations * android.scaler.availableRawMinDurations

*/ public static final Key SENSOR_INFO_MAX_FRAME_DURATION = new Key("android.sensor.info.maxFrameDuration", long.class); /** *

The physical dimensions of the full pixel * array

*

Needed for FOV calculation for old API

*/ public static final Key SENSOR_INFO_PHYSICAL_SIZE = new Key("android.sensor.info.physicalSize", float[].class); /** *

Gain factor from electrons to raw units when * ISO=100

*

Optional - This value may be {@code null} on some devices.

*

Full capability - * Present on all camera devices that report being {@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL_FULL HARDWARE_LEVEL_FULL} devices in the * {@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL android.info.supportedHardwareLevel} key

* * @see CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL */ public static final Key SENSOR_BASE_GAIN_FACTOR = new Key("android.sensor.baseGainFactor", Rational.class); /** *

Maximum sensitivity that is implemented * purely through analog gain.

*

For {@link CaptureRequest#SENSOR_SENSITIVITY android.sensor.sensitivity} values less than or * equal to this, all applied gain must be analog. For * values above this, the gain applied can be a mix of analog and * digital.

*

Optional - This value may be {@code null} on some devices.

*

Full capability - * Present on all camera devices that report being {@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL_FULL HARDWARE_LEVEL_FULL} devices in the * {@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL android.info.supportedHardwareLevel} key

* * @see CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL * @see CaptureRequest#SENSOR_SENSITIVITY */ public static final Key SENSOR_MAX_ANALOG_SENSITIVITY = new Key("android.sensor.maxAnalogSensitivity", int.class); /** *

Clockwise angle through which the output * image needs to be rotated to be upright on the device * screen in its native orientation. Also defines the * direction of rolling shutter readout, which is from top * to bottom in the sensor's coordinate system

*/ public static final Key SENSOR_ORIENTATION = new Key("android.sensor.orientation", int.class); /** *

Optional. Defaults to [OFF]. Lists the supported test * pattern modes for android.test.patternMode.

*

Optional - This value may be {@code null} on some devices.

*/ public static final Key SENSOR_AVAILABLE_TEST_PATTERN_MODES = new Key("android.sensor.availableTestPatternModes", byte.class); /** *

Which face detection modes are available, * if any

*

OFF means face detection is disabled, it must * be included in the list.

*

SIMPLE means the device supports the * android.statistics.faceRectangles and * android.statistics.faceScores outputs.

*

FULL means the device additionally supports the * android.statistics.faceIds and * android.statistics.faceLandmarks outputs.

*/ public static final Key STATISTICS_INFO_AVAILABLE_FACE_DETECT_MODES = new Key("android.statistics.info.availableFaceDetectModes", byte[].class); /** *

Maximum number of simultaneously detectable * faces

*/ public static final Key STATISTICS_INFO_MAX_FACE_COUNT = new Key("android.statistics.info.maxFaceCount", int.class); /** *

Maximum number of supported points in the * tonemap curve that can be used for {@link CaptureRequest#TONEMAP_CURVE_RED android.tonemap.curveRed}, or * {@link CaptureRequest#TONEMAP_CURVE_GREEN android.tonemap.curveGreen}, or {@link CaptureRequest#TONEMAP_CURVE_BLUE android.tonemap.curveBlue}.

*

If the actual number of points provided by the application (in * android.tonemap.curve*) is less than max, the camera device will * resample the curve to its internal representation, using linear * interpolation.

*

The output curves in the result metadata may have a different number * of points than the input curves, and will represent the actual * hardware curves used as closely as possible when linearly interpolated.

* * @see CaptureRequest#TONEMAP_CURVE_BLUE * @see CaptureRequest#TONEMAP_CURVE_GREEN * @see CaptureRequest#TONEMAP_CURVE_RED */ public static final Key TONEMAP_MAX_CURVE_POINTS = new Key("android.tonemap.maxCurvePoints", int.class); /** *

A list of camera LEDs that are available on this system.

* @see #LED_AVAILABLE_LEDS_TRANSMIT * @hide */ public static final Key LED_AVAILABLE_LEDS = new Key("android.led.availableLeds", int[].class); /** *

The camera 3 HAL device can implement one of two possible * operational modes; limited and full. Full support is * expected from new higher-end devices. Limited mode has * hardware requirements roughly in line with those for a * camera HAL device v1 implementation, and is expected from * older or inexpensive devices. Full is a strict superset of * limited, and they share the same essential operational flow.

*

For full details refer to "S3. Operational Modes" in camera3.h

* @see #INFO_SUPPORTED_HARDWARE_LEVEL_LIMITED * @see #INFO_SUPPORTED_HARDWARE_LEVEL_FULL */ public static final Key INFO_SUPPORTED_HARDWARE_LEVEL = new Key("android.info.supportedHardwareLevel", int.class); /** *

The maximum number of frames that can occur after a request * (different than the previous) has been submitted, and before the * result's state becomes synchronized (by setting * android.sync.frameNumber to a non-negative value).

*

This defines the maximum distance (in number of metadata results), * between android.sync.frameNumber and the equivalent * android.request.frameCount.

*

In other words this acts as an upper boundary for how many frames * must occur before the camera device knows for a fact that the new * submitted camera settings have been applied in outgoing frames.

*

For example if the distance was 2,

*
initial request = X (repeating)
     * request1 = X
     * request2 = Y
     * request3 = Y
     * request4 = Y
     *
     * where requestN has frameNumber N, and the first of the repeating
     * initial request's has frameNumber F (and F < 1).
     *
     * initial result = X' + { android.sync.frameNumber == F }
     * result1 = X' + { android.sync.frameNumber == F }
     * result2 = X' + { android.sync.frameNumber == CONVERGING }
     * result3 = X' + { android.sync.frameNumber == CONVERGING }
     * result4 = X' + { android.sync.frameNumber == 2 }
     *
     * where resultN has frameNumber N.
     * 
*

Since result4 has a frameNumber == 4 and * android.sync.frameNumber == 2, the distance is clearly * 4 - 2 = 2.

* @see #SYNC_MAX_LATENCY_PER_FRAME_CONTROL * @see #SYNC_MAX_LATENCY_UNKNOWN */ public static final Key SYNC_MAX_LATENCY = new Key("android.sync.maxLatency", int.class); /*~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~ * End generated code *~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~O@*/ }