/* * Copyright (C) 2013 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package android.hardware.camera2; import android.hardware.camera2.impl.CameraMetadataNative; import java.util.Collections; import java.util.List; /** *
The properties describing a * {@link CameraDevice CameraDevice}.
* *These properties are fixed for a given CameraDevice, and can be queried * through the {@link CameraManager CameraManager} * interface in addition to through the CameraDevice interface.
* * @see CameraDevice * @see CameraManager */ public final class CameraCharacteristics extends CameraMetadata { private final CameraMetadataNative mProperties; private ListThe list returned is not modifiable, so any attempts to modify it will throw * a {@code UnsupportedOperationException}.
* *Each key is only listed once in the list. The order of the keys is undefined.
* *Note that there is no {@code getAvailableCameraCharacteristicsKeys()} -- use * {@link #getKeys()} instead.
* * @return List of keys supported by this CameraDevice for CaptureRequests. */ public ListThe list returned is not modifiable, so any attempts to modify it will throw * a {@code UnsupportedOperationException}.
* *Each key is only listed once in the list. The order of the keys is undefined.
* *Note that there is no {@code getAvailableCameraCharacteristicsKeys()} -- use * {@link #getKeys()} instead.
* * @return List of keys supported by this CameraDevice for CaptureResults. */ public ListThe list returned is not modifiable, so any attempts to modify it will throw * a {@code UnsupportedOperationException}.
* *Each key is only listed once in the list. The order of the keys is undefined.
* * @param metadataClass The subclass of CameraMetadata that you want to get the keys for. * * @return List of keys supported by this CameraDevice for metadataClass. * * @throws IllegalArgumentException if metadataClass is not a subclass of CameraMetadata */ privateThe set of auto-exposure antibanding modes that are * supported by this camera device.
*Not all of the auto-exposure anti-banding modes may be * supported by a given camera device. This field lists the * valid anti-banding modes that the application may request * for this camera device; they must include AUTO.
*/ public static final KeyThe set of auto-exposure modes that are supported by this * camera device.
*Not all the auto-exposure modes may be supported by a * given camera device, especially if no flash unit is * available. This entry lists the valid modes for * {@link CaptureRequest#CONTROL_AE_MODE android.control.aeMode} for this camera device.
*All camera devices support ON, and all camera devices with * flash units support ON_AUTO_FLASH and * ON_ALWAYS_FLASH.
*Full-capability camera devices always support OFF mode, * which enables application control of camera exposure time, * sensitivity, and frame duration.
* * @see CaptureRequest#CONTROL_AE_MODE */ public static final KeyList of frame rate ranges supported by the * AE algorithm/hardware
*/ public static final KeyMaximum and minimum exposure compensation * setting, in counts of * {@link CameraCharacteristics#CONTROL_AE_COMPENSATION_STEP android.control.aeCompensationStep}.
* * @see CameraCharacteristics#CONTROL_AE_COMPENSATION_STEP */ public static final KeySmallest step by which exposure compensation * can be changed
*/ public static final KeyList of AF modes that can be * selected with {@link CaptureRequest#CONTROL_AF_MODE android.control.afMode}.
*Not all the auto-focus modes may be supported by a * given camera device. This entry lists the valid modes for * {@link CaptureRequest#CONTROL_AF_MODE android.control.afMode} for this camera device.
*All camera devices will support OFF mode, and all camera devices with
* adjustable focuser units ({@link CameraCharacteristics#LENS_INFO_MINIMUM_FOCUS_DISTANCE android.lens.info.minimumFocusDistance} > 0
)
* will support AUTO mode.
List containing the subset of color effects * specified in {@link CaptureRequest#CONTROL_EFFECT_MODE android.control.effectMode} that is supported by * this device.
*This list contains the color effect modes that can be applied to * images produced by the camera device. Only modes that have * been fully implemented for the current device may be included here. * Implementations are not expected to be consistent across all devices. * If no color effect modes are available for a device, this should * simply be set to OFF.
*A color effect will only be applied if * {@link CaptureRequest#CONTROL_MODE android.control.mode} != OFF.
* * @see CaptureRequest#CONTROL_EFFECT_MODE * @see CaptureRequest#CONTROL_MODE */ public static final KeyList containing a subset of scene modes * specified in {@link CaptureRequest#CONTROL_SCENE_MODE android.control.sceneMode}.
*This list contains scene modes that can be set for the camera device.
* Only scene modes that have been fully implemented for the
* camera device may be included here. Implementations are not expected
* to be consistent across all devices. If no scene modes are supported
* by the camera device, this will be set to [DISABLED]
.
List of video stabilization modes that can * be supported
*/ public static final KeyThe set of auto-white-balance modes ({@link CaptureRequest#CONTROL_AWB_MODE android.control.awbMode}) * that are supported by this camera device.
*Not all the auto-white-balance modes may be supported by a * given camera device. This entry lists the valid modes for * {@link CaptureRequest#CONTROL_AWB_MODE android.control.awbMode} for this camera device.
*All camera devices will support ON mode.
*Full-capability camera devices will always support OFF mode, * which enables application control of white balance, by using * {@link CaptureRequest#COLOR_CORRECTION_TRANSFORM android.colorCorrection.transform} and {@link CaptureRequest#COLOR_CORRECTION_GAINS android.colorCorrection.gains}({@link CaptureRequest#COLOR_CORRECTION_MODE android.colorCorrection.mode} must be set to TRANSFORM_MATRIX).
* * @see CaptureRequest#COLOR_CORRECTION_GAINS * @see CaptureRequest#COLOR_CORRECTION_MODE * @see CaptureRequest#COLOR_CORRECTION_TRANSFORM * @see CaptureRequest#CONTROL_AWB_MODE */ public static final KeyList of the maximum number of regions that can be used for metering in * auto-exposure (AE), auto-white balance (AWB), and auto-focus (AF); * this corresponds to the the maximum number of elements in * {@link CaptureRequest#CONTROL_AE_REGIONS android.control.aeRegions}, {@link CaptureRequest#CONTROL_AWB_REGIONS android.control.awbRegions}, * and {@link CaptureRequest#CONTROL_AF_REGIONS android.control.afRegions}.
* * @see CaptureRequest#CONTROL_AE_REGIONS * @see CaptureRequest#CONTROL_AF_REGIONS * @see CaptureRequest#CONTROL_AWB_REGIONS */ public static final KeyWhether this camera device has a * flash.
*If no flash, none of the flash controls do * anything. All other metadata should return 0.
*/ public static final KeySupported resolutions for the JPEG thumbnail
*Below condiditions will be satisfied for this size list:
*List of supported aperture * values.
*If the camera device doesn't support variable apertures, * listed value will be the fixed aperture.
*If the camera device supports variable apertures, the aperture value * in this list will be sorted in ascending order.
*/ public static final KeyList of supported neutral density filter values for * {@link CaptureRequest#LENS_FILTER_DENSITY android.lens.filterDensity}.
*If changing {@link CaptureRequest#LENS_FILTER_DENSITY android.lens.filterDensity} is not supported, * availableFilterDensities must contain only 0. Otherwise, this * list contains only the exact filter density values available on * this camera device.
* * @see CaptureRequest#LENS_FILTER_DENSITY */ public static final KeyThe available focal lengths for this device for use with * {@link CaptureRequest#LENS_FOCAL_LENGTH android.lens.focalLength}.
*If optical zoom is not supported, this will only report * a single value corresponding to the static focal length of the * device. Otherwise, this will report every focal length supported * by the device.
* * @see CaptureRequest#LENS_FOCAL_LENGTH */ public static final KeyList containing a subset of the optical image * stabilization (OIS) modes specified in * {@link CaptureRequest#LENS_OPTICAL_STABILIZATION_MODE android.lens.opticalStabilizationMode}.
*If OIS is not implemented for a given camera device, this should * contain only OFF.
* * @see CaptureRequest#LENS_OPTICAL_STABILIZATION_MODE */ public static final KeyOptional. Hyperfocal distance for this lens.
*If the lens is fixed focus, the camera device will report 0.
*If the lens is not fixed focus, the camera device will report this * field when {@link CameraCharacteristics#LENS_INFO_FOCUS_DISTANCE_CALIBRATION android.lens.info.focusDistanceCalibration} is APPROXIMATE or CALIBRATED.
*Optional - This value may be {@code null} on some devices.
* * @see CameraCharacteristics#LENS_INFO_FOCUS_DISTANCE_CALIBRATION */ public static final KeyShortest distance from frontmost surface * of the lens that can be focused correctly.
*If the lens is fixed-focus, this should be * 0.
*/ public static final KeyDimensions of lens shading map.
*The map should be on the order of 30-40 rows and columns, and * must be smaller than 64x64.
*/ public static final KeyThe lens focus distance calibration quality.
*The lens focus distance calibration quality determines the reliability of * focus related metadata entries, i.e. {@link CaptureRequest#LENS_FOCUS_DISTANCE android.lens.focusDistance}, * {@link CaptureResult#LENS_FOCUS_RANGE android.lens.focusRange}, {@link CameraCharacteristics#LENS_INFO_HYPERFOCAL_DISTANCE android.lens.info.hyperfocalDistance}, and * {@link CameraCharacteristics#LENS_INFO_MINIMUM_FOCUS_DISTANCE android.lens.info.minimumFocusDistance}.
* * @see CaptureRequest#LENS_FOCUS_DISTANCE * @see CaptureResult#LENS_FOCUS_RANGE * @see CameraCharacteristics#LENS_INFO_HYPERFOCAL_DISTANCE * @see CameraCharacteristics#LENS_INFO_MINIMUM_FOCUS_DISTANCE * @see #LENS_INFO_FOCUS_DISTANCE_CALIBRATION_UNCALIBRATED * @see #LENS_INFO_FOCUS_DISTANCE_CALIBRATION_APPROXIMATE * @see #LENS_INFO_FOCUS_DISTANCE_CALIBRATION_CALIBRATED */ public static final KeyDirection the camera faces relative to * device screen
* @see #LENS_FACING_FRONT * @see #LENS_FACING_BACK */ public static final KeyIf set to 1, the HAL will always split result * metadata for a single capture into multiple buffers, * returned using multiple process_capture_result calls.
*Does not need to be listed in static * metadata. Support for partial results will be reworked in * future versions of camera service. This quirk will stop * working at that point; DO NOT USE without careful * consideration of future support.
*Optional - This value may be {@code null} on some devices.
* @hide */ public static final KeyThe maximum numbers of different types of output streams * that can be configured and used simultaneously by a camera device.
*This is a 3 element tuple that contains the max number of output simultaneous
* streams for raw sensor, processed (and uncompressed), and JPEG formats respectively.
* For example, if max raw sensor format output stream number is 1, max YUV streams
* number is 3, and max JPEG stream number is 2, then this tuple should be (1, 3, 2)
.
This lists the upper bound of the number of output streams supported by * the camera device. Using more streams simultaneously may require more hardware and * CPU resources that will consume more power. The image format for a output stream can * be any supported format provided by {@link CameraCharacteristics#SCALER_AVAILABLE_FORMATS android.scaler.availableFormats}. The formats * defined in {@link CameraCharacteristics#SCALER_AVAILABLE_FORMATS android.scaler.availableFormats} can be catergorized into the 3 stream types * as below:
*The maximum numbers of any type of input streams * that can be configured and used simultaneously by a camera device.
*When set to 0, it means no input stream is supported.
*The image format for a input stream can be any supported * format provided by * {@link CameraCharacteristics#SCALER_AVAILABLE_INPUT_OUTPUT_FORMATS_MAP android.scaler.availableInputOutputFormatsMap}. When using an * input stream, there must be at least one output stream * configured to to receive the reprocessed images.
*For example, for Zero Shutter Lag (ZSL) still capture use case, the input * stream image format will be RAW_OPAQUE, the associated output stream image format * should be JPEG.
* * @see CameraCharacteristics#SCALER_AVAILABLE_INPUT_OUTPUT_FORMATS_MAP */ public static final KeySpecifies the number of maximum pipeline stages a frame * has to go through from when it's exposed to when it's available * to the framework.
*A typical minimum value for this is 2 (one stage to expose, * one stage to readout) from the sensor. The ISP then usually adds * its own stages to do custom HW processing. Further stages may be * added by SW processing.
*Depending on what settings are used (e.g. YUV, JPEG) and what * processing is enabled (e.g. face detection), the actual pipeline * depth (specified by {@link CaptureResult#REQUEST_PIPELINE_DEPTH android.request.pipelineDepth}) may be less than * the max pipeline depth.
*A pipeline depth of X stages is equivalent to a pipeline latency of * X frame intervals.
*This value will be 8 or less.
* * @see CaptureResult#REQUEST_PIPELINE_DEPTH */ public static final KeyOptional. Defaults to 1. Defines how many sub-components * a result will be composed of.
*In order to combat the pipeline latency, partial results * may be delivered to the application layer from the camera device as * soon as they are available.
*A value of 1 means that partial results are not supported.
*A typical use case for this might be: after requesting an AF lock the * new AF state might be available 50% of the way through the pipeline. * The camera device could then immediately dispatch this state via a * partial result to the framework/application layer, and the rest of * the metadata via later partial results.
*/ public static final KeyList of capabilities that the camera device * advertises as fully supporting.
*A capability is a contract that the camera device makes in order * to be able to satisfy one or more use cases.
*Listing a capability guarantees that the whole set of features * required to support a common use will all be available.
*Using a subset of the functionality provided by an unsupported * capability may be possible on a specific camera device implementation; * to do this query each of android.request.availableRequestKeys, * android.request.availableResultKeys, * android.request.availableCharacteristicsKeys.
*XX: Maybe these should go into {@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL android.info.supportedHardwareLevel} * as a table instead?
*The following capabilities are guaranteed to be available on
* {@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL android.info.supportedHardwareLevel} ==
FULL devices:
Other capabilities may be available on either FULL or LIMITED * devices, but the app. should query this field to be sure.
* * @see CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL * @see #REQUEST_AVAILABLE_CAPABILITIES_BACKWARD_COMPATIBLE * @see #REQUEST_AVAILABLE_CAPABILITIES_OPTIONAL * @see #REQUEST_AVAILABLE_CAPABILITIES_MANUAL_SENSOR * @see #REQUEST_AVAILABLE_CAPABILITIES_GCAM * @see #REQUEST_AVAILABLE_CAPABILITIES_ZSL * @see #REQUEST_AVAILABLE_CAPABILITIES_DNG */ public static final KeyA list of all keys that the camera device has available * to use with CaptureRequest.
*Attempting to set a key into a CaptureRequest that is not * listed here will result in an invalid request and will be rejected * by the camera device.
*This field can be used to query the feature set of a camera device * at a more granular level than capabilities. This is especially * important for optional keys that are not listed under any capability * in {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES android.request.availableCapabilities}.
*TODO: This should be used by #getAvailableCaptureRequestKeys.
* * @see CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES * @hide */ public static final KeyA list of all keys that the camera device has available * to use with CaptureResult.
*Attempting to get a key from a CaptureResult that is not
* listed here will always return a null
value. Getting a key from
* a CaptureResult that is listed here must never return a null
* value.
The following keys may return null
unless they are enabled:
(Those sometimes-null keys should nevertheless be listed here * if they are available.)
*This field can be used to query the feature set of a camera device * at a more granular level than capabilities. This is especially * important for optional keys that are not listed under any capability * in {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES android.request.availableCapabilities}.
*TODO: This should be used by #getAvailableCaptureResultKeys.
* * @see CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES * @see CaptureResult#STATISTICS_LENS_SHADING_MAP * @see CaptureRequest#STATISTICS_LENS_SHADING_MAP_MODE * @hide */ public static final KeyA list of all keys that the camera device has available * to use with CameraCharacteristics.
*This entry follows the same rules as * android.request.availableResultKeys (except that it applies for * CameraCharacteristics instead of CaptureResult). See above for more * details.
*TODO: This should be used by CameraCharacteristics#getKeys.
* @hide */ public static final KeyThe list of image formats that are supported by this * camera device for output streams.
*All camera devices will support JPEG and YUV_420_888 formats.
*When set to YUV_420_888, application can access the YUV420 data directly.
*/ public static final KeyThe minimum frame duration that is supported * for each resolution in {@link CameraCharacteristics#SCALER_AVAILABLE_JPEG_SIZES android.scaler.availableJpegSizes}.
*This corresponds to the minimum steady-state frame duration when only * that JPEG stream is active and captured in a burst, with all * processing (typically in android.*.mode) set to FAST.
*When multiple streams are configured, the minimum * frame duration will be >= max(individual stream min * durations)
* * @see CameraCharacteristics#SCALER_AVAILABLE_JPEG_SIZES */ public static final KeyThe JPEG resolutions that are supported by this camera device.
*The resolutions are listed as (width, height)
pairs. All camera devices will support
* sensor maximum resolution (defined by {@link CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE android.sensor.info.activeArraySize}).
The maximum ratio between active area width * and crop region width, or between active area height and * crop region height, if the crop region height is larger * than width
*/ public static final KeyFor each available processed output size (defined in * {@link CameraCharacteristics#SCALER_AVAILABLE_PROCESSED_SIZES android.scaler.availableProcessedSizes}), this property lists the * minimum supportable frame duration for that size.
*This should correspond to the frame duration when only that processed * stream is active, with all processing (typically in android.*.mode) * set to FAST.
*When multiple streams are configured, the minimum frame duration will * be >= max(individual stream min durations).
* * @see CameraCharacteristics#SCALER_AVAILABLE_PROCESSED_SIZES */ public static final KeyThe resolutions available for use with * processed output streams, such as YV12, NV12, and * platform opaque YUV/RGB streams to the GPU or video * encoders.
*The resolutions are listed as (width, height)
pairs.
For a given use case, the actual maximum supported resolution * may be lower than what is listed here, depending on the destination * Surface for the image data. For example, for recording video, * the video encoder chosen may have a maximum size limit (e.g. 1080p) * smaller than what the camera (e.g. maximum resolution is 3264x2448) * can provide.
*Please reference the documentation for the image data destination to * check if it limits the maximum size for image data.
*/ public static final KeyThe mapping of image formats that are supported by this * camera device for input streams, to their corresponding output formats.
*All camera devices with at least 1 * {@link CameraCharacteristics#REQUEST_MAX_NUM_INPUT_STREAMS android.request.maxNumInputStreams} will have at least one * available input format.
*The camera device will support the following map of formats, * if its dependent capability is supported:
*Input Format | *Output Format | *Capability | *
---|---|---|
RAW_OPAQUE | *JPEG | *ZSL | *
RAW_OPAQUE | *YUV_420_888 | *ZSL | *
RAW_OPAQUE | *RAW16 | *DNG | *
RAW16 | *YUV_420_888 | *DNG | *
RAW16 | *JPEG | *DNG | *
For ZSL-capable camera devices, using the RAW_OPAQUE format * as either input or output will never hurt maximum frame rate (i.e. * {@link CameraCharacteristics#SCALER_AVAILABLE_STALL_DURATIONS android.scaler.availableStallDurations} will not have RAW_OPAQUE).
*Attempting to configure an input stream with output streams not * listed as available in this map is not valid.
*TODO: Add java type mapping for this property.
* * @see CameraCharacteristics#REQUEST_MAX_NUM_INPUT_STREAMS * @see CameraCharacteristics#SCALER_AVAILABLE_STALL_DURATIONS */ public static final KeyThe available stream configurations that this * camera device supports * (i.e. format, width, height, output/input stream).
*The configurations are listed as (format, width, height, input?)
* tuples.
All camera devices will support sensor maximum resolution (defined by * {@link CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE android.sensor.info.activeArraySize}) for the JPEG format.
*For a given use case, the actual maximum supported resolution * may be lower than what is listed here, depending on the destination * Surface for the image data. For example, for recording video, * the video encoder chosen may have a maximum size limit (e.g. 1080p) * smaller than what the camera (e.g. maximum resolution is 3264x2448) * can provide.
*Please reference the documentation for the image data destination to * check if it limits the maximum size for image data.
*Not all output formats may be supported in a configuration with * an input stream of a particular format. For more details, see * {@link CameraCharacteristics#SCALER_AVAILABLE_INPUT_OUTPUT_FORMATS_MAP android.scaler.availableInputOutputFormatsMap}.
*The following table describes the minimum required output stream * configurations based on the hardware level * ({@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL android.info.supportedHardwareLevel}):
*Format | *Size | *Hardware Level | *Notes | *
---|---|---|---|
JPEG | *{@link CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE android.sensor.info.activeArraySize} | *Any | ** |
JPEG | *1920x1080 (1080p) | *Any | *if 1080p <= activeArraySize | *
JPEG | *1280x720 (720) | *Any | *if 720p <= activeArraySize | *
JPEG | *640x480 (480p) | *Any | *if 480p <= activeArraySize | *
JPEG | *320x240 (240p) | *Any | *if 240p <= activeArraySize | *
YUV_420_888 | *all output sizes available for JPEG | *FULL | ** |
YUV_420_888 | *all output sizes available for JPEG, up to the maximum video size | *LIMITED | ** |
IMPLEMENTATION_DEFINED | *same as YUV_420_888 | *Any | ** |
Refer to {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES android.request.availableCapabilities} for additional * mandatory stream configurations on a per-capability basis.
* * @see CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL * @see CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES * @see CameraCharacteristics#SCALER_AVAILABLE_INPUT_OUTPUT_FORMATS_MAP * @see CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE * @see #SCALER_AVAILABLE_STREAM_CONFIGURATIONS_OUTPUT * @see #SCALER_AVAILABLE_STREAM_CONFIGURATIONS_INPUT */ public static final KeyThis lists the minimum frame duration for each * format/size combination.
*This should correspond to the frame duration when only that * stream is active, with all processing (typically in android.*.mode) * set to either OFF or FAST.
*When multiple streams are used in a request, the minimum frame * duration will be max(individual stream min durations).
*The minimum frame duration of a stream (of a particular format, size) * is the same regardless of whether the stream is input or output.
*See {@link CaptureRequest#SENSOR_FRAME_DURATION android.sensor.frameDuration} and * {@link CameraCharacteristics#SCALER_AVAILABLE_STALL_DURATIONS android.scaler.availableStallDurations} for more details about * calculating the max frame rate.
* * @see CameraCharacteristics#SCALER_AVAILABLE_STALL_DURATIONS * @see CaptureRequest#SENSOR_FRAME_DURATION */ public static final KeyThis lists the maximum stall duration for each * format/size combination.
*A stall duration is how much extra time would get added * to the normal minimum frame duration for a repeating request * that has streams with non-zero stall.
*For example, consider JPEG captures which have the following * characteristics:
*In other words, using a repeating YUV request would result * in a steady frame rate (let's say it's 30 FPS). If a single * JPEG request is submitted periodically, the frame rate will stay * at 30 FPS (as long as we wait for the previous JPEG to return each * time). If we try to submit a repeating YUV + JPEG request, then * the frame rate will drop from 30 FPS.
*In general, submitting a new request with a non-0 stall time * stream will not cause a frame rate drop unless there are still * outstanding buffers for that stream from previous requests.
*Submitting a repeating request with streams (call this S
)
* is the same as setting the minimum frame duration from
* the normal minimum frame duration corresponding to S
, added with
* the maximum stall duration for S
.
If interleaving requests with and without a stall duration, * a request will stall by the maximum of the remaining times * for each can-stall stream with outstanding buffers.
*This means that a stalling request will not have an exposure start * until the stall has completed.
*This should correspond to the stall duration when only that stream is * active, with all processing (typically in android.*.mode) set to FAST * or OFF. Setting any of the processing modes to HIGH_QUALITY * effectively results in an indeterminate stall duration for all * streams in a request (the regular stall calculation rules are * ignored).
*The following formats may always have a stall duration:
*The following formats will never have a stall duration:
*All other formats may or may not have an allowed stall duration on * a per-capability basis; refer to {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES android.request.availableCapabilities} * for more details.
*See {@link CaptureRequest#SENSOR_FRAME_DURATION android.sensor.frameDuration} for more information about * calculating the max frame rate (absent stalls).
* * @see CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES * @see CaptureRequest#SENSOR_FRAME_DURATION */ public static final KeyArea of raw data which corresponds to only * active pixels.
*It is smaller or equal to * sensor full pixel array, which could include the black calibration pixels.
*/ public static final KeyRange of valid sensitivities
*/ public static final KeyRange of valid exposure * times used by {@link CaptureRequest#SENSOR_EXPOSURE_TIME android.sensor.exposureTime}.
* * @see CaptureRequest#SENSOR_EXPOSURE_TIME */ public static final KeyMaximum possible frame duration (minimum frame * rate).
*The largest possible {@link CaptureRequest#SENSOR_FRAME_DURATION android.sensor.frameDuration} * that will be accepted by the camera device. Attempting to use * frame durations beyond the maximum will result in the frame duration * being clipped to the maximum. See that control * for a full definition of frame durations.
*Refer to * {@link CameraCharacteristics#SCALER_AVAILABLE_PROCESSED_MIN_DURATIONS android.scaler.availableProcessedMinDurations}, * {@link CameraCharacteristics#SCALER_AVAILABLE_JPEG_MIN_DURATIONS android.scaler.availableJpegMinDurations}, and * android.scaler.availableRawMinDurations for the minimum * frame duration values.
* * @see CameraCharacteristics#SCALER_AVAILABLE_JPEG_MIN_DURATIONS * @see CameraCharacteristics#SCALER_AVAILABLE_PROCESSED_MIN_DURATIONS * @see CaptureRequest#SENSOR_FRAME_DURATION */ public static final KeyThe physical dimensions of the full pixel * array
*Needed for FOV calculation for old API
*/ public static final KeyDimensions of full pixel array, possibly * including black calibration pixels.
*Maximum output resolution for raw format must * match this in * {@link CameraCharacteristics#SCALER_AVAILABLE_STREAM_CONFIGURATIONS android.scaler.availableStreamConfigurations}.
* * @see CameraCharacteristics#SCALER_AVAILABLE_STREAM_CONFIGURATIONS */ public static final KeyMaximum raw value output by sensor.
*This specifies the fully-saturated encoding level for the raw * sample values from the sensor. This is typically caused by the * sensor becoming highly non-linear or clipping. The minimum for * each channel is specified by the offset in the * {@link CameraCharacteristics#SENSOR_BLACK_LEVEL_PATTERN android.sensor.blackLevelPattern} tag.
*The white level is typically determined either by sensor bit depth * (10-14 bits is expected), or by the point where the sensor response * becomes too non-linear to be useful. The default value for this is * maximum representable value for a 16-bit raw sample (2^16 - 1).
* * @see CameraCharacteristics#SENSOR_BLACK_LEVEL_PATTERN */ public static final KeyGain factor from electrons to raw units when * ISO=100
*Optional - This value may be {@code null} on some devices.
*Full capability - * Present on all camera devices that report being {@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL_FULL HARDWARE_LEVEL_FULL} devices in the * {@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL android.info.supportedHardwareLevel} key
* * @see CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL */ public static final KeyA fixed black level offset for each of the color filter arrangement * (CFA) mosaic channels.
*This tag specifies the zero light value for each of the CFA mosaic * channels in the camera sensor. The maximal value output by the * sensor is represented by the value in {@link CameraCharacteristics#SENSOR_INFO_WHITE_LEVEL android.sensor.info.whiteLevel}.
*The values are given in row-column scan order, with the first value * corresponding to the element of the CFA in row=0, column=0.
*Optional - This value may be {@code null} on some devices.
* * @see CameraCharacteristics#SENSOR_INFO_WHITE_LEVEL */ public static final KeyMaximum sensitivity that is implemented * purely through analog gain.
*For {@link CaptureRequest#SENSOR_SENSITIVITY android.sensor.sensitivity} values less than or * equal to this, all applied gain must be analog. For * values above this, the gain applied can be a mix of analog and * digital.
*Optional - This value may be {@code null} on some devices.
*Full capability - * Present on all camera devices that report being {@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL_FULL HARDWARE_LEVEL_FULL} devices in the * {@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL android.info.supportedHardwareLevel} key
* * @see CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL * @see CaptureRequest#SENSOR_SENSITIVITY */ public static final KeyClockwise angle through which the output * image needs to be rotated to be upright on the device * screen in its native orientation. Also defines the * direction of rolling shutter readout, which is from top * to bottom in the sensor's coordinate system
*/ public static final KeyThe number of input samples for each dimension of * {@link CaptureResult#SENSOR_PROFILE_HUE_SAT_MAP android.sensor.profileHueSatMap}.
*The number of input samples for the hue, saturation, and value * dimension of {@link CaptureResult#SENSOR_PROFILE_HUE_SAT_MAP android.sensor.profileHueSatMap}. The order of the * dimensions given is hue, saturation, value; where hue is the 0th * element.
*Optional - This value may be {@code null} on some devices.
* * @see CaptureResult#SENSOR_PROFILE_HUE_SAT_MAP */ public static final KeyOptional. Defaults to [OFF]. Lists the supported test * pattern modes for {@link CaptureRequest#SENSOR_TEST_PATTERN_MODE android.sensor.testPatternMode}.
*Optional - This value may be {@code null} on some devices.
* * @see CaptureRequest#SENSOR_TEST_PATTERN_MODE */ public static final KeyWhich face detection modes are available, * if any
*OFF means face detection is disabled, it must * be included in the list.
*SIMPLE means the device supports the * android.statistics.faceRectangles and * android.statistics.faceScores outputs.
*FULL means the device additionally supports the * android.statistics.faceIds and * android.statistics.faceLandmarks outputs.
*/ public static final KeyMaximum number of simultaneously detectable * faces
*/ public static final KeyMaximum number of supported points in the * tonemap curve that can be used for {@link CaptureRequest#TONEMAP_CURVE_RED android.tonemap.curveRed}, or * {@link CaptureRequest#TONEMAP_CURVE_GREEN android.tonemap.curveGreen}, or {@link CaptureRequest#TONEMAP_CURVE_BLUE android.tonemap.curveBlue}.
*If the actual number of points provided by the application (in * android.tonemap.curve*) is less than max, the camera device will * resample the curve to its internal representation, using linear * interpolation.
*The output curves in the result metadata may have a different number * of points than the input curves, and will represent the actual * hardware curves used as closely as possible when linearly interpolated.
* * @see CaptureRequest#TONEMAP_CURVE_BLUE * @see CaptureRequest#TONEMAP_CURVE_GREEN * @see CaptureRequest#TONEMAP_CURVE_RED */ public static final KeyA list of camera LEDs that are available on this system.
* @see #LED_AVAILABLE_LEDS_TRANSMIT * @hide */ public static final KeyGenerally classifies the overall set of the camera device functionality.
*Camera devices will come in two flavors: LIMITED and FULL.
*A FULL device has the most support possible and will enable the * widest range of use cases such as:
*A LIMITED device may have some or none of the above characteristics. * To find out more refer to {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES android.request.availableCapabilities}.
* * @see CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES * @see #INFO_SUPPORTED_HARDWARE_LEVEL_LIMITED * @see #INFO_SUPPORTED_HARDWARE_LEVEL_FULL */ public static final KeyThe maximum number of frames that can occur after a request * (different than the previous) has been submitted, and before the * result's state becomes synchronized (by setting * android.sync.frameNumber to a non-negative value).
*This defines the maximum distance (in number of metadata results), * between android.sync.frameNumber and the equivalent * android.request.frameCount.
*In other words this acts as an upper boundary for how many frames * must occur before the camera device knows for a fact that the new * submitted camera settings have been applied in outgoing frames.
*For example if the distance was 2,
*initial request = X (repeating)
* request1 = X
* request2 = Y
* request3 = Y
* request4 = Y
*
* where requestN has frameNumber N, and the first of the repeating
* initial request's has frameNumber F (and F < 1).
*
* initial result = X' + { android.sync.frameNumber == F }
* result1 = X' + { android.sync.frameNumber == F }
* result2 = X' + { android.sync.frameNumber == CONVERGING }
* result3 = X' + { android.sync.frameNumber == CONVERGING }
* result4 = X' + { android.sync.frameNumber == 2 }
*
* where resultN has frameNumber N.
*
* Since result4
has a frameNumber == 4
and
* android.sync.frameNumber == 2
, the distance is clearly
* 4 - 2 = 2
.