/* * Copyright (C) 2013 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package android.hardware.camera2; import android.hardware.camera2.CaptureResult.Key; import android.hardware.camera2.impl.CameraMetadataNative; import android.hardware.camera2.utils.TypeReference; import android.util.Rational; import java.util.Collections; import java.util.List; /** *

The properties describing a * {@link CameraDevice CameraDevice}.

* *

These properties are fixed for a given CameraDevice, and can be queried * through the {@link CameraManager CameraManager} * interface with {@link CameraManager#getCameraCharacteristics}.

* *

{@link CameraCharacteristics} objects are immutable.

* * @see CameraDevice * @see CameraManager */ public final class CameraCharacteristics extends CameraMetadata> { /** * A {@code Key} is used to do camera characteristics field lookups with * {@link CameraCharacteristics#get}. * *

For example, to get the stream configuration map: *

     * StreamConfigurationMap map = cameraCharacteristics.get(
     *      CameraCharacteristics.SCALER_STREAM_CONFIGURATION_MAP);
     * 
*

* *

To enumerate over all possible keys for {@link CameraCharacteristics}, see * {@link CameraCharacteristics#getKeys()}.

* * @see CameraCharacteristics#get * @see CameraCharacteristics#getKeys() */ public static final class Key { private final CameraMetadataNative.Key mKey; /** * Visible for testing and vendor extensions only. * * @hide */ public Key(String name, Class type) { mKey = new CameraMetadataNative.Key(name, type); } /** * Visible for testing and vendor extensions only. * * @hide */ public Key(String name, TypeReference typeReference) { mKey = new CameraMetadataNative.Key(name, typeReference); } /** * Return a camelCase, period separated name formatted like: * {@code "root.section[.subsections].name"}. * *

Built-in keys exposed by the Android SDK are always prefixed with {@code "android."}; * keys that are device/platform-specific are prefixed with {@code "com."}.

* *

For example, {@code CameraCharacteristics.SCALER_STREAM_CONFIGURATION_MAP} would * have a name of {@code "android.scaler.streamConfigurationMap"}; whereas a device * specific key might look like {@code "com.google.nexus.data.private"}.

* * @return String representation of the key name */ public String getName() { return mKey.getName(); } /** * {@inheritDoc} */ @Override public final int hashCode() { return mKey.hashCode(); } /** * {@inheritDoc} */ @SuppressWarnings("unchecked") @Override public final boolean equals(Object o) { return o instanceof Key && ((Key)o).mKey.equals(mKey); } /** * Visible for CameraMetadataNative implementation only; do not use. * * TODO: Make this private or remove it altogether. * * @hide */ public CameraMetadataNative.Key getNativeKey() { return mKey; } @SuppressWarnings({ "unused", "unchecked" }) private Key(CameraMetadataNative.Key nativeKey) { mKey = (CameraMetadataNative.Key) nativeKey; } } private final CameraMetadataNative mProperties; private List> mAvailableRequestKeys; private List> mAvailableResultKeys; /** * Takes ownership of the passed-in properties object * @hide */ public CameraCharacteristics(CameraMetadataNative properties) { mProperties = CameraMetadataNative.move(properties); } /** * Returns a copy of the underlying {@link CameraMetadataNative}. * @hide */ public CameraMetadataNative getNativeCopy() { return new CameraMetadataNative(mProperties); } /** * Get a camera characteristics field value. * *

The field definitions can be * found in {@link CameraCharacteristics}.

* *

Querying the value for the same key more than once will return a value * which is equal to the previous queried value.

* * @throws IllegalArgumentException if the key was not valid * * @param key The characteristics field to read. * @return The value of that key, or {@code null} if the field is not set. */ public T get(Key key) { return mProperties.get(key); } /** * {@inheritDoc} * @hide */ @SuppressWarnings("unchecked") @Override protected T getProtected(Key key) { return (T) mProperties.get(key); } /** * {@inheritDoc} * @hide */ @SuppressWarnings("unchecked") @Override protected Class> getKeyClass() { Object thisClass = Key.class; return (Class>)thisClass; } /** * {@inheritDoc} */ @Override public List> getKeys() { // Force the javadoc for this function to show up on the CameraCharacteristics page return super.getKeys(); } /** * Returns the list of keys supported by this {@link CameraDevice} for querying * with a {@link CaptureRequest}. * *

The list returned is not modifiable, so any attempts to modify it will throw * a {@code UnsupportedOperationException}.

* *

Each key is only listed once in the list. The order of the keys is undefined.

* *

Note that there is no {@code getAvailableCameraCharacteristicsKeys()} -- use * {@link #getKeys()} instead.

* * @return List of keys supported by this CameraDevice for CaptureRequests. */ @SuppressWarnings({"unchecked"}) public List> getAvailableCaptureRequestKeys() { if (mAvailableRequestKeys == null) { Object crKey = CaptureRequest.Key.class; Class> crKeyTyped = (Class>)crKey; mAvailableRequestKeys = Collections.unmodifiableList( getAvailableKeyList(CaptureRequest.class, crKeyTyped)); } return mAvailableRequestKeys; } /** * Returns the list of keys supported by this {@link CameraDevice} for querying * with a {@link CaptureResult}. * *

The list returned is not modifiable, so any attempts to modify it will throw * a {@code UnsupportedOperationException}.

* *

Each key is only listed once in the list. The order of the keys is undefined.

* *

Note that there is no {@code getAvailableCameraCharacteristicsKeys()} -- use * {@link #getKeys()} instead.

* * @return List of keys supported by this CameraDevice for CaptureResults. */ @SuppressWarnings({"unchecked"}) public List> getAvailableCaptureResultKeys() { if (mAvailableResultKeys == null) { Object crKey = CaptureResult.Key.class; Class> crKeyTyped = (Class>)crKey; mAvailableResultKeys = Collections.unmodifiableList( getAvailableKeyList(CaptureResult.class, crKeyTyped)); } return mAvailableResultKeys; } /** * Returns the list of keys supported by this {@link CameraDevice} by metadataClass. * *

The list returned is not modifiable, so any attempts to modify it will throw * a {@code UnsupportedOperationException}.

* *

Each key is only listed once in the list. The order of the keys is undefined.

* * @param metadataClass The subclass of CameraMetadata that you want to get the keys for. * @param keyClass The class of the metadata key, e.g. CaptureRequest.Key.class * * @return List of keys supported by this CameraDevice for metadataClass. * * @throws IllegalArgumentException if metadataClass is not a subclass of CameraMetadata */ private List getAvailableKeyList(Class metadataClass, Class keyClass) { if (metadataClass.equals(CameraMetadata.class)) { throw new AssertionError( "metadataClass must be a strict subclass of CameraMetadata"); } else if (!CameraMetadata.class.isAssignableFrom(metadataClass)) { throw new AssertionError( "metadataClass must be a subclass of CameraMetadata"); } List staticKeyList = CameraCharacteristics.getKeysStatic( metadataClass, keyClass, /*instance*/null); return Collections.unmodifiableList(staticKeyList); } /*@O~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~ * The key entries below this point are generated from metadata * definitions in /system/media/camera/docs. Do not modify by hand or * modify the comment blocks at the start or end. *~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~*/ /** *

The set of auto-exposure antibanding modes that are * supported by this camera device.

*

Not all of the auto-exposure anti-banding modes may be * supported by a given camera device. This field lists the * valid anti-banding modes that the application may request * for this camera device; they must include AUTO.

*/ public static final Key CONTROL_AE_AVAILABLE_ANTIBANDING_MODES = new Key("android.control.aeAvailableAntibandingModes", int[].class); /** *

The set of auto-exposure modes that are supported by this * camera device.

*

Not all the auto-exposure modes may be supported by a * given camera device, especially if no flash unit is * available. This entry lists the valid modes for * {@link CaptureRequest#CONTROL_AE_MODE android.control.aeMode} for this camera device.

*

All camera devices support ON, and all camera devices with * flash units support ON_AUTO_FLASH and * ON_ALWAYS_FLASH.

*

FULL mode camera devices always support OFF mode, * which enables application control of camera exposure time, * sensitivity, and frame duration.

* * @see CaptureRequest#CONTROL_AE_MODE */ public static final Key CONTROL_AE_AVAILABLE_MODES = new Key("android.control.aeAvailableModes", int[].class); /** *

List of frame rate ranges supported by the * auto-exposure (AE) algorithm/hardware

*/ public static final Key[]> CONTROL_AE_AVAILABLE_TARGET_FPS_RANGES = new Key[]>("android.control.aeAvailableTargetFpsRanges", new TypeReference[]>() {{ }}); /** *

Maximum and minimum exposure compensation * setting, in counts of * {@link CameraCharacteristics#CONTROL_AE_COMPENSATION_STEP android.control.aeCompensationStep}.

* * @see CameraCharacteristics#CONTROL_AE_COMPENSATION_STEP */ public static final Key> CONTROL_AE_COMPENSATION_RANGE = new Key>("android.control.aeCompensationRange", new TypeReference>() {{ }}); /** *

Smallest step by which exposure compensation * can be changed

*/ public static final Key CONTROL_AE_COMPENSATION_STEP = new Key("android.control.aeCompensationStep", Rational.class); /** *

List of auto-focus (AF) modes that can be * selected with {@link CaptureRequest#CONTROL_AF_MODE android.control.afMode}.

*

Not all the auto-focus modes may be supported by a * given camera device. This entry lists the valid modes for * {@link CaptureRequest#CONTROL_AF_MODE android.control.afMode} for this camera device.

*

All camera devices will support OFF mode, and all camera devices with * adjustable focuser units ({@link CameraCharacteristics#LENS_INFO_MINIMUM_FOCUS_DISTANCE android.lens.info.minimumFocusDistance} > 0) * will support AUTO mode.

* * @see CaptureRequest#CONTROL_AF_MODE * @see CameraCharacteristics#LENS_INFO_MINIMUM_FOCUS_DISTANCE */ public static final Key CONTROL_AF_AVAILABLE_MODES = new Key("android.control.afAvailableModes", int[].class); /** *

List containing the subset of color effects * specified in {@link CaptureRequest#CONTROL_EFFECT_MODE android.control.effectMode} that is supported by * this device.

*

This list contains the color effect modes that can be applied to * images produced by the camera device. Only modes that have * been fully implemented for the current device may be included here. * Implementations are not expected to be consistent across all devices. * If no color effect modes are available for a device, this should * simply be set to OFF.

*

A color effect will only be applied if * {@link CaptureRequest#CONTROL_MODE android.control.mode} != OFF.

* * @see CaptureRequest#CONTROL_EFFECT_MODE * @see CaptureRequest#CONTROL_MODE */ public static final Key CONTROL_AVAILABLE_EFFECTS = new Key("android.control.availableEffects", int[].class); /** *

List containing a subset of scene modes * specified in {@link CaptureRequest#CONTROL_SCENE_MODE android.control.sceneMode}.

*

This list contains scene modes that can be set for the camera device. * Only scene modes that have been fully implemented for the * camera device may be included here. Implementations are not expected * to be consistent across all devices. If no scene modes are supported * by the camera device, this will be set to [DISABLED].

* * @see CaptureRequest#CONTROL_SCENE_MODE */ public static final Key CONTROL_AVAILABLE_SCENE_MODES = new Key("android.control.availableSceneModes", int[].class); /** *

List of video stabilization modes that can * be supported

*/ public static final Key CONTROL_AVAILABLE_VIDEO_STABILIZATION_MODES = new Key("android.control.availableVideoStabilizationModes", int[].class); /** *

The set of auto-white-balance modes ({@link CaptureRequest#CONTROL_AWB_MODE android.control.awbMode}) * that are supported by this camera device.

*

Not all the auto-white-balance modes may be supported by a * given camera device. This entry lists the valid modes for * {@link CaptureRequest#CONTROL_AWB_MODE android.control.awbMode} for this camera device.

*

All camera devices will support ON mode.

*

FULL mode camera devices will always support OFF mode, * which enables application control of white balance, by using * {@link CaptureRequest#COLOR_CORRECTION_TRANSFORM android.colorCorrection.transform} and {@link CaptureRequest#COLOR_CORRECTION_GAINS android.colorCorrection.gains}({@link CaptureRequest#COLOR_CORRECTION_MODE android.colorCorrection.mode} must be set to TRANSFORM_MATRIX).

* * @see CaptureRequest#COLOR_CORRECTION_GAINS * @see CaptureRequest#COLOR_CORRECTION_MODE * @see CaptureRequest#COLOR_CORRECTION_TRANSFORM * @see CaptureRequest#CONTROL_AWB_MODE */ public static final Key CONTROL_AWB_AVAILABLE_MODES = new Key("android.control.awbAvailableModes", int[].class); /** *

List of the maximum number of regions that can be used for metering in * auto-exposure (AE), auto-white balance (AWB), and auto-focus (AF); * this corresponds to the the maximum number of elements in * {@link CaptureRequest#CONTROL_AE_REGIONS android.control.aeRegions}, {@link CaptureRequest#CONTROL_AWB_REGIONS android.control.awbRegions}, * and {@link CaptureRequest#CONTROL_AF_REGIONS android.control.afRegions}.

* * @see CaptureRequest#CONTROL_AE_REGIONS * @see CaptureRequest#CONTROL_AF_REGIONS * @see CaptureRequest#CONTROL_AWB_REGIONS * @hide */ public static final Key CONTROL_MAX_REGIONS = new Key("android.control.maxRegions", int[].class); /** *

List of the maximum number of regions that can be used for metering in * auto-exposure (AE); * this corresponds to the the maximum number of elements in * {@link CaptureRequest#CONTROL_AE_REGIONS android.control.aeRegions}.

* * @see CaptureRequest#CONTROL_AE_REGIONS */ public static final Key CONTROL_MAX_REGIONS_AE = new Key("android.control.maxRegionsAe", int.class); /** *

List of the maximum number of regions that can be used for metering in * auto-white balance (AWB); * this corresponds to the the maximum number of elements in * {@link CaptureRequest#CONTROL_AWB_REGIONS android.control.awbRegions}.

* * @see CaptureRequest#CONTROL_AWB_REGIONS */ public static final Key CONTROL_MAX_REGIONS_AWB = new Key("android.control.maxRegionsAwb", int.class); /** *

List of the maximum number of regions that can be used for metering in * auto-focus (AF); * this corresponds to the the maximum number of elements in * {@link CaptureRequest#CONTROL_AF_REGIONS android.control.afRegions}.

* * @see CaptureRequest#CONTROL_AF_REGIONS */ public static final Key CONTROL_MAX_REGIONS_AF = new Key("android.control.maxRegionsAf", int.class); /** *

The set of edge enhancement modes supported by this camera device.

*

This tag lists the valid modes for {@link CaptureRequest#EDGE_MODE android.edge.mode}.

*

Full-capability camera devices must always support OFF and FAST.

* * @see CaptureRequest#EDGE_MODE */ public static final Key EDGE_AVAILABLE_EDGE_MODES = new Key("android.edge.availableEdgeModes", int[].class); /** *

Whether this camera device has a * flash.

*

If no flash, none of the flash controls do * anything. All other metadata should return 0.

*/ public static final Key FLASH_INFO_AVAILABLE = new Key("android.flash.info.available", boolean.class); /** *

The set of hot pixel correction modes that are supported by this * camera device.

*

This tag lists valid modes for {@link CaptureRequest#HOT_PIXEL_MODE android.hotPixel.mode}.

*

FULL mode camera devices will always support FAST.

* * @see CaptureRequest#HOT_PIXEL_MODE */ public static final Key HOT_PIXEL_AVAILABLE_HOT_PIXEL_MODES = new Key("android.hotPixel.availableHotPixelModes", int[].class); /** *

Supported resolutions for the JPEG thumbnail.

*

Below condiditions will be satisfied for this size list:

*
    *
  • The sizes will be sorted by increasing pixel area (width x height). * If several resolutions have the same area, they will be sorted by increasing width.
  • *
  • The aspect ratio of the largest thumbnail size will be same as the * aspect ratio of largest JPEG output size in android.scaler.availableStreamConfigurations. * The largest size is defined as the size that has the largest pixel area * in a given size list.
  • *
  • Each output JPEG size in android.scaler.availableStreamConfigurations will have at least * one corresponding size that has the same aspect ratio in availableThumbnailSizes, * and vice versa.
  • *
  • All non (0, 0) sizes will have non-zero widths and heights.
  • *
*/ public static final Key JPEG_AVAILABLE_THUMBNAIL_SIZES = new Key("android.jpeg.availableThumbnailSizes", android.util.Size[].class); /** *

List of supported aperture * values.

*

If the camera device doesn't support variable apertures, * listed value will be the fixed aperture.

*

If the camera device supports variable apertures, the aperture value * in this list will be sorted in ascending order.

*/ public static final Key LENS_INFO_AVAILABLE_APERTURES = new Key("android.lens.info.availableApertures", float[].class); /** *

List of supported neutral density filter values for * {@link CaptureRequest#LENS_FILTER_DENSITY android.lens.filterDensity}.

*

If changing {@link CaptureRequest#LENS_FILTER_DENSITY android.lens.filterDensity} is not supported, * availableFilterDensities must contain only 0. Otherwise, this * list contains only the exact filter density values available on * this camera device.

* * @see CaptureRequest#LENS_FILTER_DENSITY */ public static final Key LENS_INFO_AVAILABLE_FILTER_DENSITIES = new Key("android.lens.info.availableFilterDensities", float[].class); /** *

The available focal lengths for this device for use with * {@link CaptureRequest#LENS_FOCAL_LENGTH android.lens.focalLength}.

*

If optical zoom is not supported, this will only report * a single value corresponding to the static focal length of the * device. Otherwise, this will report every focal length supported * by the device.

* * @see CaptureRequest#LENS_FOCAL_LENGTH */ public static final Key LENS_INFO_AVAILABLE_FOCAL_LENGTHS = new Key("android.lens.info.availableFocalLengths", float[].class); /** *

List containing a subset of the optical image * stabilization (OIS) modes specified in * {@link CaptureRequest#LENS_OPTICAL_STABILIZATION_MODE android.lens.opticalStabilizationMode}.

*

If OIS is not implemented for a given camera device, this will * contain only OFF.

* * @see CaptureRequest#LENS_OPTICAL_STABILIZATION_MODE */ public static final Key LENS_INFO_AVAILABLE_OPTICAL_STABILIZATION = new Key("android.lens.info.availableOpticalStabilization", int[].class); /** *

Optional. Hyperfocal distance for this lens.

*

If the lens is not fixed focus, the camera device will report this * field when {@link CameraCharacteristics#LENS_INFO_FOCUS_DISTANCE_CALIBRATION android.lens.info.focusDistanceCalibration} is APPROXIMATE or CALIBRATED.

*

Optional - This value may be {@code null} on some devices.

* * @see CameraCharacteristics#LENS_INFO_FOCUS_DISTANCE_CALIBRATION */ public static final Key LENS_INFO_HYPERFOCAL_DISTANCE = new Key("android.lens.info.hyperfocalDistance", float.class); /** *

Shortest distance from frontmost surface * of the lens that can be focused correctly.

*

If the lens is fixed-focus, this should be * 0.

*/ public static final Key LENS_INFO_MINIMUM_FOCUS_DISTANCE = new Key("android.lens.info.minimumFocusDistance", float.class); /** *

Dimensions of lens shading map.

*

The map should be on the order of 30-40 rows and columns, and * must be smaller than 64x64.

* @hide */ public static final Key LENS_INFO_SHADING_MAP_SIZE = new Key("android.lens.info.shadingMapSize", android.util.Size.class); /** *

The lens focus distance calibration quality.

*

The lens focus distance calibration quality determines the reliability of * focus related metadata entries, i.e. {@link CaptureRequest#LENS_FOCUS_DISTANCE android.lens.focusDistance}, * {@link CaptureResult#LENS_FOCUS_RANGE android.lens.focusRange}, {@link CameraCharacteristics#LENS_INFO_HYPERFOCAL_DISTANCE android.lens.info.hyperfocalDistance}, and * {@link CameraCharacteristics#LENS_INFO_MINIMUM_FOCUS_DISTANCE android.lens.info.minimumFocusDistance}.

* * @see CaptureRequest#LENS_FOCUS_DISTANCE * @see CaptureResult#LENS_FOCUS_RANGE * @see CameraCharacteristics#LENS_INFO_HYPERFOCAL_DISTANCE * @see CameraCharacteristics#LENS_INFO_MINIMUM_FOCUS_DISTANCE * @see #LENS_INFO_FOCUS_DISTANCE_CALIBRATION_UNCALIBRATED * @see #LENS_INFO_FOCUS_DISTANCE_CALIBRATION_APPROXIMATE * @see #LENS_INFO_FOCUS_DISTANCE_CALIBRATION_CALIBRATED */ public static final Key LENS_INFO_FOCUS_DISTANCE_CALIBRATION = new Key("android.lens.info.focusDistanceCalibration", int.class); /** *

Direction the camera faces relative to * device screen.

* @see #LENS_FACING_FRONT * @see #LENS_FACING_BACK */ public static final Key LENS_FACING = new Key("android.lens.facing", int.class); /** *

The set of noise reduction modes supported by this camera device.

*

This tag lists the valid modes for {@link CaptureRequest#NOISE_REDUCTION_MODE android.noiseReduction.mode}.

*

Full-capability camera devices must always support OFF and FAST.

* * @see CaptureRequest#NOISE_REDUCTION_MODE */ public static final Key NOISE_REDUCTION_AVAILABLE_NOISE_REDUCTION_MODES = new Key("android.noiseReduction.availableNoiseReductionModes", int[].class); /** *

If set to 1, the HAL will always split result * metadata for a single capture into multiple buffers, * returned using multiple process_capture_result calls.

*

Does not need to be listed in static * metadata. Support for partial results will be reworked in * future versions of camera service. This quirk will stop * working at that point; DO NOT USE without careful * consideration of future support.

*

Optional - This value may be {@code null} on some devices.

* @deprecated * @hide */ @Deprecated public static final Key QUIRKS_USE_PARTIAL_RESULT = new Key("android.quirks.usePartialResult", byte.class); /** *

The maximum numbers of different types of output streams * that can be configured and used simultaneously by a camera device.

*

This is a 3 element tuple that contains the max number of output simultaneous * streams for raw sensor, processed (but not stalling), and processed (and stalling) * formats respectively. For example, assuming that JPEG is typically a processed and * stalling stream, if max raw sensor format output stream number is 1, max YUV streams * number is 3, and max JPEG stream number is 2, then this tuple should be (1, 3, 2).

*

This lists the upper bound of the number of output streams supported by * the camera device. Using more streams simultaneously may require more hardware and * CPU resources that will consume more power. The image format for an output stream can * be any supported format provided by android.scaler.availableStreamConfigurations. * The formats defined in android.scaler.availableStreamConfigurations can be catergorized * into the 3 stream types as below:

*
    *
  • Processed (but stalling): any non-RAW format with a stallDurations > 0. * Typically JPEG format (ImageFormat#JPEG).
  • *
  • Raw formats: ImageFormat#RAW_SENSOR and ImageFormat#RAW_OPAQUE.
  • *
  • Processed (but not-stalling): any non-RAW format without a stall duration. * Typically ImageFormat#YUV_420_888, ImageFormat#NV21, ImageFormat#YV12.
  • *
* @hide */ public static final Key REQUEST_MAX_NUM_OUTPUT_STREAMS = new Key("android.request.maxNumOutputStreams", int[].class); /** *

The maximum numbers of different types of output streams * that can be configured and used simultaneously by a camera device * for any RAW formats.

*

This value contains the max number of output simultaneous * streams from the raw sensor.

*

This lists the upper bound of the number of output streams supported by * the camera device. Using more streams simultaneously may require more hardware and * CPU resources that will consume more power. The image format for this kind of an output stream can * be any RAW and supported format provided by {@link CameraCharacteristics#SCALER_STREAM_CONFIGURATION_MAP android.scaler.streamConfigurationMap}.

*

In particular, a RAW format is typically one of:

*
    *
  • ImageFormat#RAW_SENSOR
  • *
  • Opaque RAW
  • *
* * @see CameraCharacteristics#SCALER_STREAM_CONFIGURATION_MAP */ public static final Key REQUEST_MAX_NUM_OUTPUT_RAW = new Key("android.request.maxNumOutputRaw", int.class); /** *

The maximum numbers of different types of output streams * that can be configured and used simultaneously by a camera device * for any processed (but not-stalling) formats.

*

This value contains the max number of output simultaneous * streams for any processed (but not-stalling) formats.

*

This lists the upper bound of the number of output streams supported by * the camera device. Using more streams simultaneously may require more hardware and * CPU resources that will consume more power. The image format for this kind of an output stream can * be any non-RAW and supported format provided by {@link CameraCharacteristics#SCALER_STREAM_CONFIGURATION_MAP android.scaler.streamConfigurationMap}.

*

Processed (but not-stalling) is defined as any non-RAW format without a stall duration. * Typically:

*
    *
  • ImageFormat#YUV_420_888
  • *
  • ImageFormat#NV21
  • *
  • ImageFormat#YV12
  • *
  • Implementation-defined formats, i.e. StreamConfiguration#isOutputSupportedFor(Class)
  • *
*

For full guarantees, query StreamConfigurationMap#getOutputStallDuration with * a processed format -- it will return 0 for a non-stalling stream.

* * @see CameraCharacteristics#SCALER_STREAM_CONFIGURATION_MAP */ public static final Key REQUEST_MAX_NUM_OUTPUT_PROC = new Key("android.request.maxNumOutputProc", int.class); /** *

The maximum numbers of different types of output streams * that can be configured and used simultaneously by a camera device * for any processed (and stalling) formats.

*

This value contains the max number of output simultaneous * streams for any processed (but not-stalling) formats.

*

This lists the upper bound of the number of output streams supported by * the camera device. Using more streams simultaneously may require more hardware and * CPU resources that will consume more power. The image format for this kind of an output stream can * be any non-RAW and supported format provided by {@link CameraCharacteristics#SCALER_STREAM_CONFIGURATION_MAP android.scaler.streamConfigurationMap}.

*

A processed and stalling format is defined as any non-RAW format with a stallDurations > 0. * Typically only the JPEG format (ImageFormat#JPEG)

*

For full guarantees, query StreamConfigurationMap#getOutputStallDuration with * a processed format -- it will return a non-0 value for a stalling stream.

* * @see CameraCharacteristics#SCALER_STREAM_CONFIGURATION_MAP */ public static final Key REQUEST_MAX_NUM_OUTPUT_PROC_STALLING = new Key("android.request.maxNumOutputProcStalling", int.class); /** *

The maximum numbers of any type of input streams * that can be configured and used simultaneously by a camera device.

*

When set to 0, it means no input stream is supported.

*

The image format for a input stream can be any supported * format provided by * android.scaler.availableInputOutputFormatsMap. When using an * input stream, there must be at least one output stream * configured to to receive the reprocessed images.

*

For example, for Zero Shutter Lag (ZSL) still capture use case, the input * stream image format will be RAW_OPAQUE, the associated output stream image format * should be JPEG.

* @hide */ public static final Key REQUEST_MAX_NUM_INPUT_STREAMS = new Key("android.request.maxNumInputStreams", int.class); /** *

Specifies the number of maximum pipeline stages a frame * has to go through from when it's exposed to when it's available * to the framework.

*

A typical minimum value for this is 2 (one stage to expose, * one stage to readout) from the sensor. The ISP then usually adds * its own stages to do custom HW processing. Further stages may be * added by SW processing.

*

Depending on what settings are used (e.g. YUV, JPEG) and what * processing is enabled (e.g. face detection), the actual pipeline * depth (specified by {@link CaptureResult#REQUEST_PIPELINE_DEPTH android.request.pipelineDepth}) may be less than * the max pipeline depth.

*

A pipeline depth of X stages is equivalent to a pipeline latency of * X frame intervals.

*

This value will be 8 or less.

* * @see CaptureResult#REQUEST_PIPELINE_DEPTH */ public static final Key REQUEST_PIPELINE_MAX_DEPTH = new Key("android.request.pipelineMaxDepth", byte.class); /** *

Defines how many sub-components * a result will be composed of.

*

In order to combat the pipeline latency, partial results * may be delivered to the application layer from the camera device as * soon as they are available.

*

Optional; defaults to 1. A value of 1 means that partial * results are not supported, and only the final TotalCaptureResult will * be produced by the camera device.

*

A typical use case for this might be: after requesting an * auto-focus (AF) lock the new AF state might be available 50% * of the way through the pipeline. The camera device could * then immediately dispatch this state via a partial result to * the application, and the rest of the metadata via later * partial results.

*/ public static final Key REQUEST_PARTIAL_RESULT_COUNT = new Key("android.request.partialResultCount", int.class); /** *

List of capabilities that the camera device * advertises as fully supporting.

*

A capability is a contract that the camera device makes in order * to be able to satisfy one or more use cases.

*

Listing a capability guarantees that the whole set of features * required to support a common use will all be available.

*

Using a subset of the functionality provided by an unsupported * capability may be possible on a specific camera device implementation; * to do this query each of android.request.availableRequestKeys, * android.request.availableResultKeys, * android.request.availableCharacteristicsKeys.

*

The following capabilities are guaranteed to be available on * {@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL android.info.supportedHardwareLevel} == FULL devices:

*
    *
  • MANUAL_SENSOR
  • *
  • MANUAL_POST_PROCESSING
  • *
*

Other capabilities may be available on either FULL or LIMITED * devices, but the application should query this field to be sure.

* * @see CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL * @see #REQUEST_AVAILABLE_CAPABILITIES_MANUAL_SENSOR * @see #REQUEST_AVAILABLE_CAPABILITIES_MANUAL_POST_PROCESSING * @see #REQUEST_AVAILABLE_CAPABILITIES_DNG */ public static final Key REQUEST_AVAILABLE_CAPABILITIES = new Key("android.request.availableCapabilities", int[].class); /** *

A list of all keys that the camera device has available * to use with CaptureRequest.

*

Attempting to set a key into a CaptureRequest that is not * listed here will result in an invalid request and will be rejected * by the camera device.

*

This field can be used to query the feature set of a camera device * at a more granular level than capabilities. This is especially * important for optional keys that are not listed under any capability * in {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES android.request.availableCapabilities}.

* * @see CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES * @hide */ public static final Key REQUEST_AVAILABLE_REQUEST_KEYS = new Key("android.request.availableRequestKeys", int[].class); /** *

A list of all keys that the camera device has available * to use with CaptureResult.

*

Attempting to get a key from a CaptureResult that is not * listed here will always return a null value. Getting a key from * a CaptureResult that is listed here must never return a null * value.

*

The following keys may return null unless they are enabled:

*
    *
  • android.statistics.lensShadingMap (non-null iff {@link CaptureRequest#STATISTICS_LENS_SHADING_MAP_MODE android.statistics.lensShadingMapMode} == ON)
  • *
*

(Those sometimes-null keys should nevertheless be listed here * if they are available.)

*

This field can be used to query the feature set of a camera device * at a more granular level than capabilities. This is especially * important for optional keys that are not listed under any capability * in {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES android.request.availableCapabilities}.

* * @see CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES * @see CaptureRequest#STATISTICS_LENS_SHADING_MAP_MODE * @hide */ public static final Key REQUEST_AVAILABLE_RESULT_KEYS = new Key("android.request.availableResultKeys", int[].class); /** *

A list of all keys that the camera device has available * to use with CameraCharacteristics.

*

This entry follows the same rules as * android.request.availableResultKeys (except that it applies for * CameraCharacteristics instead of CaptureResult). See above for more * details.

* @hide */ public static final Key REQUEST_AVAILABLE_CHARACTERISTICS_KEYS = new Key("android.request.availableCharacteristicsKeys", int[].class); /** *

The list of image formats that are supported by this * camera device for output streams.

*

All camera devices will support JPEG and YUV_420_888 formats.

*

When set to YUV_420_888, application can access the YUV420 data directly.

* @deprecated * @hide */ @Deprecated public static final Key SCALER_AVAILABLE_FORMATS = new Key("android.scaler.availableFormats", int[].class); /** *

The minimum frame duration that is supported * for each resolution in android.scaler.availableJpegSizes.

*

This corresponds to the minimum steady-state frame duration when only * that JPEG stream is active and captured in a burst, with all * processing (typically in android.*.mode) set to FAST.

*

When multiple streams are configured, the minimum * frame duration will be >= max(individual stream min * durations)

* @deprecated * @hide */ @Deprecated public static final Key SCALER_AVAILABLE_JPEG_MIN_DURATIONS = new Key("android.scaler.availableJpegMinDurations", long[].class); /** *

The JPEG resolutions that are supported by this camera device.

*

The resolutions are listed as (width, height) pairs. All camera devices will support * sensor maximum resolution (defined by {@link CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE android.sensor.info.activeArraySize}).

* * @see CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE * @deprecated * @hide */ @Deprecated public static final Key SCALER_AVAILABLE_JPEG_SIZES = new Key("android.scaler.availableJpegSizes", android.util.Size[].class); /** *

The maximum ratio between both active area width * and crop region width, and active area height and * crop region height.

*

This represents the maximum amount of zooming possible by * the camera device, or equivalently, the minimum cropping * window size.

*

Crop regions that have a width or height that is smaller * than this ratio allows will be rounded up to the minimum * allowed size by the camera device.

*/ public static final Key SCALER_AVAILABLE_MAX_DIGITAL_ZOOM = new Key("android.scaler.availableMaxDigitalZoom", float.class); /** *

For each available processed output size (defined in * android.scaler.availableProcessedSizes), this property lists the * minimum supportable frame duration for that size.

*

This should correspond to the frame duration when only that processed * stream is active, with all processing (typically in android.*.mode) * set to FAST.

*

When multiple streams are configured, the minimum frame duration will * be >= max(individual stream min durations).

* @deprecated * @hide */ @Deprecated public static final Key SCALER_AVAILABLE_PROCESSED_MIN_DURATIONS = new Key("android.scaler.availableProcessedMinDurations", long[].class); /** *

The resolutions available for use with * processed output streams, such as YV12, NV12, and * platform opaque YUV/RGB streams to the GPU or video * encoders.

*

The resolutions are listed as (width, height) pairs.

*

For a given use case, the actual maximum supported resolution * may be lower than what is listed here, depending on the destination * Surface for the image data. For example, for recording video, * the video encoder chosen may have a maximum size limit (e.g. 1080p) * smaller than what the camera (e.g. maximum resolution is 3264x2448) * can provide.

*

Please reference the documentation for the image data destination to * check if it limits the maximum size for image data.

* @deprecated * @hide */ @Deprecated public static final Key SCALER_AVAILABLE_PROCESSED_SIZES = new Key("android.scaler.availableProcessedSizes", android.util.Size[].class); /** *

The mapping of image formats that are supported by this * camera device for input streams, to their corresponding output formats.

*

All camera devices with at least 1 * android.request.maxNumInputStreams will have at least one * available input format.

*

The camera device will support the following map of formats, * if its dependent capability is supported:

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
Input FormatOutput FormatCapability
RAW_OPAQUEJPEGZSL
RAW_OPAQUEYUV_420_888ZSL
RAW_OPAQUERAW16DNG
RAW16YUV_420_888DNG
RAW16JPEGDNG
*

For ZSL-capable camera devices, using the RAW_OPAQUE format * as either input or output will never hurt maximum frame rate (i.e. * StreamConfigurationMap#getOutputStallDuration(int,Size) * for a format = RAW_OPAQUE is always 0).

*

Attempting to configure an input stream with output streams not * listed as available in this map is not valid.

*

TODO: typedef to ReprocessFormatMap

* @hide */ public static final Key SCALER_AVAILABLE_INPUT_OUTPUT_FORMATS_MAP = new Key("android.scaler.availableInputOutputFormatsMap", int[].class); /** *

The available stream configurations that this * camera device supports * (i.e. format, width, height, output/input stream).

*

The configurations are listed as (format, width, height, input?) * tuples.

*

For a given use case, the actual maximum supported resolution * may be lower than what is listed here, depending on the destination * Surface for the image data. For example, for recording video, * the video encoder chosen may have a maximum size limit (e.g. 1080p) * smaller than what the camera (e.g. maximum resolution is 3264x2448) * can provide.

*

Please reference the documentation for the image data destination to * check if it limits the maximum size for image data.

*

Not all output formats may be supported in a configuration with * an input stream of a particular format. For more details, see * android.scaler.availableInputOutputFormatsMap.

*

The following table describes the minimum required output stream * configurations based on the hardware level * ({@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL android.info.supportedHardwareLevel}):

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
FormatSizeHardware LevelNotes
JPEG{@link CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE android.sensor.info.activeArraySize}Any
JPEG1920x1080 (1080p)Anyif 1080p <= activeArraySize
JPEG1280x720 (720)Anyif 720p <= activeArraySize
JPEG640x480 (480p)Anyif 480p <= activeArraySize
JPEG320x240 (240p)Anyif 240p <= activeArraySize
YUV_420_888all output sizes available for JPEGFULL
YUV_420_888all output sizes available for JPEG, up to the maximum video sizeLIMITED
IMPLEMENTATION_DEFINEDsame as YUV_420_888Any
*

Refer to {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES android.request.availableCapabilities} for additional * mandatory stream configurations on a per-capability basis.

* * @see CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL * @see CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES * @see CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE * @hide */ public static final Key SCALER_AVAILABLE_STREAM_CONFIGURATIONS = new Key("android.scaler.availableStreamConfigurations", android.hardware.camera2.params.StreamConfiguration[].class); /** *

This lists the minimum frame duration for each * format/size combination.

*

This should correspond to the frame duration when only that * stream is active, with all processing (typically in android.*.mode) * set to either OFF or FAST.

*

When multiple streams are used in a request, the minimum frame * duration will be max(individual stream min durations).

*

The minimum frame duration of a stream (of a particular format, size) * is the same regardless of whether the stream is input or output.

*

See {@link CaptureRequest#SENSOR_FRAME_DURATION android.sensor.frameDuration} and * android.scaler.availableStallDurations for more details about * calculating the max frame rate.

*

(Keep in sync with * StreamConfigurationMap#getOutputMinFrameDuration)

* * @see CaptureRequest#SENSOR_FRAME_DURATION * @hide */ public static final Key SCALER_AVAILABLE_MIN_FRAME_DURATIONS = new Key("android.scaler.availableMinFrameDurations", android.hardware.camera2.params.StreamConfigurationDuration[].class); /** *

This lists the maximum stall duration for each * format/size combination.

*

A stall duration is how much extra time would get added * to the normal minimum frame duration for a repeating request * that has streams with non-zero stall.

*

For example, consider JPEG captures which have the following * characteristics:

*
    *
  • JPEG streams act like processed YUV streams in requests for which * they are not included; in requests in which they are directly * referenced, they act as JPEG streams. This is because supporting a * JPEG stream requires the underlying YUV data to always be ready for * use by a JPEG encoder, but the encoder will only be used (and impact * frame duration) on requests that actually reference a JPEG stream.
  • *
  • The JPEG processor can run concurrently to the rest of the camera * pipeline, but cannot process more than 1 capture at a time.
  • *
*

In other words, using a repeating YUV request would result * in a steady frame rate (let's say it's 30 FPS). If a single * JPEG request is submitted periodically, the frame rate will stay * at 30 FPS (as long as we wait for the previous JPEG to return each * time). If we try to submit a repeating YUV + JPEG request, then * the frame rate will drop from 30 FPS.

*

In general, submitting a new request with a non-0 stall time * stream will not cause a frame rate drop unless there are still * outstanding buffers for that stream from previous requests.

*

Submitting a repeating request with streams (call this S) * is the same as setting the minimum frame duration from * the normal minimum frame duration corresponding to S, added with * the maximum stall duration for S.

*

If interleaving requests with and without a stall duration, * a request will stall by the maximum of the remaining times * for each can-stall stream with outstanding buffers.

*

This means that a stalling request will not have an exposure start * until the stall has completed.

*

This should correspond to the stall duration when only that stream is * active, with all processing (typically in android.*.mode) set to FAST * or OFF. Setting any of the processing modes to HIGH_QUALITY * effectively results in an indeterminate stall duration for all * streams in a request (the regular stall calculation rules are * ignored).

*

The following formats may always have a stall duration:

*
    *
  • JPEG
  • *
  • RAW16
  • *
*

The following formats will never have a stall duration:

*
    *
  • YUV_420_888
  • *
  • IMPLEMENTATION_DEFINED
  • *
*

All other formats may or may not have an allowed stall duration on * a per-capability basis; refer to {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES android.request.availableCapabilities} * for more details.

*

See {@link CaptureRequest#SENSOR_FRAME_DURATION android.sensor.frameDuration} for more information about * calculating the max frame rate (absent stalls).

*

(Keep up to date with * StreamConfigurationMap#getOutputStallDuration(int, Size) )

* * @see CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES * @see CaptureRequest#SENSOR_FRAME_DURATION * @hide */ public static final Key SCALER_AVAILABLE_STALL_DURATIONS = new Key("android.scaler.availableStallDurations", android.hardware.camera2.params.StreamConfigurationDuration[].class); /** *

The available stream configurations that this * camera device supports; also includes the minimum frame durations * and the stall durations for each format/size combination.

*

All camera devices will support sensor maximum resolution (defined by * {@link CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE android.sensor.info.activeArraySize}) for the JPEG format.

*

For a given use case, the actual maximum supported resolution * may be lower than what is listed here, depending on the destination * Surface for the image data. For example, for recording video, * the video encoder chosen may have a maximum size limit (e.g. 1080p) * smaller than what the camera (e.g. maximum resolution is 3264x2448) * can provide.

*

Please reference the documentation for the image data destination to * check if it limits the maximum size for image data.

*

The following table describes the minimum required output stream * configurations based on the hardware level * ({@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL android.info.supportedHardwareLevel}):

* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
FormatSizeHardware LevelNotes
JPEG{@link CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE android.sensor.info.activeArraySize}Any
JPEG1920x1080 (1080p)Anyif 1080p <= activeArraySize
JPEG1280x720 (720)Anyif 720p <= activeArraySize
JPEG640x480 (480p)Anyif 480p <= activeArraySize
JPEG320x240 (240p)Anyif 240p <= activeArraySize
YUV_420_888all output sizes available for JPEGFULL
YUV_420_888all output sizes available for JPEG, up to the maximum video sizeLIMITED
IMPLEMENTATION_DEFINEDsame as YUV_420_888Any
*

Refer to {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES android.request.availableCapabilities} for additional * mandatory stream configurations on a per-capability basis.

* * @see CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL * @see CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES * @see CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE */ public static final Key SCALER_STREAM_CONFIGURATION_MAP = new Key("android.scaler.streamConfigurationMap", android.hardware.camera2.params.StreamConfigurationMap.class); /** *

The crop type that this camera device supports.

*

When passing a non-centered crop region ({@link CaptureRequest#SCALER_CROP_REGION android.scaler.cropRegion}) to a camera * device that only supports CENTER_ONLY cropping, the camera device will move the * crop region to the center of the sensor active array ({@link CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE android.sensor.info.activeArraySize}) * and keep the crop region width and height unchanged. The camera device will return the * final used crop region in metadata result {@link CaptureRequest#SCALER_CROP_REGION android.scaler.cropRegion}.

*

Camera devices that support FREEFORM cropping will support any crop region that * is inside of the active array. The camera device will apply the same crop region and * return the final used crop region in capture result metadata {@link CaptureRequest#SCALER_CROP_REGION android.scaler.cropRegion}.

*

FULL capability devices ({@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL android.info.supportedHardwareLevel} == FULL) will support * FREEFORM cropping.

* * @see CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL * @see CaptureRequest#SCALER_CROP_REGION * @see CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE * @see #SCALER_CROPPING_TYPE_CENTER_ONLY * @see #SCALER_CROPPING_TYPE_FREEFORM */ public static final Key SCALER_CROPPING_TYPE = new Key("android.scaler.croppingType", int.class); /** *

Area of raw data which corresponds to only * active pixels.

*

It is smaller or equal to * sensor full pixel array, which could include the black calibration pixels.

*/ public static final Key SENSOR_INFO_ACTIVE_ARRAY_SIZE = new Key("android.sensor.info.activeArraySize", android.graphics.Rect.class); /** *

Range of valid sensitivities.

*

The minimum and maximum valid values for the * {@link CaptureRequest#SENSOR_SENSITIVITY android.sensor.sensitivity} control.

*

The values are the standard ISO sensitivity values, * as defined in ISO 12232:2006.

* * @see CaptureRequest#SENSOR_SENSITIVITY */ public static final Key> SENSOR_INFO_SENSITIVITY_RANGE = new Key>("android.sensor.info.sensitivityRange", new TypeReference>() {{ }}); /** *

The arrangement of color filters on sensor; * represents the colors in the top-left 2x2 section of * the sensor, in reading order.

* @see #SENSOR_INFO_COLOR_FILTER_ARRANGEMENT_RGGB * @see #SENSOR_INFO_COLOR_FILTER_ARRANGEMENT_GRBG * @see #SENSOR_INFO_COLOR_FILTER_ARRANGEMENT_GBRG * @see #SENSOR_INFO_COLOR_FILTER_ARRANGEMENT_BGGR * @see #SENSOR_INFO_COLOR_FILTER_ARRANGEMENT_RGB */ public static final Key SENSOR_INFO_COLOR_FILTER_ARRANGEMENT = new Key("android.sensor.info.colorFilterArrangement", int.class); /** *

Range of valid exposure * times used by {@link CaptureRequest#SENSOR_EXPOSURE_TIME android.sensor.exposureTime}.

* * @see CaptureRequest#SENSOR_EXPOSURE_TIME */ public static final Key> SENSOR_INFO_EXPOSURE_TIME_RANGE = new Key>("android.sensor.info.exposureTimeRange", new TypeReference>() {{ }}); /** *

Maximum possible frame duration (minimum frame * rate).

*

The largest possible {@link CaptureRequest#SENSOR_FRAME_DURATION android.sensor.frameDuration} * that will be accepted by the camera device. Attempting to use * frame durations beyond the maximum will result in the frame duration * being clipped to the maximum. See that control * for a full definition of frame durations.

*

Refer to * StreamConfigurationMap#getOutputMinFrameDuration(int,Size) * for the minimum frame duration values.

* * @see CaptureRequest#SENSOR_FRAME_DURATION */ public static final Key SENSOR_INFO_MAX_FRAME_DURATION = new Key("android.sensor.info.maxFrameDuration", long.class); /** *

The physical dimensions of the full pixel * array.

*

This is the physical size of the sensor pixel * array defined by {@link CameraCharacteristics#SENSOR_INFO_PIXEL_ARRAY_SIZE android.sensor.info.pixelArraySize}.

* * @see CameraCharacteristics#SENSOR_INFO_PIXEL_ARRAY_SIZE */ public static final Key SENSOR_INFO_PHYSICAL_SIZE = new Key("android.sensor.info.physicalSize", android.util.SizeF.class); /** *

Dimensions of full pixel array, possibly * including black calibration pixels.

*

The pixel count of the full pixel array, * which covers {@link CameraCharacteristics#SENSOR_INFO_PHYSICAL_SIZE android.sensor.info.physicalSize} area.

*

If a camera device supports raw sensor formats, either this * or {@link CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE android.sensor.info.activeArraySize} is the maximum output * raw size listed in {@link CameraCharacteristics#SCALER_STREAM_CONFIGURATION_MAP android.scaler.streamConfigurationMap}. * If a size corresponding to pixelArraySize is listed, the resulting * raw sensor image will include black pixels.

* * @see CameraCharacteristics#SCALER_STREAM_CONFIGURATION_MAP * @see CameraCharacteristics#SENSOR_INFO_ACTIVE_ARRAY_SIZE * @see CameraCharacteristics#SENSOR_INFO_PHYSICAL_SIZE */ public static final Key SENSOR_INFO_PIXEL_ARRAY_SIZE = new Key("android.sensor.info.pixelArraySize", android.util.Size.class); /** *

Maximum raw value output by sensor.

*

This specifies the fully-saturated encoding level for the raw * sample values from the sensor. This is typically caused by the * sensor becoming highly non-linear or clipping. The minimum for * each channel is specified by the offset in the * {@link CameraCharacteristics#SENSOR_BLACK_LEVEL_PATTERN android.sensor.blackLevelPattern} tag.

*

The white level is typically determined either by sensor bit depth * (8-14 bits is expected), or by the point where the sensor response * becomes too non-linear to be useful. The default value for this is * maximum representable value for a 16-bit raw sample (2^16 - 1).

* * @see CameraCharacteristics#SENSOR_BLACK_LEVEL_PATTERN */ public static final Key SENSOR_INFO_WHITE_LEVEL = new Key("android.sensor.info.whiteLevel", int.class); /** *

The sensor timestamp calibration quality.

*

The sensor timestamp calibration quality determines the reliability of * {@link CaptureResult#SENSOR_TIMESTAMP android.sensor.timestamp} provided by the camera device.

* * @see CaptureResult#SENSOR_TIMESTAMP * @see #SENSOR_INFO_TIMESTAMP_CALIBRATION_UNCALIBRATED * @see #SENSOR_INFO_TIMESTAMP_CALIBRATION_CALIBRATED */ public static final Key SENSOR_INFO_TIMESTAMP_CALIBRATION = new Key("android.sensor.info.timestampCalibration", int.class); /** *

The standard reference illuminant used as the scene light source when * calculating the {@link CameraCharacteristics#SENSOR_COLOR_TRANSFORM1 android.sensor.colorTransform1}, * {@link CameraCharacteristics#SENSOR_CALIBRATION_TRANSFORM1 android.sensor.calibrationTransform1}, and * {@link CameraCharacteristics#SENSOR_FORWARD_MATRIX1 android.sensor.forwardMatrix1} matrices.

*

The values in this tag correspond to the values defined for the * EXIF LightSource tag. These illuminants are standard light sources * that are often used calibrating camera devices.

*

If this tag is present, then {@link CameraCharacteristics#SENSOR_COLOR_TRANSFORM1 android.sensor.colorTransform1}, * {@link CameraCharacteristics#SENSOR_CALIBRATION_TRANSFORM1 android.sensor.calibrationTransform1}, and * {@link CameraCharacteristics#SENSOR_FORWARD_MATRIX1 android.sensor.forwardMatrix1} will also be present.

*

Some devices may choose to provide a second set of calibration * information for improved quality, including * {@link CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT2 android.sensor.referenceIlluminant2} and its corresponding matrices.

* * @see CameraCharacteristics#SENSOR_CALIBRATION_TRANSFORM1 * @see CameraCharacteristics#SENSOR_COLOR_TRANSFORM1 * @see CameraCharacteristics#SENSOR_FORWARD_MATRIX1 * @see CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT2 * @see #SENSOR_REFERENCE_ILLUMINANT1_DAYLIGHT * @see #SENSOR_REFERENCE_ILLUMINANT1_FLUORESCENT * @see #SENSOR_REFERENCE_ILLUMINANT1_TUNGSTEN * @see #SENSOR_REFERENCE_ILLUMINANT1_FLASH * @see #SENSOR_REFERENCE_ILLUMINANT1_FINE_WEATHER * @see #SENSOR_REFERENCE_ILLUMINANT1_CLOUDY_WEATHER * @see #SENSOR_REFERENCE_ILLUMINANT1_SHADE * @see #SENSOR_REFERENCE_ILLUMINANT1_DAYLIGHT_FLUORESCENT * @see #SENSOR_REFERENCE_ILLUMINANT1_DAY_WHITE_FLUORESCENT * @see #SENSOR_REFERENCE_ILLUMINANT1_COOL_WHITE_FLUORESCENT * @see #SENSOR_REFERENCE_ILLUMINANT1_WHITE_FLUORESCENT * @see #SENSOR_REFERENCE_ILLUMINANT1_STANDARD_A * @see #SENSOR_REFERENCE_ILLUMINANT1_STANDARD_B * @see #SENSOR_REFERENCE_ILLUMINANT1_STANDARD_C * @see #SENSOR_REFERENCE_ILLUMINANT1_D55 * @see #SENSOR_REFERENCE_ILLUMINANT1_D65 * @see #SENSOR_REFERENCE_ILLUMINANT1_D75 * @see #SENSOR_REFERENCE_ILLUMINANT1_D50 * @see #SENSOR_REFERENCE_ILLUMINANT1_ISO_STUDIO_TUNGSTEN */ public static final Key SENSOR_REFERENCE_ILLUMINANT1 = new Key("android.sensor.referenceIlluminant1", int.class); /** *

The standard reference illuminant used as the scene light source when * calculating the {@link CameraCharacteristics#SENSOR_COLOR_TRANSFORM2 android.sensor.colorTransform2}, * {@link CameraCharacteristics#SENSOR_CALIBRATION_TRANSFORM2 android.sensor.calibrationTransform2}, and * {@link CameraCharacteristics#SENSOR_FORWARD_MATRIX2 android.sensor.forwardMatrix2} matrices.

*

See {@link CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT1 android.sensor.referenceIlluminant1} for more details. * Valid values for this are the same as those given for the first * reference illuminant.

*

If this tag is present, then {@link CameraCharacteristics#SENSOR_COLOR_TRANSFORM2 android.sensor.colorTransform2}, * {@link CameraCharacteristics#SENSOR_CALIBRATION_TRANSFORM2 android.sensor.calibrationTransform2}, and * {@link CameraCharacteristics#SENSOR_FORWARD_MATRIX2 android.sensor.forwardMatrix2} will also be present.

* * @see CameraCharacteristics#SENSOR_CALIBRATION_TRANSFORM2 * @see CameraCharacteristics#SENSOR_COLOR_TRANSFORM2 * @see CameraCharacteristics#SENSOR_FORWARD_MATRIX2 * @see CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT1 */ public static final Key SENSOR_REFERENCE_ILLUMINANT2 = new Key("android.sensor.referenceIlluminant2", byte.class); /** *

A per-device calibration transform matrix that maps from the * reference sensor colorspace to the actual device sensor colorspace.

*

This matrix is used to correct for per-device variations in the * sensor colorspace, and is used for processing raw buffer data.

*

The matrix is expressed as a 3x3 matrix in row-major-order, and * contains a per-device calibration transform that maps colors * from reference sensor color space (i.e. the "golden module" * colorspace) into this camera device's native sensor color * space under the first reference illuminant * ({@link CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT1 android.sensor.referenceIlluminant1}).

*

Optional - This value may be {@code null} on some devices.

* * @see CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT1 */ public static final Key SENSOR_CALIBRATION_TRANSFORM1 = new Key("android.sensor.calibrationTransform1", android.hardware.camera2.params.ColorSpaceTransform.class); /** *

A per-device calibration transform matrix that maps from the * reference sensor colorspace to the actual device sensor colorspace * (this is the colorspace of the raw buffer data).

*

This matrix is used to correct for per-device variations in the * sensor colorspace, and is used for processing raw buffer data.

*

The matrix is expressed as a 3x3 matrix in row-major-order, and * contains a per-device calibration transform that maps colors * from reference sensor color space (i.e. the "golden module" * colorspace) into this camera device's native sensor color * space under the second reference illuminant * ({@link CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT2 android.sensor.referenceIlluminant2}).

*

This matrix will only be present if the second reference * illuminant is present.

*

Optional - This value may be {@code null} on some devices.

* * @see CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT2 */ public static final Key SENSOR_CALIBRATION_TRANSFORM2 = new Key("android.sensor.calibrationTransform2", android.hardware.camera2.params.ColorSpaceTransform.class); /** *

A matrix that transforms color values from CIE XYZ color space to * reference sensor color space.

*

This matrix is used to convert from the standard CIE XYZ color * space to the reference sensor colorspace, and is used when processing * raw buffer data.

*

The matrix is expressed as a 3x3 matrix in row-major-order, and * contains a color transform matrix that maps colors from the CIE * XYZ color space to the reference sensor color space (i.e. the * "golden module" colorspace) under the first reference illuminant * ({@link CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT1 android.sensor.referenceIlluminant1}).

*

The white points chosen in both the reference sensor color space * and the CIE XYZ colorspace when calculating this transform will * match the standard white point for the first reference illuminant * (i.e. no chromatic adaptation will be applied by this transform).

*

Optional - This value may be {@code null} on some devices.

* * @see CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT1 */ public static final Key SENSOR_COLOR_TRANSFORM1 = new Key("android.sensor.colorTransform1", android.hardware.camera2.params.ColorSpaceTransform.class); /** *

A matrix that transforms color values from CIE XYZ color space to * reference sensor color space.

*

This matrix is used to convert from the standard CIE XYZ color * space to the reference sensor colorspace, and is used when processing * raw buffer data.

*

The matrix is expressed as a 3x3 matrix in row-major-order, and * contains a color transform matrix that maps colors from the CIE * XYZ color space to the reference sensor color space (i.e. the * "golden module" colorspace) under the second reference illuminant * ({@link CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT2 android.sensor.referenceIlluminant2}).

*

The white points chosen in both the reference sensor color space * and the CIE XYZ colorspace when calculating this transform will * match the standard white point for the second reference illuminant * (i.e. no chromatic adaptation will be applied by this transform).

*

This matrix will only be present if the second reference * illuminant is present.

*

Optional - This value may be {@code null} on some devices.

* * @see CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT2 */ public static final Key SENSOR_COLOR_TRANSFORM2 = new Key("android.sensor.colorTransform2", android.hardware.camera2.params.ColorSpaceTransform.class); /** *

A matrix that transforms white balanced camera colors from the reference * sensor colorspace to the CIE XYZ colorspace with a D50 whitepoint.

*

This matrix is used to convert to the standard CIE XYZ colorspace, and * is used when processing raw buffer data.

*

This matrix is expressed as a 3x3 matrix in row-major-order, and contains * a color transform matrix that maps white balanced colors from the * reference sensor color space to the CIE XYZ color space with a D50 white * point.

*

Under the first reference illuminant ({@link CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT1 android.sensor.referenceIlluminant1}) * this matrix is chosen so that the standard white point for this reference * illuminant in the reference sensor colorspace is mapped to D50 in the * CIE XYZ colorspace.

*

Optional - This value may be {@code null} on some devices.

* * @see CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT1 */ public static final Key SENSOR_FORWARD_MATRIX1 = new Key("android.sensor.forwardMatrix1", android.hardware.camera2.params.ColorSpaceTransform.class); /** *

A matrix that transforms white balanced camera colors from the reference * sensor colorspace to the CIE XYZ colorspace with a D50 whitepoint.

*

This matrix is used to convert to the standard CIE XYZ colorspace, and * is used when processing raw buffer data.

*

This matrix is expressed as a 3x3 matrix in row-major-order, and contains * a color transform matrix that maps white balanced colors from the * reference sensor color space to the CIE XYZ color space with a D50 white * point.

*

Under the second reference illuminant ({@link CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT2 android.sensor.referenceIlluminant2}) * this matrix is chosen so that the standard white point for this reference * illuminant in the reference sensor colorspace is mapped to D50 in the * CIE XYZ colorspace.

*

This matrix will only be present if the second reference * illuminant is present.

*

Optional - This value may be {@code null} on some devices.

* * @see CameraCharacteristics#SENSOR_REFERENCE_ILLUMINANT2 */ public static final Key SENSOR_FORWARD_MATRIX2 = new Key("android.sensor.forwardMatrix2", android.hardware.camera2.params.ColorSpaceTransform.class); /** *

A fixed black level offset for each of the color filter arrangement * (CFA) mosaic channels.

*

This tag specifies the zero light value for each of the CFA mosaic * channels in the camera sensor. The maximal value output by the * sensor is represented by the value in {@link CameraCharacteristics#SENSOR_INFO_WHITE_LEVEL android.sensor.info.whiteLevel}.

*

The values are given in row-column scan order, with the first value * corresponding to the element of the CFA in row=0, column=0.

*

Optional - This value may be {@code null} on some devices.

* * @see CameraCharacteristics#SENSOR_INFO_WHITE_LEVEL */ public static final Key SENSOR_BLACK_LEVEL_PATTERN = new Key("android.sensor.blackLevelPattern", int[].class); /** *

Maximum sensitivity that is implemented * purely through analog gain.

*

For {@link CaptureRequest#SENSOR_SENSITIVITY android.sensor.sensitivity} values less than or * equal to this, all applied gain must be analog. For * values above this, the gain applied can be a mix of analog and * digital.

*

Optional - This value may be {@code null} on some devices.

*

Full capability - * Present on all camera devices that report being {@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL_FULL HARDWARE_LEVEL_FULL} devices in the * {@link CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL android.info.supportedHardwareLevel} key

* * @see CameraCharacteristics#INFO_SUPPORTED_HARDWARE_LEVEL * @see CaptureRequest#SENSOR_SENSITIVITY */ public static final Key SENSOR_MAX_ANALOG_SENSITIVITY = new Key("android.sensor.maxAnalogSensitivity", int.class); /** *

Clockwise angle through which the output * image needs to be rotated to be upright on the device * screen in its native orientation. Also defines the * direction of rolling shutter readout, which is from top * to bottom in the sensor's coordinate system

*/ public static final Key SENSOR_ORIENTATION = new Key("android.sensor.orientation", int.class); /** *

Lists the supported sensor test pattern modes for {@link CaptureRequest#SENSOR_TEST_PATTERN_MODE android.sensor.testPatternMode}.

*

Optional. Defaults to [OFF].

*

Optional - This value may be {@code null} on some devices.

* * @see CaptureRequest#SENSOR_TEST_PATTERN_MODE */ public static final Key SENSOR_AVAILABLE_TEST_PATTERN_MODES = new Key("android.sensor.availableTestPatternModes", int[].class); /** *

The face detection modes that are available * for this camera device.

*

OFF is always supported.

*

SIMPLE means the device supports the * android.statistics.faceRectangles and * android.statistics.faceScores outputs.

*

FULL means the device additionally supports the * android.statistics.faceIds and * android.statistics.faceLandmarks outputs.

*/ public static final Key STATISTICS_INFO_AVAILABLE_FACE_DETECT_MODES = new Key("android.statistics.info.availableFaceDetectModes", int[].class); /** *

The maximum number of simultaneously detectable * faces.

*/ public static final Key STATISTICS_INFO_MAX_FACE_COUNT = new Key("android.statistics.info.maxFaceCount", int.class); /** *

The set of hot pixel map output modes supported by this camera device.

*

This tag lists valid output modes for {@link CaptureRequest#STATISTICS_HOT_PIXEL_MAP_MODE android.statistics.hotPixelMapMode}.

*

If no hotpixel map is available for this camera device, this will contain * only OFF. If the hotpixel map is available, this will include both * the ON and OFF options.

* * @see CaptureRequest#STATISTICS_HOT_PIXEL_MAP_MODE */ public static final Key STATISTICS_INFO_AVAILABLE_HOT_PIXEL_MAP_MODES = new Key("android.statistics.info.availableHotPixelMapModes", boolean[].class); /** *

Maximum number of supported points in the * tonemap curve that can be used for {@link CaptureRequest#TONEMAP_CURVE android.tonemap.curve}.

*

If the actual number of points provided by the application (in * {@link CaptureRequest#TONEMAP_CURVE android.tonemap.curve}*) is less than max, the camera device will * resample the curve to its internal representation, using linear * interpolation.

*

The output curves in the result metadata may have a different number * of points than the input curves, and will represent the actual * hardware curves used as closely as possible when linearly interpolated.

* * @see CaptureRequest#TONEMAP_CURVE */ public static final Key TONEMAP_MAX_CURVE_POINTS = new Key("android.tonemap.maxCurvePoints", int.class); /** *

The set of tonemapping modes supported by this camera device.

*

This tag lists the valid modes for {@link CaptureRequest#TONEMAP_MODE android.tonemap.mode}.

*

Full-capability camera devices must always support CONTRAST_CURVE and * FAST.

* * @see CaptureRequest#TONEMAP_MODE */ public static final Key TONEMAP_AVAILABLE_TONE_MAP_MODES = new Key("android.tonemap.availableToneMapModes", int[].class); /** *

A list of camera LEDs that are available on this system.

* @see #LED_AVAILABLE_LEDS_TRANSMIT * @hide */ public static final Key LED_AVAILABLE_LEDS = new Key("android.led.availableLeds", int[].class); /** *

Generally classifies the overall set of the camera device functionality.

*

Camera devices will come in two flavors: LIMITED and FULL.

*

A FULL device has the most support possible and will enable the * widest range of use cases such as:

*
    *
  • 30fps at maximum resolution (== sensor resolution) is preferred, more than 20fps is required.
  • *
  • Per frame control ({@link CameraCharacteristics#SYNC_MAX_LATENCY android.sync.maxLatency} == PER_FRAME_CONTROL)
  • *
  • Manual sensor control ({@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES android.request.availableCapabilities} contains MANUAL_SENSOR)
  • *
  • Manual post-processing control ({@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES android.request.availableCapabilities} contains MANUAL_POST_PROCESSING)
  • *
*

A LIMITED device may have some or none of the above characteristics. * To find out more refer to {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES android.request.availableCapabilities}.

* * @see CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES * @see CameraCharacteristics#SYNC_MAX_LATENCY * @see #INFO_SUPPORTED_HARDWARE_LEVEL_LIMITED * @see #INFO_SUPPORTED_HARDWARE_LEVEL_FULL */ public static final Key INFO_SUPPORTED_HARDWARE_LEVEL = new Key("android.info.supportedHardwareLevel", int.class); /** *

The maximum number of frames that can occur after a request * (different than the previous) has been submitted, and before the * result's state becomes synchronized (by setting * android.sync.frameNumber to a non-negative value).

*

This defines the maximum distance (in number of metadata results), * between android.sync.frameNumber and the equivalent * android.request.frameCount.

*

In other words this acts as an upper boundary for how many frames * must occur before the camera device knows for a fact that the new * submitted camera settings have been applied in outgoing frames.

*

For example if the distance was 2,

*
initial request = X (repeating)
     * request1 = X
     * request2 = Y
     * request3 = Y
     * request4 = Y
     *
     * where requestN has frameNumber N, and the first of the repeating
     * initial request's has frameNumber F (and F < 1).
     *
     * initial result = X' + { android.sync.frameNumber == F }
     * result1 = X' + { android.sync.frameNumber == F }
     * result2 = X' + { android.sync.frameNumber == CONVERGING }
     * result3 = X' + { android.sync.frameNumber == CONVERGING }
     * result4 = X' + { android.sync.frameNumber == 2 }
     *
     * where resultN has frameNumber N.
     * 
*

Since result4 has a frameNumber == 4 and * android.sync.frameNumber == 2, the distance is clearly * 4 - 2 = 2.

* @see #SYNC_MAX_LATENCY_PER_FRAME_CONTROL * @see #SYNC_MAX_LATENCY_UNKNOWN */ public static final Key SYNC_MAX_LATENCY = new Key("android.sync.maxLatency", int.class); /*~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~ * End generated code *~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~@~O@*/ }