1/*
2 * Copyright (C) 2014 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 *      http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17package android.hardware.camera2.params;
18
19import android.graphics.ImageFormat;
20import android.graphics.PixelFormat;
21import android.hardware.camera2.CameraCharacteristics;
22import android.hardware.camera2.CameraDevice;
23import android.hardware.camera2.CameraMetadata;
24import android.hardware.camera2.CaptureRequest;
25import android.hardware.camera2.utils.HashCodeHelpers;
26import android.hardware.camera2.utils.SurfaceUtils;
27import android.hardware.camera2.legacy.LegacyCameraDevice;
28import android.hardware.camera2.legacy.LegacyMetadataMapper;
29import android.view.Surface;
30import android.util.Range;
31import android.util.Size;
32import android.util.SparseIntArray;
33
34import java.util.Arrays;
35import java.util.HashMap;
36import java.util.Objects;
37import java.util.Set;
38
39import static com.android.internal.util.Preconditions.*;
40
41/**
42 * Immutable class to store the available stream
43 * {@link CameraCharacteristics#SCALER_STREAM_CONFIGURATION_MAP configurations} to set up
44 * {@link android.view.Surface Surfaces} for creating a
45 * {@link android.hardware.camera2.CameraCaptureSession capture session} with
46 * {@link android.hardware.camera2.CameraDevice#createCaptureSession}.
47 * <!-- TODO: link to input stream configuration -->
48 *
49 * <p>This is the authoritative list for all <!-- input/ -->output formats (and sizes respectively
50 * for that format) that are supported by a camera device.</p>
51 *
52 * <p>This also contains the minimum frame durations and stall durations for each format/size
53 * combination that can be used to calculate effective frame rate when submitting multiple captures.
54 * </p>
55 *
56 * <p>An instance of this object is available from {@link CameraCharacteristics} using
57 * the {@link CameraCharacteristics#SCALER_STREAM_CONFIGURATION_MAP} key and the
58 * {@link CameraCharacteristics#get} method.</p>
59 *
60 * <pre><code>{@code
61 * CameraCharacteristics characteristics = cameraManager.getCameraCharacteristics(cameraId);
62 * StreamConfigurationMap configs = characteristics.get(
63 *         CameraCharacteristics.SCALER_STREAM_CONFIGURATION_MAP);
64 * }</code></pre>
65 *
66 * @see CameraCharacteristics#SCALER_STREAM_CONFIGURATION_MAP
67 * @see CameraDevice#createCaptureSession
68 */
69public final class StreamConfigurationMap {
70
71    private static final String TAG = "StreamConfigurationMap";
72
73    /**
74     * Create a new {@link StreamConfigurationMap}.
75     *
76     * <p>The array parameters ownership is passed to this object after creation; do not
77     * write to them after this constructor is invoked.</p>
78     *
79     * @param configurations a non-{@code null} array of {@link StreamConfiguration}
80     * @param minFrameDurations a non-{@code null} array of {@link StreamConfigurationDuration}
81     * @param stallDurations a non-{@code null} array of {@link StreamConfigurationDuration}
82     * @param highSpeedVideoConfigurations an array of {@link HighSpeedVideoConfiguration}, null if
83     *        camera device does not support high speed video recording
84     * @param listHighResolution a flag indicating whether the device supports BURST_CAPTURE
85     *        and thus needs a separate list of slow high-resolution output sizes
86     * @throws NullPointerException if any of the arguments except highSpeedVideoConfigurations
87     *         were {@code null} or any subelements were {@code null}
88     *
89     * @hide
90     */
91    public StreamConfigurationMap(
92            StreamConfiguration[] configurations,
93            StreamConfigurationDuration[] minFrameDurations,
94            StreamConfigurationDuration[] stallDurations,
95            StreamConfiguration[] depthConfigurations,
96            StreamConfigurationDuration[] depthMinFrameDurations,
97            StreamConfigurationDuration[] depthStallDurations,
98            HighSpeedVideoConfiguration[] highSpeedVideoConfigurations,
99            ReprocessFormatsMap inputOutputFormatsMap,
100            boolean listHighResolution) {
101
102        if (configurations == null) {
103            // If no color configurations exist, ensure depth ones do
104            checkArrayElementsNotNull(depthConfigurations, "depthConfigurations");
105            mConfigurations = new StreamConfiguration[0];
106            mMinFrameDurations = new StreamConfigurationDuration[0];
107            mStallDurations = new StreamConfigurationDuration[0];
108        } else {
109            mConfigurations = checkArrayElementsNotNull(configurations, "configurations");
110            mMinFrameDurations = checkArrayElementsNotNull(minFrameDurations, "minFrameDurations");
111            mStallDurations = checkArrayElementsNotNull(stallDurations, "stallDurations");
112        }
113
114        mListHighResolution = listHighResolution;
115
116        if (depthConfigurations == null) {
117            mDepthConfigurations = new StreamConfiguration[0];
118            mDepthMinFrameDurations = new StreamConfigurationDuration[0];
119            mDepthStallDurations = new StreamConfigurationDuration[0];
120        } else {
121            mDepthConfigurations = checkArrayElementsNotNull(depthConfigurations,
122                    "depthConfigurations");
123            mDepthMinFrameDurations = checkArrayElementsNotNull(depthMinFrameDurations,
124                    "depthMinFrameDurations");
125            mDepthStallDurations = checkArrayElementsNotNull(depthStallDurations,
126                    "depthStallDurations");
127        }
128
129        if (highSpeedVideoConfigurations == null) {
130            mHighSpeedVideoConfigurations = new HighSpeedVideoConfiguration[0];
131        } else {
132            mHighSpeedVideoConfigurations = checkArrayElementsNotNull(
133                    highSpeedVideoConfigurations, "highSpeedVideoConfigurations");
134        }
135
136        // For each format, track how many sizes there are available to configure
137        for (StreamConfiguration config : mConfigurations) {
138            int fmt = config.getFormat();
139            SparseIntArray map = null;
140            if (config.isOutput()) {
141                mAllOutputFormats.put(fmt, mAllOutputFormats.get(fmt) + 1);
142                long duration = 0;
143                if (mListHighResolution) {
144                    for (StreamConfigurationDuration configurationDuration : mMinFrameDurations) {
145                        if (configurationDuration.getFormat() == fmt &&
146                                configurationDuration.getWidth() == config.getSize().getWidth() &&
147                                configurationDuration.getHeight() == config.getSize().getHeight()) {
148                            duration = configurationDuration.getDuration();
149                            break;
150                        }
151                    }
152                }
153                map = duration <= DURATION_20FPS_NS ?
154                        mOutputFormats : mHighResOutputFormats;
155            } else {
156                map = mInputFormats;
157            }
158            map.put(fmt, map.get(fmt) + 1);
159        }
160
161        // For each depth format, track how many sizes there are available to configure
162        for (StreamConfiguration config : mDepthConfigurations) {
163            if (!config.isOutput()) {
164                // Ignoring input depth configs
165                continue;
166            }
167
168            mDepthOutputFormats.put(config.getFormat(),
169                    mDepthOutputFormats.get(config.getFormat()) + 1);
170        }
171
172        if (configurations != null &&
173                mOutputFormats.indexOfKey(HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED) < 0) {
174            throw new AssertionError(
175                    "At least one stream configuration for IMPLEMENTATION_DEFINED must exist");
176        }
177
178        // For each Size/FPS range, track how many FPS range/Size there are available
179        for (HighSpeedVideoConfiguration config : mHighSpeedVideoConfigurations) {
180            Size size = config.getSize();
181            Range<Integer> fpsRange = config.getFpsRange();
182            Integer fpsRangeCount = mHighSpeedVideoSizeMap.get(size);
183            if (fpsRangeCount == null) {
184                fpsRangeCount = 0;
185            }
186            mHighSpeedVideoSizeMap.put(size, fpsRangeCount + 1);
187            Integer sizeCount = mHighSpeedVideoFpsRangeMap.get(fpsRange);
188            if (sizeCount == null) {
189                sizeCount = 0;
190            }
191            mHighSpeedVideoFpsRangeMap.put(fpsRange, sizeCount + 1);
192        }
193
194        mInputOutputFormatsMap = inputOutputFormatsMap;
195    }
196
197    /**
198     * Get the image {@code format} output formats in this stream configuration.
199     *
200     * <p>All image formats returned by this function will be defined in either {@link ImageFormat}
201     * or in {@link PixelFormat} (and there is no possibility of collision).</p>
202     *
203     * <p>Formats listed in this array are guaranteed to return true if queried with
204     * {@link #isOutputSupportedFor(int)}.</p>
205     *
206     * @return an array of integer format
207     *
208     * @see ImageFormat
209     * @see PixelFormat
210     */
211    public final int[] getOutputFormats() {
212        return getPublicFormats(/*output*/true);
213    }
214
215    /**
216     * Get the image {@code format} output formats for a reprocessing input format.
217     *
218     * <p>When submitting a {@link CaptureRequest} with an input Surface of a given format,
219     * the only allowed target outputs of the {@link CaptureRequest} are the ones with a format
220     * listed in the return value of this method. Including any other output Surface as a target
221     * will throw an IllegalArgumentException. If no output format is supported given the input
222     * format, an empty int[] will be returned.</p>
223     *
224     * <p>All image formats returned by this function will be defined in either {@link ImageFormat}
225     * or in {@link PixelFormat} (and there is no possibility of collision).</p>
226     *
227     * <p>Formats listed in this array are guaranteed to return true if queried with
228     * {@link #isOutputSupportedFor(int)}.</p>
229     *
230     * @return an array of integer format
231     *
232     * @see ImageFormat
233     * @see PixelFormat
234     */
235    public final int[] getValidOutputFormatsForInput(int inputFormat) {
236        if (mInputOutputFormatsMap == null) {
237            return new int[0];
238        }
239        return mInputOutputFormatsMap.getOutputs(inputFormat);
240    }
241
242    /**
243     * Get the image {@code format} input formats in this stream configuration.
244     *
245     * <p>All image formats returned by this function will be defined in either {@link ImageFormat}
246     * or in {@link PixelFormat} (and there is no possibility of collision).</p>
247     *
248     * @return an array of integer format
249     *
250     * @see ImageFormat
251     * @see PixelFormat
252     */
253    public final int[] getInputFormats() {
254        return getPublicFormats(/*output*/false);
255    }
256
257    /**
258     * Get the supported input sizes for this input format.
259     *
260     * <p>The format must have come from {@link #getInputFormats}; otherwise
261     * {@code null} is returned.</p>
262     *
263     * @param format a format from {@link #getInputFormats}
264     * @return a non-empty array of sizes, or {@code null} if the format was not available.
265     */
266    public Size[] getInputSizes(final int format) {
267        return getPublicFormatSizes(format, /*output*/false, /*highRes*/false);
268    }
269
270    /**
271     * Determine whether or not output surfaces with a particular user-defined format can be passed
272     * {@link CameraDevice#createCaptureSession createCaptureSession}.
273     *
274     * <p>This method determines that the output {@code format} is supported by the camera device;
275     * each output {@code surface} target may or may not itself support that {@code format}.
276     * Refer to the class which provides the surface for additional documentation.</p>
277     *
278     * <p>Formats for which this returns {@code true} are guaranteed to exist in the result
279     * returned by {@link #getOutputSizes}.</p>
280     *
281     * @param format an image format from either {@link ImageFormat} or {@link PixelFormat}
282     * @return
283     *          {@code true} iff using a {@code surface} with this {@code format} will be
284     *          supported with {@link CameraDevice#createCaptureSession}
285     *
286     * @throws IllegalArgumentException
287     *          if the image format was not a defined named constant
288     *          from either {@link ImageFormat} or {@link PixelFormat}
289     *
290     * @see ImageFormat
291     * @see PixelFormat
292     * @see CameraDevice#createCaptureSession
293     */
294    public boolean isOutputSupportedFor(int format) {
295        checkArgumentFormat(format);
296
297        int internalFormat = imageFormatToInternal(format);
298        int dataspace = imageFormatToDataspace(format);
299        if (dataspace == HAL_DATASPACE_DEPTH) {
300            return mDepthOutputFormats.indexOfKey(internalFormat) >= 0;
301        } else {
302            return getFormatsMap(/*output*/true).indexOfKey(internalFormat) >= 0;
303        }
304    }
305
306    /**
307     * Determine whether or not output streams can be configured with a particular class
308     * as a consumer.
309     *
310     * <p>The following list is generally usable for outputs:
311     * <ul>
312     * <li>{@link android.media.ImageReader} -
313     * Recommended for image processing or streaming to external resources (such as a file or
314     * network)
315     * <li>{@link android.media.MediaRecorder} -
316     * Recommended for recording video (simple to use)
317     * <li>{@link android.media.MediaCodec} -
318     * Recommended for recording video (more complicated to use, with more flexibility)
319     * <li>{@link android.renderscript.Allocation} -
320     * Recommended for image processing with {@link android.renderscript RenderScript}
321     * <li>{@link android.view.SurfaceHolder} -
322     * Recommended for low-power camera preview with {@link android.view.SurfaceView}
323     * <li>{@link android.graphics.SurfaceTexture} -
324     * Recommended for OpenGL-accelerated preview processing or compositing with
325     * {@link android.view.TextureView}
326     * </ul>
327     * </p>
328     *
329     * <p>Generally speaking this means that creating a {@link Surface} from that class <i>may</i>
330     * provide a producer endpoint that is suitable to be used with
331     * {@link CameraDevice#createCaptureSession}.</p>
332     *
333     * <p>Since not all of the above classes support output of all format and size combinations,
334     * the particular combination should be queried with {@link #isOutputSupportedFor(Surface)}.</p>
335     *
336     * @param klass a non-{@code null} {@link Class} object reference
337     * @return {@code true} if this class is supported as an output, {@code false} otherwise
338     *
339     * @throws NullPointerException if {@code klass} was {@code null}
340     *
341     * @see CameraDevice#createCaptureSession
342     * @see #isOutputSupportedFor(Surface)
343     */
344    public static <T> boolean isOutputSupportedFor(Class<T> klass) {
345        checkNotNull(klass, "klass must not be null");
346
347        if (klass == android.media.ImageReader.class) {
348            return true;
349        } else if (klass == android.media.MediaRecorder.class) {
350            return true;
351        } else if (klass == android.media.MediaCodec.class) {
352            return true;
353        } else if (klass == android.renderscript.Allocation.class) {
354            return true;
355        } else if (klass == android.view.SurfaceHolder.class) {
356            return true;
357        } else if (klass == android.graphics.SurfaceTexture.class) {
358            return true;
359        }
360
361        return false;
362    }
363
364    /**
365     * Determine whether or not the {@code surface} in its current state is suitable to be included
366     * in a {@link CameraDevice#createCaptureSession capture session} as an output.
367     *
368     * <p>Not all surfaces are usable with the {@link CameraDevice}, and not all configurations
369     * of that {@code surface} are compatible. Some classes that provide the {@code surface} are
370     * compatible with the {@link CameraDevice} in general
371     * (see {@link #isOutputSupportedFor(Class)}, but it is the caller's responsibility to put the
372     * {@code surface} into a state that will be compatible with the {@link CameraDevice}.</p>
373     *
374     * <p>Reasons for a {@code surface} being specifically incompatible might be:
375     * <ul>
376     * <li>Using a format that's not listed by {@link #getOutputFormats}
377     * <li>Using a format/size combination that's not listed by {@link #getOutputSizes}
378     * <li>The {@code surface} itself is not in a state where it can service a new producer.</p>
379     * </li>
380     * </ul>
381     *
382     * <p>Surfaces from flexible sources will return true even if the exact size of the Surface does
383     * not match a camera-supported size, as long as the format (or class) is supported and the
384     * camera device supports a size that is equal to or less than 1080p in that format. If such as
385     * Surface is used to create a capture session, it will have its size rounded to the nearest
386     * supported size, below or equal to 1080p. Flexible sources include SurfaceView, SurfaceTexture,
387     * and ImageReader.</p>
388     *
389     * <p>This is not an exhaustive list; see the particular class's documentation for further
390     * possible reasons of incompatibility.</p>
391     *
392     * @param surface a non-{@code null} {@link Surface} object reference
393     * @return {@code true} if this is supported, {@code false} otherwise
394     *
395     * @throws NullPointerException if {@code surface} was {@code null}
396     * @throws IllegalArgumentException if the Surface endpoint is no longer valid
397     *
398     * @see CameraDevice#createCaptureSession
399     * @see #isOutputSupportedFor(Class)
400     */
401    public boolean isOutputSupportedFor(Surface surface) {
402        checkNotNull(surface, "surface must not be null");
403
404        Size surfaceSize = SurfaceUtils.getSurfaceSize(surface);
405        int surfaceFormat = SurfaceUtils.getSurfaceFormat(surface);
406        int surfaceDataspace = SurfaceUtils.getSurfaceDataspace(surface);
407
408        // See if consumer is flexible.
409        boolean isFlexible = SurfaceUtils.isFlexibleConsumer(surface);
410
411        StreamConfiguration[] configs =
412                surfaceDataspace != HAL_DATASPACE_DEPTH ? mConfigurations : mDepthConfigurations;
413        for (StreamConfiguration config : configs) {
414            if (config.getFormat() == surfaceFormat && config.isOutput()) {
415                // Matching format, either need exact size match, or a flexible consumer
416                // and a size no bigger than MAX_DIMEN_FOR_ROUNDING
417                if (config.getSize().equals(surfaceSize)) {
418                    return true;
419                } else if (isFlexible &&
420                        (config.getSize().getWidth() <= LegacyCameraDevice.MAX_DIMEN_FOR_ROUNDING)) {
421                    return true;
422                }
423            }
424        }
425        return false;
426    }
427
428    /**
429     * Get a list of sizes compatible with {@code klass} to use as an output.
430     *
431     * <p>Some of the supported classes may support additional formats beyond
432     * {@link ImageFormat#PRIVATE}; this function only returns
433     * sizes for {@link ImageFormat#PRIVATE}. For example, {@link android.media.ImageReader}
434     * supports {@link ImageFormat#YUV_420_888} and {@link ImageFormat#PRIVATE}, this method will
435     * only return the sizes for {@link ImageFormat#PRIVATE} for {@link android.media.ImageReader}
436     * class.</p>
437     *
438     * <p>If a well-defined format such as {@code NV21} is required, use
439     * {@link #getOutputSizes(int)} instead.</p>
440     *
441     * <p>The {@code klass} should be a supported output, that querying
442     * {@code #isOutputSupportedFor(Class)} should return {@code true}.</p>
443     *
444     * @param klass
445     *          a non-{@code null} {@link Class} object reference
446     * @return
447     *          an array of supported sizes for {@link ImageFormat#PRIVATE} format,
448     *          or {@code null} iff the {@code klass} is not a supported output.
449     *
450     *
451     * @throws NullPointerException if {@code klass} was {@code null}
452     *
453     * @see #isOutputSupportedFor(Class)
454     */
455    public <T> Size[] getOutputSizes(Class<T> klass) {
456        if (isOutputSupportedFor(klass) == false) {
457            return null;
458        }
459
460        return getInternalFormatSizes(HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED,
461                HAL_DATASPACE_UNKNOWN,/*output*/true, /*highRes*/false);
462    }
463
464    /**
465     * Get a list of sizes compatible with the requested image {@code format}.
466     *
467     * <p>The {@code format} should be a supported format (one of the formats returned by
468     * {@link #getOutputFormats}).</p>
469     *
470     * As of API level 23, the {@link #getHighResolutionOutputSizes} method can be used on devices
471     * that support the
472     * {@link android.hardware.camera2.CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES_BURST_CAPTURE BURST_CAPTURE}
473     * capability to get a list of high-resolution output sizes that cannot operate at the preferred
474     * 20fps rate. This means that for some supported formats, this method will return an empty
475     * list, if all the supported resolutions operate at below 20fps.  For devices that do not
476     * support the BURST_CAPTURE capability, all output resolutions are listed through this method.
477     *
478     * @param format an image format from {@link ImageFormat} or {@link PixelFormat}
479     * @return
480     *          an array of supported sizes,
481     *          or {@code null} if the {@code format} is not a supported output
482     *
483     * @see ImageFormat
484     * @see PixelFormat
485     * @see #getOutputFormats
486     */
487    public Size[] getOutputSizes(int format) {
488        return getPublicFormatSizes(format, /*output*/true, /*highRes*/ false);
489    }
490
491    /**
492     * Get a list of supported high speed video recording sizes.
493     * <p>
494     * When {@link CameraMetadata#REQUEST_AVAILABLE_CAPABILITIES_CONSTRAINED_HIGH_SPEED_VIDEO} is
495     * supported in {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES}, this method will
496     * list the supported high speed video size configurations. All the sizes listed will be a
497     * subset of the sizes reported by {@link #getOutputSizes} for processed non-stalling formats
498     * (typically {@link ImageFormat#PRIVATE} {@link ImageFormat#YUV_420_888}, etc.)
499     * </p>
500     * <p>
501     * To enable high speed video recording, application must create a constrained create high speed
502     * capture session via {@link CameraDevice#createConstrainedHighSpeedCaptureSession}, and submit
503     * a CaptureRequest list created by
504     * {@link android.hardware.camera2.CameraConstrainedHighSpeedCaptureSession#createHighSpeedRequestList}
505     * to this session. The application must select the video size from this method and
506     * {@link CaptureRequest#CONTROL_AE_TARGET_FPS_RANGE FPS range} from
507     * {@link #getHighSpeedVideoFpsRangesFor} to configure the constrained high speed session and
508     * generate the high speed request list. For example, if the application intends to do high
509     * speed recording, it can select the maximum size reported by this method to create high speed
510     * capture session. Note that for the use case of multiple output streams, application must
511     * select one unique size from this method to use (e.g., preview and recording streams must have
512     * the same size). Otherwise, the high speed session creation will fail. Once the size is
513     * selected, application can get the supported FPS ranges by
514     * {@link #getHighSpeedVideoFpsRangesFor}, and use these FPS ranges to setup the recording
515     * request lists via
516     * {@link android.hardware.camera2.CameraConstrainedHighSpeedCaptureSession#createHighSpeedRequestList}.
517     * </p>
518     *
519     * @return an array of supported high speed video recording sizes
520     * @see #getHighSpeedVideoFpsRangesFor(Size)
521     * @see CameraMetadata#REQUEST_AVAILABLE_CAPABILITIES_CONSTRAINED_HIGH_SPEED_VIDEO
522     * @see CameraDevice#createConstrainedHighSpeedCaptureSession
523     * @see android.hardware.camera2.CameraConstrainedHighSpeedCaptureSession#createHighSpeedRequestList
524     */
525    public Size[] getHighSpeedVideoSizes() {
526        Set<Size> keySet = mHighSpeedVideoSizeMap.keySet();
527        return keySet.toArray(new Size[keySet.size()]);
528    }
529
530    /**
531     * Get the frame per second ranges (fpsMin, fpsMax) for input high speed video size.
532     * <p>
533     * See {@link #getHighSpeedVideoFpsRanges} for how to enable high speed recording.
534     * </p>
535     * <p>
536     * The {@link CaptureRequest#CONTROL_AE_TARGET_FPS_RANGE FPS ranges} reported in this method
537     * must not be used to setup capture requests that are submitted to unconstrained capture
538     * sessions, or it will result in {@link IllegalArgumentException IllegalArgumentExceptions}.
539     * </p>
540     * <p>
541     * See {@link #getHighSpeedVideoFpsRanges} for the characteristics of the returned FPS ranges.
542     * </p>
543     *
544     * @param size one of the sizes returned by {@link #getHighSpeedVideoSizes()}
545     * @return an array of supported high speed video recording FPS ranges The upper bound of
546     *         returned ranges is guaranteed to be greater than or equal to 120.
547     * @throws IllegalArgumentException if input size does not exist in the return value of
548     *             getHighSpeedVideoSizes
549     * @see #getHighSpeedVideoSizes()
550     * @see #getHighSpeedVideoFpsRanges()
551     */
552    public Range<Integer>[] getHighSpeedVideoFpsRangesFor(Size size) {
553        Integer fpsRangeCount = mHighSpeedVideoSizeMap.get(size);
554        if (fpsRangeCount == null || fpsRangeCount == 0) {
555            throw new IllegalArgumentException(String.format(
556                    "Size %s does not support high speed video recording", size));
557        }
558
559        @SuppressWarnings("unchecked")
560        Range<Integer>[] fpsRanges = new Range[fpsRangeCount];
561        int i = 0;
562        for (HighSpeedVideoConfiguration config : mHighSpeedVideoConfigurations) {
563            if (size.equals(config.getSize())) {
564                fpsRanges[i++] = config.getFpsRange();
565            }
566        }
567        return fpsRanges;
568    }
569
570    /**
571     * Get a list of supported high speed video recording FPS ranges.
572     * <p>
573     * When {@link CameraMetadata#REQUEST_AVAILABLE_CAPABILITIES_CONSTRAINED_HIGH_SPEED_VIDEO} is
574     * supported in {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES}, this method will
575     * list the supported high speed video FPS range configurations. Application can then use
576     * {@link #getHighSpeedVideoSizesFor} to query available sizes for one of returned FPS range.
577     * </p>
578     * <p>
579     * To enable high speed video recording, application must create a constrained create high speed
580     * capture session via {@link CameraDevice#createConstrainedHighSpeedCaptureSession}, and submit
581     * a CaptureRequest list created by
582     * {@link android.hardware.camera2.CameraConstrainedHighSpeedCaptureSession#createHighSpeedRequestList}
583     * to this session. The application must select the video size from this method and
584     * {@link CaptureRequest#CONTROL_AE_TARGET_FPS_RANGE FPS range} from
585     * {@link #getHighSpeedVideoFpsRangesFor} to configure the constrained high speed session and
586     * generate the high speed request list. For example, if the application intends to do high
587     * speed recording, it can select one FPS range reported by this method, query the video sizes
588     * corresponding to this FPS range by {@link #getHighSpeedVideoSizesFor} and use one of reported
589     * sizes to create a high speed capture session. Note that for the use case of multiple output
590     * streams, application must select one unique size from this method to use (e.g., preview and
591     * recording streams must have the same size). Otherwise, the high speed session creation will
592     * fail. Once the high speed capture session is created, the application can set the FPS range
593     * in the recording request lists via
594     * {@link android.hardware.camera2.CameraConstrainedHighSpeedCaptureSession#createHighSpeedRequestList}.
595     * </p>
596     * <p>
597     * The FPS ranges reported by this method will have below characteristics:
598     * <li>The fpsMin and fpsMax will be a multiple 30fps.</li>
599     * <li>The fpsMin will be no less than 30fps, the fpsMax will be no less than 120fps.</li>
600     * <li>At least one range will be a fixed FPS range where fpsMin == fpsMax.</li>
601     * <li>For each fixed FPS range, there will be one corresponding variable FPS range [30,
602     * fps_max]. These kinds of FPS ranges are suitable for preview-only use cases where the
603     * application doesn't want the camera device always produce higher frame rate than the display
604     * refresh rate.</li>
605     * </p>
606     *
607     * @return an array of supported high speed video recording FPS ranges The upper bound of
608     *         returned ranges is guaranteed to be larger or equal to 120.
609     * @see #getHighSpeedVideoSizesFor
610     * @see CameraMetadata#REQUEST_AVAILABLE_CAPABILITIES_CONSTRAINED_HIGH_SPEED_VIDEO
611     * @see CameraDevice#createConstrainedHighSpeedCaptureSession
612     * @see CameraDevice#createHighSpeedRequestList
613     */
614    @SuppressWarnings("unchecked")
615    public Range<Integer>[] getHighSpeedVideoFpsRanges() {
616        Set<Range<Integer>> keySet = mHighSpeedVideoFpsRangeMap.keySet();
617        return keySet.toArray(new Range[keySet.size()]);
618    }
619
620    /**
621     * Get the supported video sizes for an input high speed FPS range.
622     *
623     * <p> See {@link #getHighSpeedVideoSizes} for how to enable high speed recording.</p>
624     *
625     * @param fpsRange one of the FPS range returned by {@link #getHighSpeedVideoFpsRanges()}
626     * @return An array of video sizes to create high speed capture sessions for high speed streaming
627     *         use cases.
628     *
629     * @throws IllegalArgumentException if input FPS range does not exist in the return value of
630     *         getHighSpeedVideoFpsRanges
631     * @see #getHighSpeedVideoFpsRanges()
632     */
633    public Size[] getHighSpeedVideoSizesFor(Range<Integer> fpsRange) {
634        Integer sizeCount = mHighSpeedVideoFpsRangeMap.get(fpsRange);
635        if (sizeCount == null || sizeCount == 0) {
636            throw new IllegalArgumentException(String.format(
637                    "FpsRange %s does not support high speed video recording", fpsRange));
638        }
639
640        Size[] sizes = new Size[sizeCount];
641        int i = 0;
642        for (HighSpeedVideoConfiguration config : mHighSpeedVideoConfigurations) {
643            if (fpsRange.equals(config.getFpsRange())) {
644                sizes[i++] = config.getSize();
645            }
646        }
647        return sizes;
648    }
649
650    /**
651     * Get a list of supported high resolution sizes, which cannot operate at full BURST_CAPTURE
652     * rate.
653     *
654     * <p>This includes all output sizes that cannot meet the 20 fps frame rate requirements for the
655     * {@link android.hardware.camera2.CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES_BURST_CAPTURE BURST_CAPTURE}
656     * capability.  This does not include the stall duration, so for example, a JPEG or RAW16 output
657     * resolution with a large stall duration but a minimum frame duration that's above 20 fps will
658     * still be listed in the regular {@link #getOutputSizes} list. All the sizes on this list are
659     * still guaranteed to operate at a rate of at least 10 fps, not including stall duration.</p>
660     *
661     * <p>For a device that does not support the BURST_CAPTURE capability, this list will be
662     * {@code null}, since resolutions in the {@link #getOutputSizes} list are already not
663     * guaranteed to meet &gt;= 20 fps rate requirements. For a device that does support the
664     * BURST_CAPTURE capability, this list may be empty, if all supported resolutions meet the 20
665     * fps requirement.</p>
666     *
667     * @return an array of supported slower high-resolution sizes, or {@code null} if the
668     *         BURST_CAPTURE capability is not supported
669     */
670    public Size[] getHighResolutionOutputSizes(int format) {
671        if (!mListHighResolution) return null;
672
673        return getPublicFormatSizes(format, /*output*/true, /*highRes*/ true);
674    }
675
676    /**
677     * Get the minimum {@link CaptureRequest#SENSOR_FRAME_DURATION frame duration}
678     * for the format/size combination (in nanoseconds).
679     *
680     * <p>{@code format} should be one of the ones returned by {@link #getOutputFormats()}.</p>
681     * <p>{@code size} should be one of the ones returned by
682     * {@link #getOutputSizes(int)}.</p>
683     *
684     * <p>This should correspond to the frame duration when only that stream is active, with all
685     * processing (typically in {@code android.*.mode}) set to either {@code OFF} or {@code FAST}.
686     * </p>
687     *
688     * <p>When multiple streams are used in a request, the minimum frame duration will be
689     * {@code max(individual stream min durations)}.</p>
690     *
691     * <p>For devices that do not support manual sensor control
692     * ({@link android.hardware.camera2.CameraMetadata#REQUEST_AVAILABLE_CAPABILITIES_MANUAL_SENSOR}),
693     * this function may return 0.</p>
694     *
695     * <!--
696     * TODO: uncomment after adding input stream support
697     * <p>The minimum frame duration of a stream (of a particular format, size) is the same
698     * regardless of whether the stream is input or output.</p>
699     * -->
700     *
701     * @param format an image format from {@link ImageFormat} or {@link PixelFormat}
702     * @param size an output-compatible size
703     * @return a minimum frame duration {@code >} 0 in nanoseconds, or
704     *          0 if the minimum frame duration is not available.
705     *
706     * @throws IllegalArgumentException if {@code format} or {@code size} was not supported
707     * @throws NullPointerException if {@code size} was {@code null}
708     *
709     * @see CaptureRequest#SENSOR_FRAME_DURATION
710     * @see #getOutputStallDuration(int, Size)
711     * @see ImageFormat
712     * @see PixelFormat
713     */
714    public long getOutputMinFrameDuration(int format, Size size) {
715        checkNotNull(size, "size must not be null");
716        checkArgumentFormatSupported(format, /*output*/true);
717
718        return getInternalFormatDuration(imageFormatToInternal(format),
719                imageFormatToDataspace(format),
720                size,
721                DURATION_MIN_FRAME);
722    }
723
724    /**
725     * Get the minimum {@link CaptureRequest#SENSOR_FRAME_DURATION frame duration}
726     * for the class/size combination (in nanoseconds).
727     *
728     * <p>This assumes a the {@code klass} is set up to use {@link ImageFormat#PRIVATE}.
729     * For user-defined formats, use {@link #getOutputMinFrameDuration(int, Size)}.</p>
730     *
731     * <p>{@code klass} should be one of the ones which is supported by
732     * {@link #isOutputSupportedFor(Class)}.</p>
733     *
734     * <p>{@code size} should be one of the ones returned by
735     * {@link #getOutputSizes(int)}.</p>
736     *
737     * <p>This should correspond to the frame duration when only that stream is active, with all
738     * processing (typically in {@code android.*.mode}) set to either {@code OFF} or {@code FAST}.
739     * </p>
740     *
741     * <p>When multiple streams are used in a request, the minimum frame duration will be
742     * {@code max(individual stream min durations)}.</p>
743     *
744     * <p>For devices that do not support manual sensor control
745     * ({@link android.hardware.camera2.CameraMetadata#REQUEST_AVAILABLE_CAPABILITIES_MANUAL_SENSOR}),
746     * this function may return 0.</p>
747     *
748     * <!--
749     * TODO: uncomment after adding input stream support
750     * <p>The minimum frame duration of a stream (of a particular format, size) is the same
751     * regardless of whether the stream is input or output.</p>
752     * -->
753     *
754     * @param klass
755     *          a class which is supported by {@link #isOutputSupportedFor(Class)} and has a
756     *          non-empty array returned by {@link #getOutputSizes(Class)}
757     * @param size an output-compatible size
758     * @return a minimum frame duration {@code >} 0 in nanoseconds, or
759     *          0 if the minimum frame duration is not available.
760     *
761     * @throws IllegalArgumentException if {@code klass} or {@code size} was not supported
762     * @throws NullPointerException if {@code size} or {@code klass} was {@code null}
763     *
764     * @see CaptureRequest#SENSOR_FRAME_DURATION
765     * @see ImageFormat
766     * @see PixelFormat
767     */
768    public <T> long getOutputMinFrameDuration(final Class<T> klass, final Size size) {
769        if (!isOutputSupportedFor(klass)) {
770            throw new IllegalArgumentException("klass was not supported");
771        }
772
773        return getInternalFormatDuration(HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED,
774                HAL_DATASPACE_UNKNOWN,
775                size, DURATION_MIN_FRAME);
776    }
777
778    /**
779     * Get the stall duration for the format/size combination (in nanoseconds).
780     *
781     * <p>{@code format} should be one of the ones returned by {@link #getOutputFormats()}.</p>
782     * <p>{@code size} should be one of the ones returned by
783     * {@link #getOutputSizes(int)}.</p>
784     *
785     * <p>
786     * A stall duration is how much extra time would get added to the normal minimum frame duration
787     * for a repeating request that has streams with non-zero stall.
788     *
789     * <p>For example, consider JPEG captures which have the following characteristics:
790     *
791     * <ul>
792     * <li>JPEG streams act like processed YUV streams in requests for which they are not included;
793     * in requests in which they are directly referenced, they act as JPEG streams.
794     * This is because supporting a JPEG stream requires the underlying YUV data to always be ready
795     * for use by a JPEG encoder, but the encoder will only be used (and impact frame duration) on
796     * requests that actually reference a JPEG stream.
797     * <li>The JPEG processor can run concurrently to the rest of the camera pipeline, but cannot
798     * process more than 1 capture at a time.
799     * </ul>
800     *
801     * <p>In other words, using a repeating YUV request would result in a steady frame rate
802     * (let's say it's 30 FPS). If a single JPEG request is submitted periodically,
803     * the frame rate will stay at 30 FPS (as long as we wait for the previous JPEG to return each
804     * time). If we try to submit a repeating YUV + JPEG request, then the frame rate will drop from
805     * 30 FPS.</p>
806     *
807     * <p>In general, submitting a new request with a non-0 stall time stream will <em>not</em> cause a
808     * frame rate drop unless there are still outstanding buffers for that stream from previous
809     * requests.</p>
810     *
811     * <p>Submitting a repeating request with streams (call this {@code S}) is the same as setting
812     * the minimum frame duration from the normal minimum frame duration corresponding to {@code S},
813     * added with the maximum stall duration for {@code S}.</p>
814     *
815     * <p>If interleaving requests with and without a stall duration, a request will stall by the
816     * maximum of the remaining times for each can-stall stream with outstanding buffers.</p>
817     *
818     * <p>This means that a stalling request will not have an exposure start until the stall has
819     * completed.</p>
820     *
821     * <p>This should correspond to the stall duration when only that stream is active, with all
822     * processing (typically in {@code android.*.mode}) set to {@code FAST} or {@code OFF}.
823     * Setting any of the processing modes to {@code HIGH_QUALITY} effectively results in an
824     * indeterminate stall duration for all streams in a request (the regular stall calculation
825     * rules are ignored).</p>
826     *
827     * <p>The following formats may always have a stall duration:
828     * <ul>
829     * <li>{@link ImageFormat#JPEG JPEG}
830     * <li>{@link ImageFormat#RAW_SENSOR RAW16}
831     * <li>{@link ImageFormat#RAW_PRIVATE RAW_PRIVATE}
832     * </ul>
833     * </p>
834     *
835     * <p>The following formats will never have a stall duration:
836     * <ul>
837     * <li>{@link ImageFormat#YUV_420_888 YUV_420_888}
838     * <li>{@link #isOutputSupportedFor(Class) Implementation-Defined}
839     * </ul></p>
840     *
841     * <p>
842     * All other formats may or may not have an allowed stall duration on a per-capability basis;
843     * refer to {@link CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES
844     * android.request.availableCapabilities} for more details.</p>
845     * </p>
846     *
847     * <p>See {@link CaptureRequest#SENSOR_FRAME_DURATION android.sensor.frameDuration}
848     * for more information about calculating the max frame rate (absent stalls).</p>
849     *
850     * @param format an image format from {@link ImageFormat} or {@link PixelFormat}
851     * @param size an output-compatible size
852     * @return a stall duration {@code >=} 0 in nanoseconds
853     *
854     * @throws IllegalArgumentException if {@code format} or {@code size} was not supported
855     * @throws NullPointerException if {@code size} was {@code null}
856     *
857     * @see CaptureRequest#SENSOR_FRAME_DURATION
858     * @see ImageFormat
859     * @see PixelFormat
860     */
861    public long getOutputStallDuration(int format, Size size) {
862        checkArgumentFormatSupported(format, /*output*/true);
863
864        return getInternalFormatDuration(imageFormatToInternal(format),
865                imageFormatToDataspace(format),
866                size,
867                DURATION_STALL);
868    }
869
870    /**
871     * Get the stall duration for the class/size combination (in nanoseconds).
872     *
873     * <p>This assumes a the {@code klass} is set up to use {@link ImageFormat#PRIVATE}.
874     * For user-defined formats, use {@link #getOutputMinFrameDuration(int, Size)}.</p>
875     *
876     * <p>{@code klass} should be one of the ones with a non-empty array returned by
877     * {@link #getOutputSizes(Class)}.</p>
878     *
879     * <p>{@code size} should be one of the ones returned by
880     * {@link #getOutputSizes(Class)}.</p>
881     *
882     * <p>See {@link #getOutputStallDuration(int, Size)} for a definition of a
883     * <em>stall duration</em>.</p>
884     *
885     * @param klass
886     *          a class which is supported by {@link #isOutputSupportedFor(Class)} and has a
887     *          non-empty array returned by {@link #getOutputSizes(Class)}
888     * @param size an output-compatible size
889     * @return a minimum frame duration {@code >=} 0 in nanoseconds
890     *
891     * @throws IllegalArgumentException if {@code klass} or {@code size} was not supported
892     * @throws NullPointerException if {@code size} or {@code klass} was {@code null}
893     *
894     * @see CaptureRequest#SENSOR_FRAME_DURATION
895     * @see ImageFormat
896     * @see PixelFormat
897     */
898    public <T> long getOutputStallDuration(final Class<T> klass, final Size size) {
899        if (!isOutputSupportedFor(klass)) {
900            throw new IllegalArgumentException("klass was not supported");
901        }
902
903        return getInternalFormatDuration(HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED,
904                HAL_DATASPACE_UNKNOWN, size, DURATION_STALL);
905    }
906
907    /**
908     * Check if this {@link StreamConfigurationMap} is equal to another
909     * {@link StreamConfigurationMap}.
910     *
911     * <p>Two vectors are only equal if and only if each of the respective elements is equal.</p>
912     *
913     * @return {@code true} if the objects were equal, {@code false} otherwise
914     */
915    @Override
916    public boolean equals(final Object obj) {
917        if (obj == null) {
918            return false;
919        }
920        if (this == obj) {
921            return true;
922        }
923        if (obj instanceof StreamConfigurationMap) {
924            final StreamConfigurationMap other = (StreamConfigurationMap) obj;
925            // XX: do we care about order?
926            return Arrays.equals(mConfigurations, other.mConfigurations) &&
927                    Arrays.equals(mMinFrameDurations, other.mMinFrameDurations) &&
928                    Arrays.equals(mStallDurations, other.mStallDurations) &&
929                    Arrays.equals(mDepthConfigurations, other.mDepthConfigurations) &&
930                    Arrays.equals(mHighSpeedVideoConfigurations,
931                            other.mHighSpeedVideoConfigurations);
932        }
933        return false;
934    }
935
936    /**
937     * {@inheritDoc}
938     */
939    @Override
940    public int hashCode() {
941        // XX: do we care about order?
942        return HashCodeHelpers.hashCodeGeneric(
943                mConfigurations, mMinFrameDurations,
944                mStallDurations,
945                mDepthConfigurations, mHighSpeedVideoConfigurations);
946    }
947
948    // Check that the argument is supported by #getOutputFormats or #getInputFormats
949    private int checkArgumentFormatSupported(int format, boolean output) {
950        checkArgumentFormat(format);
951
952        int internalFormat = imageFormatToInternal(format);
953        int internalDataspace = imageFormatToDataspace(format);
954
955        if (output) {
956            if (internalDataspace == HAL_DATASPACE_DEPTH) {
957                if (mDepthOutputFormats.indexOfKey(internalFormat) >= 0) {
958                    return format;
959                }
960            } else {
961                if (mAllOutputFormats.indexOfKey(internalFormat) >= 0) {
962                    return format;
963                }
964            }
965        } else {
966            if (mInputFormats.indexOfKey(internalFormat) >= 0) {
967                return format;
968            }
969        }
970
971        throw new IllegalArgumentException(String.format(
972                "format %x is not supported by this stream configuration map", format));
973    }
974
975    /**
976     * Ensures that the format is either user-defined or implementation defined.
977     *
978     * <p>If a format has a different internal representation than the public representation,
979     * passing in the public representation here will fail.</p>
980     *
981     * <p>For example if trying to use {@link ImageFormat#JPEG}:
982     * it has a different public representation than the internal representation
983     * {@code HAL_PIXEL_FORMAT_BLOB}, this check will fail.</p>
984     *
985     * <p>Any invalid/undefined formats will raise an exception.</p>
986     *
987     * @param format image format
988     * @return the format
989     *
990     * @throws IllegalArgumentException if the format was invalid
991     */
992    static int checkArgumentFormatInternal(int format) {
993        switch (format) {
994            case HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED:
995            case HAL_PIXEL_FORMAT_BLOB:
996            case HAL_PIXEL_FORMAT_RAW_OPAQUE:
997            case HAL_PIXEL_FORMAT_Y16:
998                return format;
999            case ImageFormat.JPEG:
1000                throw new IllegalArgumentException(
1001                        "ImageFormat.JPEG is an unknown internal format");
1002            default:
1003                return checkArgumentFormat(format);
1004        }
1005    }
1006
1007    /**
1008     * Ensures that the format is publicly user-defined in either ImageFormat or PixelFormat.
1009     *
1010     * <p>If a format has a different public representation than the internal representation,
1011     * passing in the internal representation here will fail.</p>
1012     *
1013     * <p>For example if trying to use {@code HAL_PIXEL_FORMAT_BLOB}:
1014     * it has a different internal representation than the public representation
1015     * {@link ImageFormat#JPEG}, this check will fail.</p>
1016     *
1017     * <p>Any invalid/undefined formats will raise an exception, including implementation-defined.
1018     * </p>
1019     *
1020     * <p>Note that {@code @hide} and deprecated formats will not pass this check.</p>
1021     *
1022     * @param format image format
1023     * @return the format
1024     *
1025     * @throws IllegalArgumentException if the format was not user-defined
1026     */
1027    static int checkArgumentFormat(int format) {
1028        if (!ImageFormat.isPublicFormat(format) && !PixelFormat.isPublicFormat(format)) {
1029            throw new IllegalArgumentException(String.format(
1030                    "format 0x%x was not defined in either ImageFormat or PixelFormat", format));
1031        }
1032
1033        return format;
1034    }
1035
1036    /**
1037     * Convert an internal format compatible with {@code graphics.h} into public-visible
1038     * {@code ImageFormat}. This assumes the dataspace of the format is not HAL_DATASPACE_DEPTH.
1039     *
1040     * <p>In particular these formats are converted:
1041     * <ul>
1042     * <li>HAL_PIXEL_FORMAT_BLOB => ImageFormat.JPEG</li>
1043     * </ul>
1044     * </p>
1045     *
1046     * <p>Passing in a format which has no public equivalent will fail;
1047     * as will passing in a public format which has a different internal format equivalent.
1048     * See {@link #checkArgumentFormat} for more details about a legal public format.</p>
1049     *
1050     * <p>All other formats are returned as-is, no further invalid check is performed.</p>
1051     *
1052     * <p>This function is the dual of {@link #imageFormatToInternal} for dataspaces other than
1053     * HAL_DATASPACE_DEPTH.</p>
1054     *
1055     * @param format image format from {@link ImageFormat} or {@link PixelFormat}
1056     * @return the converted image formats
1057     *
1058     * @throws IllegalArgumentException
1059     *          if {@code format} is {@code HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED} or
1060     *          {@link ImageFormat#JPEG}
1061     *
1062     * @see ImageFormat
1063     * @see PixelFormat
1064     * @see #checkArgumentFormat
1065     */
1066    static int imageFormatToPublic(int format) {
1067        switch (format) {
1068            case HAL_PIXEL_FORMAT_BLOB:
1069                return ImageFormat.JPEG;
1070            case ImageFormat.JPEG:
1071                throw new IllegalArgumentException(
1072                        "ImageFormat.JPEG is an unknown internal format");
1073            default:
1074                return format;
1075        }
1076    }
1077
1078    /**
1079     * Convert an internal format compatible with {@code graphics.h} into public-visible
1080     * {@code ImageFormat}. This assumes the dataspace of the format is HAL_DATASPACE_DEPTH.
1081     *
1082     * <p>In particular these formats are converted:
1083     * <ul>
1084     * <li>HAL_PIXEL_FORMAT_BLOB => ImageFormat.DEPTH_POINT_CLOUD
1085     * <li>HAL_PIXEL_FORMAT_Y16 => ImageFormat.DEPTH16
1086     * </ul>
1087     * </p>
1088     *
1089     * <p>Passing in an implementation-defined format which has no public equivalent will fail;
1090     * as will passing in a public format which has a different internal format equivalent.
1091     * See {@link #checkArgumentFormat} for more details about a legal public format.</p>
1092     *
1093     * <p>All other formats are returned as-is, no further invalid check is performed.</p>
1094     *
1095     * <p>This function is the dual of {@link #imageFormatToInternal} for formats associated with
1096     * HAL_DATASPACE_DEPTH.</p>
1097     *
1098     * @param format image format from {@link ImageFormat} or {@link PixelFormat}
1099     * @return the converted image formats
1100     *
1101     * @throws IllegalArgumentException
1102     *          if {@code format} is {@code HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED} or
1103     *          {@link ImageFormat#JPEG}
1104     *
1105     * @see ImageFormat
1106     * @see PixelFormat
1107     * @see #checkArgumentFormat
1108     */
1109    static int depthFormatToPublic(int format) {
1110        switch (format) {
1111            case HAL_PIXEL_FORMAT_BLOB:
1112                return ImageFormat.DEPTH_POINT_CLOUD;
1113            case HAL_PIXEL_FORMAT_Y16:
1114                return ImageFormat.DEPTH16;
1115            case ImageFormat.JPEG:
1116                throw new IllegalArgumentException(
1117                        "ImageFormat.JPEG is an unknown internal format");
1118            case HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED:
1119                throw new IllegalArgumentException(
1120                        "IMPLEMENTATION_DEFINED must not leak to public API");
1121            default:
1122                throw new IllegalArgumentException(
1123                        "Unknown DATASPACE_DEPTH format " + format);
1124        }
1125    }
1126
1127    /**
1128     * Convert image formats from internal to public formats (in-place).
1129     *
1130     * @param formats an array of image formats
1131     * @return {@code formats}
1132     *
1133     * @see #imageFormatToPublic
1134     */
1135    static int[] imageFormatToPublic(int[] formats) {
1136        if (formats == null) {
1137            return null;
1138        }
1139
1140        for (int i = 0; i < formats.length; ++i) {
1141            formats[i] = imageFormatToPublic(formats[i]);
1142        }
1143
1144        return formats;
1145    }
1146
1147    /**
1148     * Convert a public format compatible with {@code ImageFormat} to an internal format
1149     * from {@code graphics.h}.
1150     *
1151     * <p>In particular these formats are converted:
1152     * <ul>
1153     * <li>ImageFormat.JPEG => HAL_PIXEL_FORMAT_BLOB
1154     * <li>ImageFormat.DEPTH_POINT_CLOUD => HAL_PIXEL_FORMAT_BLOB
1155     * <li>ImageFormat.DEPTH16 => HAL_PIXEL_FORMAT_Y16
1156     * </ul>
1157     * </p>
1158     *
1159     * <p>Passing in an internal format which has a different public format equivalent will fail.
1160     * See {@link #checkArgumentFormat} for more details about a legal public format.</p>
1161     *
1162     * <p>All other formats are returned as-is, no invalid check is performed.</p>
1163     *
1164     * <p>This function is the dual of {@link #imageFormatToPublic}.</p>
1165     *
1166     * @param format public image format from {@link ImageFormat} or {@link PixelFormat}
1167     * @return the converted image formats
1168     *
1169     * @see ImageFormat
1170     * @see PixelFormat
1171     *
1172     * @throws IllegalArgumentException
1173     *              if {@code format} was {@code HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED}
1174     */
1175    static int imageFormatToInternal(int format) {
1176        switch (format) {
1177            case ImageFormat.JPEG:
1178            case ImageFormat.DEPTH_POINT_CLOUD:
1179                return HAL_PIXEL_FORMAT_BLOB;
1180            case ImageFormat.DEPTH16:
1181                return HAL_PIXEL_FORMAT_Y16;
1182            default:
1183                return format;
1184        }
1185    }
1186
1187    /**
1188     * Convert a public format compatible with {@code ImageFormat} to an internal dataspace
1189     * from {@code graphics.h}.
1190     *
1191     * <p>In particular these formats are converted:
1192     * <ul>
1193     * <li>ImageFormat.JPEG => HAL_DATASPACE_V0_JFIF
1194     * <li>ImageFormat.DEPTH_POINT_CLOUD => HAL_DATASPACE_DEPTH
1195     * <li>ImageFormat.DEPTH16 => HAL_DATASPACE_DEPTH
1196     * <li>others => HAL_DATASPACE_UNKNOWN
1197     * </ul>
1198     * </p>
1199     *
1200     * <p>Passing in an implementation-defined format here will fail (it's not a public format);
1201     * as will passing in an internal format which has a different public format equivalent.
1202     * See {@link #checkArgumentFormat} for more details about a legal public format.</p>
1203     *
1204     * <p>All other formats are returned as-is, no invalid check is performed.</p>
1205     *
1206     * <p>This function is the dual of {@link #imageFormatToPublic}.</p>
1207     *
1208     * @param format public image format from {@link ImageFormat} or {@link PixelFormat}
1209     * @return the converted image formats
1210     *
1211     * @see ImageFormat
1212     * @see PixelFormat
1213     *
1214     * @throws IllegalArgumentException
1215     *              if {@code format} was {@code HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED}
1216     */
1217    static int imageFormatToDataspace(int format) {
1218        switch (format) {
1219            case ImageFormat.JPEG:
1220                return HAL_DATASPACE_V0_JFIF;
1221            case ImageFormat.DEPTH_POINT_CLOUD:
1222            case ImageFormat.DEPTH16:
1223                return HAL_DATASPACE_DEPTH;
1224            default:
1225                return HAL_DATASPACE_UNKNOWN;
1226        }
1227    }
1228
1229    /**
1230     * Convert image formats from public to internal formats (in-place).
1231     *
1232     * @param formats an array of image formats
1233     * @return {@code formats}
1234     *
1235     * @see #imageFormatToInternal
1236     *
1237     * @hide
1238     */
1239    public static int[] imageFormatToInternal(int[] formats) {
1240        if (formats == null) {
1241            return null;
1242        }
1243
1244        for (int i = 0; i < formats.length; ++i) {
1245            formats[i] = imageFormatToInternal(formats[i]);
1246        }
1247
1248        return formats;
1249    }
1250
1251    private Size[] getPublicFormatSizes(int format, boolean output, boolean highRes) {
1252        try {
1253            checkArgumentFormatSupported(format, output);
1254        } catch (IllegalArgumentException e) {
1255            return null;
1256        }
1257
1258        int internalFormat = imageFormatToInternal(format);
1259        int dataspace = imageFormatToDataspace(format);
1260
1261        return getInternalFormatSizes(internalFormat, dataspace, output, highRes);
1262    }
1263
1264    private Size[] getInternalFormatSizes(int format, int dataspace,
1265            boolean output, boolean highRes) {
1266        // All depth formats are non-high-res.
1267        if (dataspace == HAL_DATASPACE_DEPTH && highRes) {
1268            return new Size[0];
1269        }
1270
1271        SparseIntArray formatsMap =
1272                !output ? mInputFormats :
1273                dataspace == HAL_DATASPACE_DEPTH ? mDepthOutputFormats :
1274                highRes ? mHighResOutputFormats :
1275                mOutputFormats;
1276
1277        int sizesCount = formatsMap.get(format);
1278        if ( ((!output || dataspace == HAL_DATASPACE_DEPTH) && sizesCount == 0) ||
1279                (output && dataspace != HAL_DATASPACE_DEPTH && mAllOutputFormats.get(format) == 0)) {
1280            // Only throw if this is really not supported at all
1281            throw new IllegalArgumentException("format not available");
1282        }
1283
1284        Size[] sizes = new Size[sizesCount];
1285        int sizeIndex = 0;
1286
1287        StreamConfiguration[] configurations =
1288                (dataspace == HAL_DATASPACE_DEPTH) ? mDepthConfigurations : mConfigurations;
1289        StreamConfigurationDuration[] minFrameDurations =
1290                (dataspace == HAL_DATASPACE_DEPTH) ? mDepthMinFrameDurations : mMinFrameDurations;
1291
1292        for (StreamConfiguration config : configurations) {
1293            int fmt = config.getFormat();
1294            if (fmt == format && config.isOutput() == output) {
1295                if (output && mListHighResolution) {
1296                    // Filter slow high-res output formats; include for
1297                    // highRes, remove for !highRes
1298                    long duration = 0;
1299                    for (int i = 0; i < minFrameDurations.length; i++) {
1300                        StreamConfigurationDuration d = minFrameDurations[i];
1301                        if (d.getFormat() == fmt &&
1302                                d.getWidth() == config.getSize().getWidth() &&
1303                                d.getHeight() == config.getSize().getHeight()) {
1304                            duration = d.getDuration();
1305                            break;
1306                        }
1307                    }
1308                    if (dataspace != HAL_DATASPACE_DEPTH &&
1309                            highRes != (duration > DURATION_20FPS_NS)) {
1310                        continue;
1311                    }
1312                }
1313                sizes[sizeIndex++] = config.getSize();
1314            }
1315        }
1316
1317        if (sizeIndex != sizesCount) {
1318            throw new AssertionError(
1319                    "Too few sizes (expected " + sizesCount + ", actual " + sizeIndex + ")");
1320        }
1321
1322        return sizes;
1323    }
1324
1325    /** Get the list of publically visible output formats; does not include IMPL_DEFINED */
1326    private int[] getPublicFormats(boolean output) {
1327        int[] formats = new int[getPublicFormatCount(output)];
1328
1329        int i = 0;
1330
1331        SparseIntArray map = getFormatsMap(output);
1332        for (int j = 0; j < map.size(); j++) {
1333            int format = map.keyAt(j);
1334            formats[i++] = imageFormatToPublic(format);
1335        }
1336        if (output) {
1337            for (int j = 0; j < mDepthOutputFormats.size(); j++) {
1338                formats[i++] = depthFormatToPublic(mDepthOutputFormats.keyAt(j));
1339            }
1340        }
1341        if (formats.length != i) {
1342            throw new AssertionError("Too few formats " + i + ", expected " + formats.length);
1343        }
1344
1345        return formats;
1346    }
1347
1348    /** Get the format -> size count map for either output or input formats */
1349    private SparseIntArray getFormatsMap(boolean output) {
1350        return output ? mAllOutputFormats : mInputFormats;
1351    }
1352
1353    private long getInternalFormatDuration(int format, int dataspace, Size size, int duration) {
1354        // assume format is already checked, since its internal
1355
1356        if (!isSupportedInternalConfiguration(format, dataspace, size)) {
1357            throw new IllegalArgumentException("size was not supported");
1358        }
1359
1360        StreamConfigurationDuration[] durations = getDurations(duration, dataspace);
1361
1362        for (StreamConfigurationDuration configurationDuration : durations) {
1363            if (configurationDuration.getFormat() == format &&
1364                    configurationDuration.getWidth() == size.getWidth() &&
1365                    configurationDuration.getHeight() == size.getHeight()) {
1366                return configurationDuration.getDuration();
1367            }
1368        }
1369        // Default duration is '0' (unsupported/no extra stall)
1370        return 0;
1371    }
1372
1373    /**
1374     * Get the durations array for the kind of duration
1375     *
1376     * @see #DURATION_MIN_FRAME
1377     * @see #DURATION_STALL
1378     * */
1379    private StreamConfigurationDuration[] getDurations(int duration, int dataspace) {
1380        switch (duration) {
1381            case DURATION_MIN_FRAME:
1382                return (dataspace == HAL_DATASPACE_DEPTH) ?
1383                        mDepthMinFrameDurations : mMinFrameDurations;
1384            case DURATION_STALL:
1385                return (dataspace == HAL_DATASPACE_DEPTH) ?
1386                        mDepthStallDurations : mStallDurations;
1387            default:
1388                throw new IllegalArgumentException("duration was invalid");
1389        }
1390    }
1391
1392    /** Count the number of publicly-visible output formats */
1393    private int getPublicFormatCount(boolean output) {
1394        SparseIntArray formatsMap = getFormatsMap(output);
1395        int size = formatsMap.size();
1396        if (output) {
1397            size += mDepthOutputFormats.size();
1398        }
1399
1400        return size;
1401    }
1402
1403    private static <T> boolean arrayContains(T[] array, T element) {
1404        if (array == null) {
1405            return false;
1406        }
1407
1408        for (T el : array) {
1409            if (Objects.equals(el, element)) {
1410                return true;
1411            }
1412        }
1413
1414        return false;
1415    }
1416
1417    private boolean isSupportedInternalConfiguration(int format, int dataspace,
1418            Size size) {
1419        StreamConfiguration[] configurations =
1420                (dataspace == HAL_DATASPACE_DEPTH) ? mDepthConfigurations : mConfigurations;
1421
1422        for (int i = 0; i < configurations.length; i++) {
1423            if (configurations[i].getFormat() == format &&
1424                    configurations[i].getSize().equals(size)) {
1425                return true;
1426            }
1427        }
1428
1429        return false;
1430    }
1431
1432    /**
1433     * Return this {@link StreamConfigurationMap} as a string representation.
1434     *
1435     * <p>{@code "StreamConfigurationMap(Outputs([w:%d, h:%d, format:%s(%d), min_duration:%d,
1436     * stall:%d], ... [w:%d, h:%d, format:%s(%d), min_duration:%d, stall:%d]), Inputs([w:%d, h:%d,
1437     * format:%s(%d)], ... [w:%d, h:%d, format:%s(%d)]), ValidOutputFormatsForInput(
1438     * [in:%d, out:%d, ... %d], ... [in:%d, out:%d, ... %d]), HighSpeedVideoConfigurations(
1439     * [w:%d, h:%d, min_fps:%d, max_fps:%d], ... [w:%d, h:%d, min_fps:%d, max_fps:%d]))"}.</p>
1440     *
1441     * <p>{@code Outputs([w:%d, h:%d, format:%s(%d), min_duration:%d, stall:%d], ...
1442     * [w:%d, h:%d, format:%s(%d), min_duration:%d, stall:%d])}, where
1443     * {@code [w:%d, h:%d, format:%s(%d), min_duration:%d, stall:%d]} represents an output
1444     * configuration's width, height, format, minimal frame duration in nanoseconds, and stall
1445     * duration in nanoseconds.</p>
1446     *
1447     * <p>{@code Inputs([w:%d, h:%d, format:%s(%d)], ... [w:%d, h:%d, format:%s(%d)])}, where
1448     * {@code [w:%d, h:%d, format:%s(%d)]} represents an input configuration's width, height, and
1449     * format.</p>
1450     *
1451     * <p>{@code ValidOutputFormatsForInput([in:%s(%d), out:%s(%d), ... %s(%d)],
1452     * ... [in:%s(%d), out:%s(%d), ... %s(%d)])}, where {@code [in:%s(%d), out:%s(%d), ... %s(%d)]}
1453     * represents an input fomat and its valid output formats.</p>
1454     *
1455     * <p>{@code HighSpeedVideoConfigurations([w:%d, h:%d, min_fps:%d, max_fps:%d],
1456     * ... [w:%d, h:%d, min_fps:%d, max_fps:%d])}, where
1457     * {@code [w:%d, h:%d, min_fps:%d, max_fps:%d]} represents a high speed video output
1458     * configuration's width, height, minimal frame rate, and maximal frame rate.</p>
1459     *
1460     * @return string representation of {@link StreamConfigurationMap}
1461     */
1462    @Override
1463    public String toString() {
1464        StringBuilder sb = new StringBuilder("StreamConfiguration(");
1465        appendOutputsString(sb);
1466        sb.append(", ");
1467        appendHighResOutputsString(sb);
1468        sb.append(", ");
1469        appendInputsString(sb);
1470        sb.append(", ");
1471        appendValidOutputFormatsForInputString(sb);
1472        sb.append(", ");
1473        appendHighSpeedVideoConfigurationsString(sb);
1474        sb.append(")");
1475
1476        return sb.toString();
1477    }
1478
1479    private void appendOutputsString(StringBuilder sb) {
1480        sb.append("Outputs(");
1481        int[] formats = getOutputFormats();
1482        for (int format : formats) {
1483            Size[] sizes = getOutputSizes(format);
1484            for (Size size : sizes) {
1485                long minFrameDuration = getOutputMinFrameDuration(format, size);
1486                long stallDuration = getOutputStallDuration(format, size);
1487                sb.append(String.format("[w:%d, h:%d, format:%s(%d), min_duration:%d, " +
1488                        "stall:%d], ", size.getWidth(), size.getHeight(), formatToString(format),
1489                        format, minFrameDuration, stallDuration));
1490            }
1491        }
1492        // Remove the pending ", "
1493        if (sb.charAt(sb.length() - 1) == ' ') {
1494            sb.delete(sb.length() - 2, sb.length());
1495        }
1496        sb.append(")");
1497    }
1498
1499    private void appendHighResOutputsString(StringBuilder sb) {
1500        sb.append("HighResolutionOutputs(");
1501        int[] formats = getOutputFormats();
1502        for (int format : formats) {
1503            Size[] sizes = getHighResolutionOutputSizes(format);
1504            if (sizes == null) continue;
1505            for (Size size : sizes) {
1506                long minFrameDuration = getOutputMinFrameDuration(format, size);
1507                long stallDuration = getOutputStallDuration(format, size);
1508                sb.append(String.format("[w:%d, h:%d, format:%s(%d), min_duration:%d, " +
1509                        "stall:%d], ", size.getWidth(), size.getHeight(), formatToString(format),
1510                        format, minFrameDuration, stallDuration));
1511            }
1512        }
1513        // Remove the pending ", "
1514        if (sb.charAt(sb.length() - 1) == ' ') {
1515            sb.delete(sb.length() - 2, sb.length());
1516        }
1517        sb.append(")");
1518    }
1519
1520    private void appendInputsString(StringBuilder sb) {
1521        sb.append("Inputs(");
1522        int[] formats = getInputFormats();
1523        for (int format : formats) {
1524            Size[] sizes = getInputSizes(format);
1525            for (Size size : sizes) {
1526                sb.append(String.format("[w:%d, h:%d, format:%s(%d)], ", size.getWidth(),
1527                        size.getHeight(), formatToString(format), format));
1528            }
1529        }
1530        // Remove the pending ", "
1531        if (sb.charAt(sb.length() - 1) == ' ') {
1532            sb.delete(sb.length() - 2, sb.length());
1533        }
1534        sb.append(")");
1535    }
1536
1537    private void appendValidOutputFormatsForInputString(StringBuilder sb) {
1538        sb.append("ValidOutputFormatsForInput(");
1539        int[] inputFormats = getInputFormats();
1540        for (int inputFormat : inputFormats) {
1541            sb.append(String.format("[in:%s(%d), out:", formatToString(inputFormat), inputFormat));
1542            int[] outputFormats = getValidOutputFormatsForInput(inputFormat);
1543            for (int i = 0; i < outputFormats.length; i++) {
1544                sb.append(String.format("%s(%d)", formatToString(outputFormats[i]),
1545                        outputFormats[i]));
1546                if (i < outputFormats.length - 1) {
1547                    sb.append(", ");
1548                }
1549            }
1550            sb.append("], ");
1551        }
1552        // Remove the pending ", "
1553        if (sb.charAt(sb.length() - 1) == ' ') {
1554            sb.delete(sb.length() - 2, sb.length());
1555        }
1556        sb.append(")");
1557    }
1558
1559    private void appendHighSpeedVideoConfigurationsString(StringBuilder sb) {
1560        sb.append("HighSpeedVideoConfigurations(");
1561        Size[] sizes = getHighSpeedVideoSizes();
1562        for (Size size : sizes) {
1563            Range<Integer>[] ranges = getHighSpeedVideoFpsRangesFor(size);
1564            for (Range<Integer> range : ranges) {
1565                sb.append(String.format("[w:%d, h:%d, min_fps:%d, max_fps:%d], ", size.getWidth(),
1566                        size.getHeight(), range.getLower(), range.getUpper()));
1567            }
1568        }
1569        // Remove the pending ", "
1570        if (sb.charAt(sb.length() - 1) == ' ') {
1571            sb.delete(sb.length() - 2, sb.length());
1572        }
1573        sb.append(")");
1574    }
1575
1576    private String formatToString(int format) {
1577        switch (format) {
1578            case ImageFormat.YV12:
1579                return "YV12";
1580            case ImageFormat.YUV_420_888:
1581                return "YUV_420_888";
1582            case ImageFormat.NV21:
1583                return "NV21";
1584            case ImageFormat.NV16:
1585                return "NV16";
1586            case PixelFormat.RGB_565:
1587                return "RGB_565";
1588            case PixelFormat.RGBA_8888:
1589                return "RGBA_8888";
1590            case PixelFormat.RGBX_8888:
1591                return "RGBX_8888";
1592            case PixelFormat.RGB_888:
1593                return "RGB_888";
1594            case ImageFormat.JPEG:
1595                return "JPEG";
1596            case ImageFormat.YUY2:
1597                return "YUY2";
1598            case ImageFormat.Y8:
1599                return "Y8";
1600            case ImageFormat.Y16:
1601                return "Y16";
1602            case ImageFormat.RAW_SENSOR:
1603                return "RAW_SENSOR";
1604            case ImageFormat.RAW_PRIVATE:
1605                return "RAW_PRIVATE";
1606            case ImageFormat.RAW10:
1607                return "RAW10";
1608            case ImageFormat.DEPTH16:
1609                return "DEPTH16";
1610            case ImageFormat.DEPTH_POINT_CLOUD:
1611                return "DEPTH_POINT_CLOUD";
1612            case ImageFormat.PRIVATE:
1613                return "PRIVATE";
1614            default:
1615                return "UNKNOWN";
1616        }
1617    }
1618
1619    // from system/core/include/system/graphics.h
1620    private static final int HAL_PIXEL_FORMAT_RAW16 = 0x20;
1621    private static final int HAL_PIXEL_FORMAT_BLOB = 0x21;
1622    private static final int HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED = 0x22;
1623    private static final int HAL_PIXEL_FORMAT_YCbCr_420_888 = 0x23;
1624    private static final int HAL_PIXEL_FORMAT_RAW_OPAQUE = 0x24;
1625    private static final int HAL_PIXEL_FORMAT_RAW10 = 0x25;
1626    private static final int HAL_PIXEL_FORMAT_RAW12 = 0x26;
1627    private static final int HAL_PIXEL_FORMAT_Y16 = 0x20363159;
1628
1629
1630    private static final int HAL_DATASPACE_STANDARD_SHIFT = 16;
1631    private static final int HAL_DATASPACE_TRANSFER_SHIFT = 22;
1632    private static final int HAL_DATASPACE_RANGE_SHIFT = 27;
1633
1634    private static final int HAL_DATASPACE_UNKNOWN = 0x0;
1635    private static final int HAL_DATASPACE_V0_JFIF =
1636            (2 << HAL_DATASPACE_STANDARD_SHIFT) |
1637            (3 << HAL_DATASPACE_TRANSFER_SHIFT) |
1638            (1 << HAL_DATASPACE_RANGE_SHIFT);
1639
1640    private static final int HAL_DATASPACE_DEPTH = 0x1000;
1641
1642    private static final long DURATION_20FPS_NS = 50000000L;
1643    /**
1644     * @see #getDurations(int, int)
1645     */
1646    private static final int DURATION_MIN_FRAME = 0;
1647    private static final int DURATION_STALL = 1;
1648
1649    private final StreamConfiguration[] mConfigurations;
1650    private final StreamConfigurationDuration[] mMinFrameDurations;
1651    private final StreamConfigurationDuration[] mStallDurations;
1652
1653    private final StreamConfiguration[] mDepthConfigurations;
1654    private final StreamConfigurationDuration[] mDepthMinFrameDurations;
1655    private final StreamConfigurationDuration[] mDepthStallDurations;
1656
1657    private final HighSpeedVideoConfiguration[] mHighSpeedVideoConfigurations;
1658    private final ReprocessFormatsMap mInputOutputFormatsMap;
1659
1660    private final boolean mListHighResolution;
1661
1662    /** internal format -> num output sizes mapping, not including slow high-res sizes, for
1663     * non-depth dataspaces */
1664    private final SparseIntArray mOutputFormats = new SparseIntArray();
1665    /** internal format -> num output sizes mapping for slow high-res sizes, for non-depth
1666     * dataspaces */
1667    private final SparseIntArray mHighResOutputFormats = new SparseIntArray();
1668    /** internal format -> num output sizes mapping for all non-depth dataspaces */
1669    private final SparseIntArray mAllOutputFormats = new SparseIntArray();
1670    /** internal format -> num input sizes mapping, for input reprocessing formats */
1671    private final SparseIntArray mInputFormats = new SparseIntArray();
1672    /** internal format -> num depth output sizes mapping, for HAL_DATASPACE_DEPTH */
1673    private final SparseIntArray mDepthOutputFormats = new SparseIntArray();
1674    /** High speed video Size -> FPS range count mapping*/
1675    private final HashMap</*HighSpeedVideoSize*/Size, /*Count*/Integer> mHighSpeedVideoSizeMap =
1676            new HashMap<Size, Integer>();
1677    /** High speed video FPS range -> Size count mapping*/
1678    private final HashMap</*HighSpeedVideoFpsRange*/Range<Integer>, /*Count*/Integer>
1679            mHighSpeedVideoFpsRangeMap = new HashMap<Range<Integer>, Integer>();
1680
1681}
1682