1/*
2 * Copyright (C) 2010 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 *      http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17package android.graphics;
18
19public class ImageFormat {
20    /*
21     * these constants are chosen to be binary compatible with their previous
22     * location in PixelFormat.java
23     */
24
25    public static final int UNKNOWN = 0;
26
27    /**
28     * RGB format used for pictures encoded as RGB_565. See
29     * {@link android.hardware.Camera.Parameters#setPictureFormat(int)}.
30     */
31    public static final int RGB_565 = 4;
32
33    /**
34     * <p>Android YUV format.</p>
35     *
36     * <p>This format is exposed to software decoders and applications.</p>
37     *
38     * <p>YV12 is a 4:2:0 YCrCb planar format comprised of a WxH Y plane followed
39     * by (W/2) x (H/2) Cr and Cb planes.</p>
40     *
41     * <p>This format assumes
42     * <ul>
43     * <li>an even width</li>
44     * <li>an even height</li>
45     * <li>a horizontal stride multiple of 16 pixels</li>
46     * <li>a vertical stride equal to the height</li>
47     * </ul>
48     * </p>
49     *
50     * <pre> y_size = stride * height
51     * c_stride = ALIGN(stride/2, 16)
52     * c_size = c_stride * height/2
53     * size = y_size + c_size * 2
54     * cr_offset = y_size
55     * cb_offset = y_size + c_size</pre>
56     *
57     * <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is
58     * recommended for YUV output instead.</p>
59     *
60     * <p>For the older camera API, this format is guaranteed to be supported for
61     * {@link android.hardware.Camera} preview images since API level 12; for earlier API versions,
62     * check {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}.
63     *
64     * <p>Note that for camera preview callback use (see
65     * {@link android.hardware.Camera#setPreviewCallback}), the
66     * <var>stride</var> value is the smallest possible; that is, it is equal
67     * to:
68     *
69     * <pre>stride = ALIGN(width, 16)</pre>
70     *
71     * @see android.hardware.Camera.Parameters#setPreviewCallback
72     * @see android.hardware.Camera.Parameters#setPreviewFormat
73     * @see android.hardware.Camera.Parameters#getSupportedPreviewFormats
74     * </p>
75     */
76    public static final int YV12 = 0x32315659;
77
78    /**
79     * <p>Android Y8 format.</p>
80     *
81     * <p>Y8 is a YUV planar format comprised of a WxH Y plane only, with each pixel
82     * being represented by 8 bits. It is equivalent to just the Y plane from {@link #YV12}
83     * format.</p>
84     *
85     * <p>This format assumes
86     * <ul>
87     * <li>an even width</li>
88     * <li>an even height</li>
89     * <li>a horizontal stride multiple of 16 pixels</li>
90     * </ul>
91     * </p>
92     *
93     * <pre> y_size = stride * height </pre>
94     *
95     * <p>For example, the {@link android.media.Image} object can provide data
96     * in this format from a {@link android.hardware.camera2.CameraDevice}
97     * through a {@link android.media.ImageReader} object if this format is
98     * supported by {@link android.hardware.camera2.CameraDevice}.</p>
99     *
100     * @see android.media.Image
101     * @see android.media.ImageReader
102     * @see android.hardware.camera2.CameraDevice
103     *
104     * @hide
105     */
106    public static final int Y8 = 0x20203859;
107
108    /**
109     * <p>Android Y16 format.</p>
110     *
111     * Y16 is a YUV planar format comprised of a WxH Y plane, with each pixel
112     * being represented by 16 bits. It is just like {@link #Y8}, but has 16
113     * bits per pixel (little endian).</p>
114     *
115     * <p>This format assumes
116     * <ul>
117     * <li>an even width</li>
118     * <li>an even height</li>
119     * <li>a horizontal stride multiple of 16 pixels</li>
120     * </ul>
121     * </p>
122     *
123     * <pre> y_size = stride * height </pre>
124     *
125     * <p>For example, the {@link android.media.Image} object can provide data
126     * in this format from a {@link android.hardware.camera2.CameraDevice}
127     * through a {@link android.media.ImageReader} object if this format is
128     * supported by {@link android.hardware.camera2.CameraDevice}.</p>
129     *
130     * @see android.media.Image
131     * @see android.media.ImageReader
132     * @see android.hardware.camera2.CameraDevice
133     *
134     * @hide
135     */
136    public static final int Y16 = 0x20363159;
137
138    /**
139     * YCbCr format, used for video.
140     *
141     * <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is
142     * recommended for YUV output instead.</p>
143     *
144     * <p>Whether this format is supported by the old camera API can be determined by
145     * {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}.</p>
146     *
147     */
148    public static final int NV16 = 0x10;
149
150    /**
151     * YCrCb format used for images, which uses the NV21 encoding format.
152     *
153     * <p>This is the default format
154     * for {@link android.hardware.Camera} preview images, when not otherwise set with
155     * {@link android.hardware.Camera.Parameters#setPreviewFormat(int)}.</p>
156     *
157     * <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is
158     * recommended for YUV output instead.</p>
159     */
160    public static final int NV21 = 0x11;
161
162    /**
163     * YCbCr format used for images, which uses YUYV (YUY2) encoding format.
164     *
165     * <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is
166     * recommended for YUV output instead.</p>
167     *
168     * <p>This is an alternative format for {@link android.hardware.Camera} preview images. Whether
169     * this format is supported by the camera hardware can be determined by
170     * {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}.</p>
171     */
172    public static final int YUY2 = 0x14;
173
174    /**
175     * Compressed JPEG format.
176     *
177     * <p>This format is always supported as an output format for the
178     * {@link android.hardware.camera2} API, and as a picture format for the older
179     * {@link android.hardware.Camera} API</p>
180     */
181    public static final int JPEG = 0x100;
182
183    /**
184     * <p>Multi-plane Android YUV 420 format</p>
185     *
186     * <p>This format is a generic YCbCr format, capable of describing any 4:2:0
187     * chroma-subsampled planar or semiplanar buffer (but not fully interleaved),
188     * with 8 bits per color sample.</p>
189     *
190     * <p>Images in this format are always represented by three separate buffers
191     * of data, one for each color plane. Additional information always
192     * accompanies the buffers, describing the row stride and the pixel stride
193     * for each plane.</p>
194     *
195     * <p>The order of planes in the array returned by
196     * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
197     * plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).</p>
198     *
199     * <p>The Y-plane is guaranteed not to be interleaved with the U/V planes
200     * (in particular, pixel stride is always 1 in
201     * {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}).</p>
202     *
203     * <p>The U/V planes are guaranteed to have the same row stride and pixel stride
204     * (in particular,
205     * {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()}
206     * == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and
207     * {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()}
208     * == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()};
209     * ).</p>
210     *
211     * <p>For example, the {@link android.media.Image} object can provide data
212     * in this format from a {@link android.hardware.camera2.CameraDevice}
213     * through a {@link android.media.ImageReader} object.</p>
214     *
215     * @see android.media.Image
216     * @see android.media.ImageReader
217     * @see android.hardware.camera2.CameraDevice
218     */
219    public static final int YUV_420_888 = 0x23;
220
221    /**
222     * <p>Multi-plane Android YUV 422 format</p>
223     *
224     * <p>This format is a generic YCbCr format, capable of describing any 4:2:2
225     * chroma-subsampled (planar, semiplanar or interleaved) format,
226     * with 8 bits per color sample.</p>
227     *
228     * <p>Images in this format are always represented by three separate buffers
229     * of data, one for each color plane. Additional information always
230     * accompanies the buffers, describing the row stride and the pixel stride
231     * for each plane.</p>
232     *
233     * <p>The order of planes in the array returned by
234     * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
235     * plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).</p>
236     *
237     * <p>In contrast to the {@link #YUV_420_888} format, the Y-plane may have a pixel
238     * stride greater than 1 in
239     * {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}.</p>
240     *
241     * <p>The U/V planes are guaranteed to have the same row stride and pixel stride
242     * (in particular,
243     * {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()}
244     * == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and
245     * {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()}
246     * == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()};
247     * ).</p>
248     *
249     * <p>For example, the {@link android.media.Image} object can provide data
250     * in this format from a {@link android.media.MediaCodec}
251     * through {@link android.media.MediaCodec#getOutputImage} object.</p>
252     *
253     * @see android.media.Image
254     * @see android.media.MediaCodec
255     */
256    public static final int YUV_422_888 = 0x27;
257
258    /**
259     * <p>Multi-plane Android YUV 444 format</p>
260     *
261     * <p>This format is a generic YCbCr format, capable of describing any 4:4:4
262     * (planar, semiplanar or interleaved) format,
263     * with 8 bits per color sample.</p>
264     *
265     * <p>Images in this format are always represented by three separate buffers
266     * of data, one for each color plane. Additional information always
267     * accompanies the buffers, describing the row stride and the pixel stride
268     * for each plane.</p>
269     *
270     * <p>The order of planes in the array returned by
271     * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
272     * plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).</p>
273     *
274     * <p>In contrast to the {@link #YUV_420_888} format, the Y-plane may have a pixel
275     * stride greater than 1 in
276     * {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}.</p>
277     *
278     * <p>The U/V planes are guaranteed to have the same row stride and pixel stride
279     * (in particular,
280     * {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()}
281     * == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and
282     * {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()}
283     * == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()};
284     * ).</p>
285     *
286     * <p>For example, the {@link android.media.Image} object can provide data
287     * in this format from a {@link android.media.MediaCodec}
288     * through {@link android.media.MediaCodec#getOutputImage} object.</p>
289     *
290     * @see android.media.Image
291     * @see android.media.MediaCodec
292     */
293    public static final int YUV_444_888 = 0x28;
294
295    /**
296     * <p>Multi-plane Android RGB format</p>
297     *
298     * <p>This format is a generic RGB format, capable of describing most RGB formats,
299     * with 8 bits per color sample.</p>
300     *
301     * <p>Images in this format are always represented by three separate buffers
302     * of data, one for each color plane. Additional information always
303     * accompanies the buffers, describing the row stride and the pixel stride
304     * for each plane.</p>
305     *
306     * <p>The order of planes in the array returned by
307     * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
308     * plane #0 is always R (red), plane #1 is always G (green), and plane #2 is always B
309     * (blue).</p>
310     *
311     * <p>All three planes are guaranteed to have the same row strides and pixel strides.</p>
312     *
313     * <p>For example, the {@link android.media.Image} object can provide data
314     * in this format from a {@link android.media.MediaCodec}
315     * through {@link android.media.MediaCodec#getOutputImage} object.</p>
316     *
317     * @see android.media.Image
318     * @see android.media.MediaCodec
319     */
320    public static final int FLEX_RGB_888 = 0x29;
321
322    /**
323     * <p>Multi-plane Android RGBA format</p>
324     *
325     * <p>This format is a generic RGBA format, capable of describing most RGBA formats,
326     * with 8 bits per color sample.</p>
327     *
328     * <p>Images in this format are always represented by four separate buffers
329     * of data, one for each color plane. Additional information always
330     * accompanies the buffers, describing the row stride and the pixel stride
331     * for each plane.</p>
332     *
333     * <p>The order of planes in the array returned by
334     * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
335     * plane #0 is always R (red), plane #1 is always G (green), plane #2 is always B (blue),
336     * and plane #3 is always A (alpha). This format may represent pre-multiplied or
337     * non-premultiplied alpha.</p>
338     *
339     * <p>All four planes are guaranteed to have the same row strides and pixel strides.</p>
340     *
341     * <p>For example, the {@link android.media.Image} object can provide data
342     * in this format from a {@link android.media.MediaCodec}
343     * through {@link android.media.MediaCodec#getOutputImage} object.</p>
344     *
345     * @see android.media.Image
346     * @see android.media.MediaCodec
347     */
348    public static final int FLEX_RGBA_8888 = 0x2A;
349
350    /**
351     * <p>General raw camera sensor image format, usually representing a
352     * single-channel Bayer-mosaic image. Each pixel color sample is stored with
353     * 16 bits of precision.</p>
354     *
355     * <p>The layout of the color mosaic, the maximum and minimum encoding
356     * values of the raw pixel data, the color space of the image, and all other
357     * needed information to interpret a raw sensor image must be queried from
358     * the {@link android.hardware.camera2.CameraDevice} which produced the
359     * image.</p>
360     */
361    public static final int RAW_SENSOR = 0x20;
362
363    /**
364     * <p>Private raw camera sensor image format, a single channel image with
365     * implementation depedent pixel layout.</p>
366     *
367     * <p>RAW_PRIVATE is a format for unprocessed raw image buffers coming from an
368     * image sensor. The actual structure of buffers of this format is
369     * implementation-dependent.</p>
370     *
371     */
372    public static final int RAW_PRIVATE = 0x24;
373
374    /**
375     * <p>
376     * Android 10-bit raw format
377     * </p>
378     * <p>
379     * This is a single-plane, 10-bit per pixel, densely packed (in each row),
380     * unprocessed format, usually representing raw Bayer-pattern images coming
381     * from an image sensor.
382     * </p>
383     * <p>
384     * In an image buffer with this format, starting from the first pixel of
385     * each row, each 4 consecutive pixels are packed into 5 bytes (40 bits).
386     * Each one of the first 4 bytes contains the top 8 bits of each pixel, The
387     * fifth byte contains the 2 least significant bits of the 4 pixels, the
388     * exact layout data for each 4 consecutive pixels is illustrated below
389     * ({@code Pi[j]} stands for the jth bit of the ith pixel):
390     * </p>
391     * <table>
392     * <thead>
393     * <tr>
394     * <th align="center"></th>
395     * <th align="center">bit 7</th>
396     * <th align="center">bit 6</th>
397     * <th align="center">bit 5</th>
398     * <th align="center">bit 4</th>
399     * <th align="center">bit 3</th>
400     * <th align="center">bit 2</th>
401     * <th align="center">bit 1</th>
402     * <th align="center">bit 0</th>
403     * </tr>
404     * </thead> <tbody>
405     * <tr>
406     * <td align="center">Byte 0:</td>
407     * <td align="center">P0[9]</td>
408     * <td align="center">P0[8]</td>
409     * <td align="center">P0[7]</td>
410     * <td align="center">P0[6]</td>
411     * <td align="center">P0[5]</td>
412     * <td align="center">P0[4]</td>
413     * <td align="center">P0[3]</td>
414     * <td align="center">P0[2]</td>
415     * </tr>
416     * <tr>
417     * <td align="center">Byte 1:</td>
418     * <td align="center">P1[9]</td>
419     * <td align="center">P1[8]</td>
420     * <td align="center">P1[7]</td>
421     * <td align="center">P1[6]</td>
422     * <td align="center">P1[5]</td>
423     * <td align="center">P1[4]</td>
424     * <td align="center">P1[3]</td>
425     * <td align="center">P1[2]</td>
426     * </tr>
427     * <tr>
428     * <td align="center">Byte 2:</td>
429     * <td align="center">P2[9]</td>
430     * <td align="center">P2[8]</td>
431     * <td align="center">P2[7]</td>
432     * <td align="center">P2[6]</td>
433     * <td align="center">P2[5]</td>
434     * <td align="center">P2[4]</td>
435     * <td align="center">P2[3]</td>
436     * <td align="center">P2[2]</td>
437     * </tr>
438     * <tr>
439     * <td align="center">Byte 3:</td>
440     * <td align="center">P3[9]</td>
441     * <td align="center">P3[8]</td>
442     * <td align="center">P3[7]</td>
443     * <td align="center">P3[6]</td>
444     * <td align="center">P3[5]</td>
445     * <td align="center">P3[4]</td>
446     * <td align="center">P3[3]</td>
447     * <td align="center">P3[2]</td>
448     * </tr>
449     * <tr>
450     * <td align="center">Byte 4:</td>
451     * <td align="center">P3[1]</td>
452     * <td align="center">P3[0]</td>
453     * <td align="center">P2[1]</td>
454     * <td align="center">P2[0]</td>
455     * <td align="center">P1[1]</td>
456     * <td align="center">P1[0]</td>
457     * <td align="center">P0[1]</td>
458     * <td align="center">P0[0]</td>
459     * </tr>
460     * </tbody>
461     * </table>
462     * <p>
463     * This format assumes
464     * <ul>
465     * <li>a width multiple of 4 pixels</li>
466     * <li>an even height</li>
467     * </ul>
468     * </p>
469     *
470     * <pre>size = row stride * height</pre> where the row stride is in <em>bytes</em>,
471     * not pixels.
472     *
473     * <p>
474     * Since this is a densely packed format, the pixel stride is always 0. The
475     * application must use the pixel data layout defined in above table to
476     * access each row data. When row stride is equal to {@code width * (10 / 8)}, there
477     * will be no padding bytes at the end of each row, the entire image data is
478     * densely packed. When stride is larger than {@code width * (10 / 8)}, padding
479     * bytes will be present at the end of each row.
480     * </p>
481     * <p>
482     * For example, the {@link android.media.Image} object can provide data in
483     * this format from a {@link android.hardware.camera2.CameraDevice} (if
484     * supported) through a {@link android.media.ImageReader} object. The
485     * {@link android.media.Image#getPlanes() Image#getPlanes()} will return a
486     * single plane containing the pixel data. The pixel stride is always 0 in
487     * {@link android.media.Image.Plane#getPixelStride()}, and the
488     * {@link android.media.Image.Plane#getRowStride()} describes the vertical
489     * neighboring pixel distance (in bytes) between adjacent rows.
490     * </p>
491     *
492     * @see android.media.Image
493     * @see android.media.ImageReader
494     * @see android.hardware.camera2.CameraDevice
495     */
496    public static final int RAW10 = 0x25;
497
498    /**
499     * <p>
500     * Android 12-bit raw format
501     * </p>
502     * <p>
503     * This is a single-plane, 12-bit per pixel, densely packed (in each row),
504     * unprocessed format, usually representing raw Bayer-pattern images coming
505     * from an image sensor.
506     * </p>
507     * <p>
508     * In an image buffer with this format, starting from the first pixel of each
509     * row, each two consecutive pixels are packed into 3 bytes (24 bits). The first
510     * and second byte contains the top 8 bits of first and second pixel. The third
511     * byte contains the 4 least significant bits of the two pixels, the exact layout
512     * data for each two consecutive pixels is illustrated below (Pi[j] stands for
513     * the jth bit of the ith pixel):
514     * </p>
515     * <table>
516     * <thead>
517     * <tr>
518     * <th align="center"></th>
519     * <th align="center">bit 7</th>
520     * <th align="center">bit 6</th>
521     * <th align="center">bit 5</th>
522     * <th align="center">bit 4</th>
523     * <th align="center">bit 3</th>
524     * <th align="center">bit 2</th>
525     * <th align="center">bit 1</th>
526     * <th align="center">bit 0</th>
527     * </tr>
528     * </thead> <tbody>
529     * <tr>
530     * <td align="center">Byte 0:</td>
531     * <td align="center">P0[11]</td>
532     * <td align="center">P0[10]</td>
533     * <td align="center">P0[ 9]</td>
534     * <td align="center">P0[ 8]</td>
535     * <td align="center">P0[ 7]</td>
536     * <td align="center">P0[ 6]</td>
537     * <td align="center">P0[ 5]</td>
538     * <td align="center">P0[ 4]</td>
539     * </tr>
540     * <tr>
541     * <td align="center">Byte 1:</td>
542     * <td align="center">P1[11]</td>
543     * <td align="center">P1[10]</td>
544     * <td align="center">P1[ 9]</td>
545     * <td align="center">P1[ 8]</td>
546     * <td align="center">P1[ 7]</td>
547     * <td align="center">P1[ 6]</td>
548     * <td align="center">P1[ 5]</td>
549     * <td align="center">P1[ 4]</td>
550     * </tr>
551     * <tr>
552     * <td align="center">Byte 2:</td>
553     * <td align="center">P1[ 3]</td>
554     * <td align="center">P1[ 2]</td>
555     * <td align="center">P1[ 1]</td>
556     * <td align="center">P1[ 0]</td>
557     * <td align="center">P0[ 3]</td>
558     * <td align="center">P0[ 2]</td>
559     * <td align="center">P0[ 1]</td>
560     * <td align="center">P0[ 0]</td>
561     * </tr>
562     * </tbody>
563     * </table>
564     * <p>
565     * This format assumes
566     * <ul>
567     * <li>a width multiple of 4 pixels</li>
568     * <li>an even height</li>
569     * </ul>
570     * </p>
571     *
572     * <pre>size = row stride * height</pre> where the row stride is in <em>bytes</em>,
573     * not pixels.
574     *
575     * <p>
576     * Since this is a densely packed format, the pixel stride is always 0. The
577     * application must use the pixel data layout defined in above table to
578     * access each row data. When row stride is equal to {@code width * (12 / 8)}, there
579     * will be no padding bytes at the end of each row, the entire image data is
580     * densely packed. When stride is larger than {@code width * (12 / 8)}, padding
581     * bytes will be present at the end of each row.
582     * </p>
583     * <p>
584     * For example, the {@link android.media.Image} object can provide data in
585     * this format from a {@link android.hardware.camera2.CameraDevice} (if
586     * supported) through a {@link android.media.ImageReader} object. The
587     * {@link android.media.Image#getPlanes() Image#getPlanes()} will return a
588     * single plane containing the pixel data. The pixel stride is always 0 in
589     * {@link android.media.Image.Plane#getPixelStride()}, and the
590     * {@link android.media.Image.Plane#getRowStride()} describes the vertical
591     * neighboring pixel distance (in bytes) between adjacent rows.
592     * </p>
593     *
594     * @see android.media.Image
595     * @see android.media.ImageReader
596     * @see android.hardware.camera2.CameraDevice
597     */
598    public static final int RAW12 = 0x26;
599
600    /**
601     * <p>Android dense depth image format.</p>
602     *
603     * <p>Each pixel is 16 bits, representing a depth ranging measurement from a depth camera or
604     * similar sensor. The 16-bit sample consists of a confidence value and the actual ranging
605     * measurement.</p>
606     *
607     * <p>The confidence value is an estimate of correctness for this sample.  It is encoded in the
608     * 3 most significant bits of the sample, with a value of 0 representing 100% confidence, a
609     * value of 1 representing 0% confidence, a value of 2 representing 1/7, a value of 3
610     * representing 2/7, and so on.</p>
611     *
612     * <p>As an example, the following sample extracts the range and confidence from the first pixel
613     * of a DEPTH16-format {@link android.media.Image}, and converts the confidence to a
614     * floating-point value between 0 and 1.f inclusive, with 1.f representing maximum confidence:
615     *
616     * <pre>
617     *    ShortBuffer shortDepthBuffer = img.getPlanes()[0].getBuffer().asShortBuffer();
618     *    short depthSample = shortDepthBuffer.get()
619     *    short depthRange = (short) (depthSample & 0x1FFF);
620     *    short depthConfidence = (short) ((depthSample >> 13) & 0x7);
621     *    float depthPercentage = depthConfidence == 0 ? 1.f : (depthConfidence - 1) / 7.f;
622     * </pre>
623     * </p>
624     *
625     * <p>This format assumes
626     * <ul>
627     * <li>an even width</li>
628     * <li>an even height</li>
629     * <li>a horizontal stride multiple of 16 pixels</li>
630     * </ul>
631     * </p>
632     *
633     * <pre> y_size = stride * height </pre>
634     *
635     * When produced by a camera, the units for the range are millimeters.
636     */
637    public static final int DEPTH16 = 0x44363159;
638
639    /**
640     * Android sparse depth point cloud format.
641     *
642     * <p>A variable-length list of 3D points plus a confidence value, with each point represented
643     * by four floats; first the X, Y, Z position coordinates, and then the confidence value.</p>
644     *
645     * <p>The number of points is {@code (size of the buffer in bytes) / 16}.
646     *
647     * <p>The coordinate system and units of the position values depend on the source of the point
648     * cloud data. The confidence value is between 0.f and 1.f, inclusive, with 0 representing 0%
649     * confidence and 1.f representing 100% confidence in the measured position values.</p>
650     *
651     * <p>As an example, the following code extracts the first depth point in a DEPTH_POINT_CLOUD
652     * format {@link android.media.Image}:
653     * <pre>
654     *    FloatBuffer floatDepthBuffer = img.getPlanes()[0].getBuffer().asFloatBuffer();
655     *    float x = floatDepthBuffer.get();
656     *    float y = floatDepthBuffer.get();
657     *    float z = floatDepthBuffer.get();
658     *    float confidence = floatDepthBuffer.get();
659     * </pre>
660     *
661     */
662    public static final int DEPTH_POINT_CLOUD = 0x101;
663
664    /**
665     * Android private opaque image format.
666     * <p>
667     * The choices of the actual format and pixel data layout are entirely up to
668     * the device-specific and framework internal implementations, and may vary
669     * depending on use cases even for the same device. The buffers of this
670     * format can be produced by components like
671     * {@link android.media.ImageWriter ImageWriter} , and interpreted correctly
672     * by consumers like {@link android.hardware.camera2.CameraDevice
673     * CameraDevice} based on the device/framework private information. However,
674     * these buffers are not directly accessible to the application.
675     * </p>
676     * <p>
677     * When an {@link android.media.Image Image} of this format is obtained from
678     * an {@link android.media.ImageReader ImageReader} or
679     * {@link android.media.ImageWriter ImageWriter}, the
680     * {@link android.media.Image#getPlanes() getPlanes()} method will return an
681     * empty {@link android.media.Image.Plane Plane} array.
682     * </p>
683     * <p>
684     * If a buffer of this format is to be used as an OpenGL ES texture, the
685     * framework will assume that sampling the texture will always return an
686     * alpha value of 1.0 (i.e. the buffer contains only opaque pixel values).
687     * </p>
688     */
689    public static final int PRIVATE = 0x22;
690
691    /**
692     * Use this function to retrieve the number of bits per pixel of an
693     * ImageFormat.
694     *
695     * @param format
696     * @return the number of bits per pixel of the given format or -1 if the
697     *         format doesn't exist or is not supported.
698     */
699    public static int getBitsPerPixel(int format) {
700        switch (format) {
701            case RGB_565:
702                return 16;
703            case NV16:
704                return 16;
705            case YUY2:
706                return 16;
707            case YV12:
708                return 12;
709            case Y8:
710                return 8;
711            case Y16:
712            case DEPTH16:
713                return 16;
714            case NV21:
715                return 12;
716            case YUV_420_888:
717                return 12;
718            case YUV_422_888:
719                return 16;
720            case YUV_444_888:
721                return 24;
722            case FLEX_RGB_888:
723                return 24;
724            case FLEX_RGBA_8888:
725                return 32;
726            case RAW_SENSOR:
727                return 16;
728            case RAW10:
729                return 10;
730            case RAW12:
731                return 12;
732        }
733        return -1;
734    }
735
736    /**
737     * Determine whether or not this is a public-visible {@code format}.
738     *
739     * <p>In particular, {@code @hide} formats will return {@code false}.</p>
740     *
741     * <p>Any other formats (including UNKNOWN) will return {@code false}.</p>
742     *
743     * @param format an integer format
744     * @return a boolean
745     *
746     * @hide
747     */
748    public static boolean isPublicFormat(int format) {
749        switch (format) {
750            case RGB_565:
751            case NV16:
752            case YUY2:
753            case YV12:
754            case JPEG:
755            case NV21:
756            case YUV_420_888:
757            case YUV_422_888:
758            case YUV_444_888:
759            case FLEX_RGB_888:
760            case FLEX_RGBA_8888:
761            case RAW_SENSOR:
762            case RAW_PRIVATE:
763            case RAW10:
764            case RAW12:
765            case DEPTH16:
766            case DEPTH_POINT_CLOUD:
767            case PRIVATE:
768                return true;
769        }
770
771        return false;
772    }
773}
774