| /* |
| * Copyright (C) 2010 The Android Open Source Project |
| * |
| * Licensed under the Apache License, Version 2.0 (the "License"); |
| * you may not use this file except in compliance with the License. |
| * You may obtain a copy of the License at |
| * |
| * http://www.apache.org/licenses/LICENSE-2.0 |
| * |
| * Unless required by applicable law or agreed to in writing, software |
| * distributed under the License is distributed on an "AS IS" BASIS, |
| * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
| * See the License for the specific language governing permissions and |
| * limitations under the License. |
| */ |
| |
| package android.graphics; |
| |
| import android.annotation.FlaggedApi; |
| import android.annotation.IntDef; |
| |
| import java.lang.annotation.Retention; |
| import java.lang.annotation.RetentionPolicy; |
| |
| public class ImageFormat { |
| /** @hide */ |
| @Retention(RetentionPolicy.SOURCE) |
| @IntDef(value = { |
| UNKNOWN, |
| /* |
| * Since some APIs accept either ImageFormat or PixelFormat (and the two |
| * enums do not overlap since they're both partial versions of the |
| * internal format enum), add PixelFormat values here so linting |
| * tools won't complain when method arguments annotated with |
| * ImageFormat are provided with PixelFormat values. |
| */ |
| PixelFormat.RGBA_8888, |
| PixelFormat.RGBX_8888, |
| PixelFormat.RGB_888, |
| RGB_565, |
| YV12, |
| Y8, |
| Y16, |
| YCBCR_P010, |
| YCBCR_P210, |
| NV16, |
| NV21, |
| YUY2, |
| JPEG, |
| DEPTH_JPEG, |
| YUV_420_888, |
| YUV_422_888, |
| YUV_444_888, |
| FLEX_RGB_888, |
| FLEX_RGBA_8888, |
| RAW_SENSOR, |
| RAW_PRIVATE, |
| RAW10, |
| RAW12, |
| DEPTH16, |
| DEPTH_POINT_CLOUD, |
| RAW_DEPTH, |
| RAW_DEPTH10, |
| PRIVATE, |
| HEIC, |
| JPEG_R |
| }) |
| public @interface Format { |
| } |
| |
| /* |
| * these constants are chosen to be binary compatible with their previous |
| * location in PixelFormat.java |
| */ |
| |
| public static final int UNKNOWN = 0; |
| |
| /** |
| * RGB format used for pictures encoded as RGB_565. See |
| * {@link android.hardware.Camera.Parameters#setPictureFormat(int)}. |
| */ |
| public static final int RGB_565 = 4; |
| |
| /** |
| * <p>Android YUV format.</p> |
| * |
| * <p>This format is exposed to software decoders and applications.</p> |
| * |
| * <p>YV12 is a 4:2:0 YCrCb planar format comprised of a WxH Y plane followed |
| * by (W/2) x (H/2) Cr and Cb planes.</p> |
| * |
| * <p>This format assumes |
| * <ul> |
| * <li>an even width</li> |
| * <li>an even height</li> |
| * <li>a horizontal stride multiple of 16 pixels</li> |
| * <li>a vertical stride equal to the height</li> |
| * </ul> |
| * </p> |
| * |
| * <pre> y_size = stride * height |
| * c_stride = ALIGN(stride/2, 16) |
| * c_size = c_stride * height/2 |
| * size = y_size + c_size * 2 |
| * cr_offset = y_size |
| * cb_offset = y_size + c_size</pre> |
| * |
| * <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is |
| * recommended for YUV output instead.</p> |
| * |
| * <p>For the older camera API, this format is guaranteed to be supported for |
| * {@link android.hardware.Camera} preview images since API level 12; for earlier API versions, |
| * check {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}. |
| * |
| * <p>Note that for camera preview callback use (see |
| * {@link android.hardware.Camera#setPreviewCallback}), the |
| * <var>stride</var> value is the smallest possible; that is, it is equal |
| * to: |
| * |
| * <pre>stride = ALIGN(width, 16)</pre> |
| * |
| * @see android.hardware.Camera.Parameters#setPreviewCallback |
| * @see android.hardware.Camera.Parameters#setPreviewFormat |
| * @see android.hardware.Camera.Parameters#getSupportedPreviewFormats |
| * </p> |
| */ |
| public static final int YV12 = 0x32315659; |
| |
| /** |
| * <p>Android Y8 format.</p> |
| * |
| * <p>Y8 is a YUV planar format comprised of a WxH Y plane only, with each pixel |
| * being represented by 8 bits. It is equivalent to just the Y plane from {@link #YV12} |
| * format.</p> |
| * |
| * <p>This format assumes |
| * <ul> |
| * <li>an even width</li> |
| * <li>an even height</li> |
| * <li>a horizontal stride multiple of 16 pixels</li> |
| * </ul> |
| * </p> |
| * |
| * <pre> size = stride * height </pre> |
| * |
| * <p>For example, the {@link android.media.Image} object can provide data |
| * in this format from a {@link android.hardware.camera2.CameraDevice} (if |
| * supported) through a {@link android.media.ImageReader} object. The |
| * {@link android.media.Image#getPlanes() Image#getPlanes()} will return a |
| * single plane containing the pixel data. The pixel stride is always 1 in |
| * {@link android.media.Image.Plane#getPixelStride()}, and the |
| * {@link android.media.Image.Plane#getRowStride()} describes the vertical |
| * neighboring pixel distance (in bytes) between adjacent rows.</p> |
| * |
| * @see android.media.Image |
| * @see android.media.ImageReader |
| * @see android.hardware.camera2.CameraDevice |
| */ |
| public static final int Y8 = 0x20203859; |
| |
| /** |
| * <p>Android Y16 format.</p> |
| * |
| * Y16 is a YUV planar format comprised of a WxH Y plane, with each pixel |
| * being represented by 16 bits. It is just like {@link #Y8}, but has 16 |
| * bits per pixel (little endian).</p> |
| * |
| * <p>This format assumes |
| * <ul> |
| * <li>an even width</li> |
| * <li>an even height</li> |
| * <li>a horizontal stride multiple of 16 pixels</li> |
| * </ul> |
| * </p> |
| * |
| * <pre> y_size = stride * height </pre> |
| * |
| * <p>For example, the {@link android.media.Image} object can provide data |
| * in this format from a {@link android.hardware.camera2.CameraDevice} |
| * through a {@link android.media.ImageReader} object if this format is |
| * supported by {@link android.hardware.camera2.CameraDevice}.</p> |
| * |
| * @see android.media.Image |
| * @see android.media.ImageReader |
| * @see android.hardware.camera2.CameraDevice |
| * |
| * @hide |
| */ |
| public static final int Y16 = 0x20363159; |
| |
| /** |
| * <p>Android YUV P010 format.</p> |
| * |
| * P010 is a 4:2:0 YCbCr semiplanar format comprised of a WxH Y plane |
| * followed by a Wx(H/2) CbCr plane. Each sample is represented by a 16-bit |
| * little-endian value, with the lower 6 bits set to zero. |
| * |
| * <p>For example, the {@link android.media.Image} object can provide data |
| * in this format from a {@link android.hardware.camera2.CameraDevice} |
| * through a {@link android.media.ImageReader} object if this format is |
| * supported by {@link android.hardware.camera2.CameraDevice}.</p> |
| * |
| * @see android.media.Image |
| * @see android.media.ImageReader |
| * @see android.hardware.camera2.CameraDevice |
| * |
| */ |
| public static final int YCBCR_P010 = 0x36; |
| |
| /** |
| * <p>Android YUV P210 format.</p> |
| * |
| * P210 is a 4:2:2 YCbCr semiplanar format comprised of a WxH Y plane |
| * followed by a WxH CbCr plane. Each sample is represented by a 16-bit |
| * little-endian value, with the lower 6 bits set to zero. |
| * |
| * <p>For example, the {@link android.media.Image} object can provide data |
| * in this format from a {@link android.hardware.camera2.CameraDevice} |
| * through a {@link android.media.ImageReader} object if this format is |
| * supported by {@link android.hardware.camera2.CameraDevice}.</p> |
| * |
| * @see android.media.Image |
| * @see android.media.ImageReader |
| * @see android.hardware.camera2.CameraDevice |
| * |
| */ |
| @FlaggedApi(android.media.codec.Flags.FLAG_P210_FORMAT_SUPPORT) |
| public static final int YCBCR_P210 = 0x3c; |
| |
| /** |
| * YCbCr format, used for video. |
| * |
| * <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is |
| * recommended for YUV output instead.</p> |
| * |
| * <p>Whether this format is supported by the old camera API can be determined by |
| * {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}.</p> |
| * |
| */ |
| public static final int NV16 = 0x10; |
| |
| /** |
| * YCrCb format used for images, which uses the NV21 encoding format. |
| * |
| * <p>This is the default format |
| * for {@link android.hardware.Camera} preview images, when not otherwise set with |
| * {@link android.hardware.Camera.Parameters#setPreviewFormat(int)}.</p> |
| * |
| * <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is |
| * recommended for YUV output instead.</p> |
| */ |
| public static final int NV21 = 0x11; |
| |
| /** |
| * YCbCr format used for images, which uses YUYV (YUY2) encoding format. |
| * |
| * <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is |
| * recommended for YUV output instead.</p> |
| * |
| * <p>This is an alternative format for {@link android.hardware.Camera} preview images. Whether |
| * this format is supported by the camera hardware can be determined by |
| * {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}.</p> |
| */ |
| public static final int YUY2 = 0x14; |
| |
| /** |
| * Compressed JPEG format. |
| * |
| * <p>This format is always supported as an output format for the |
| * {@link android.hardware.camera2} API, and as a picture format for the older |
| * {@link android.hardware.Camera} API</p> |
| */ |
| public static final int JPEG = 0x100; |
| |
| /** |
| * Depth augmented compressed JPEG format. |
| * |
| * <p>JPEG compressed main image along with XMP embedded depth metadata |
| * following ISO 16684-1:2011(E).</p> |
| */ |
| public static final int DEPTH_JPEG = 0x69656963; |
| |
| /** |
| * Compressed JPEG format that includes an embedded recovery map. |
| * |
| * <p>JPEG compressed main image along with embedded recovery map following the |
| * <a href="https://developer.android.com/guide/topics/media/hdr-image-format">Ultra HDR |
| * Image format specification</a>.</p> |
| */ |
| public static final int JPEG_R = 0x1005; |
| |
| /** |
| * <p>Multi-plane Android YUV 420 format</p> |
| * |
| * <p>This format is a generic YCbCr format, capable of describing any 4:2:0 |
| * chroma-subsampled planar or semiplanar buffer (but not fully interleaved), |
| * with 8 bits per color sample.</p> |
| * |
| * <p>Images in this format are always represented by three separate buffers |
| * of data, one for each color plane. Additional information always |
| * accompanies the buffers, describing the row stride and the pixel stride |
| * for each plane.</p> |
| * |
| * <p>The order of planes in the array returned by |
| * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that |
| * plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).</p> |
| * |
| * <p>The Y-plane is guaranteed not to be interleaved with the U/V planes |
| * (in particular, pixel stride is always 1 in |
| * {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}).</p> |
| * |
| * <p>The U/V planes are guaranteed to have the same row stride and pixel stride |
| * (in particular, |
| * {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()} |
| * == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and |
| * {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()} |
| * == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()}; |
| * ).</p> |
| * |
| * <p>For example, the {@link android.media.Image} object can provide data |
| * in this format from a {@link android.hardware.camera2.CameraDevice} |
| * through a {@link android.media.ImageReader} object.</p> |
| * |
| * @see android.media.Image |
| * @see android.media.ImageReader |
| * @see android.hardware.camera2.CameraDevice |
| */ |
| public static final int YUV_420_888 = 0x23; |
| |
| /** |
| * <p>Multi-plane Android YUV 422 format</p> |
| * |
| * <p>This format is a generic YCbCr format, capable of describing any 4:2:2 |
| * chroma-subsampled (planar, semiplanar or interleaved) format, |
| * with 8 bits per color sample.</p> |
| * |
| * <p>Images in this format are always represented by three separate buffers |
| * of data, one for each color plane. Additional information always |
| * accompanies the buffers, describing the row stride and the pixel stride |
| * for each plane.</p> |
| * |
| * <p>The order of planes in the array returned by |
| * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that |
| * plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).</p> |
| * |
| * <p>In contrast to the {@link #YUV_420_888} format, the Y-plane may have a pixel |
| * stride greater than 1 in |
| * {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}.</p> |
| * |
| * <p>The U/V planes are guaranteed to have the same row stride and pixel stride |
| * (in particular, |
| * {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()} |
| * == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and |
| * {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()} |
| * == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()}; |
| * ).</p> |
| * |
| * <p>For example, the {@link android.media.Image} object can provide data |
| * in this format from a {@link android.media.MediaCodec} |
| * through {@link android.media.MediaCodec#getOutputImage} object.</p> |
| * |
| * @see android.media.Image |
| * @see android.media.MediaCodec |
| */ |
| public static final int YUV_422_888 = 0x27; |
| |
| /** |
| * <p>Multi-plane Android YUV 444 format</p> |
| * |
| * <p>This format is a generic YCbCr format, capable of describing any 4:4:4 |
| * (planar, semiplanar or interleaved) format, |
| * with 8 bits per color sample.</p> |
| * |
| * <p>Images in this format are always represented by three separate buffers |
| * of data, one for each color plane. Additional information always |
| * accompanies the buffers, describing the row stride and the pixel stride |
| * for each plane.</p> |
| * |
| * <p>The order of planes in the array returned by |
| * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that |
| * plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).</p> |
| * |
| * <p>In contrast to the {@link #YUV_420_888} format, the Y-plane may have a pixel |
| * stride greater than 1 in |
| * {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}.</p> |
| * |
| * <p>The U/V planes are guaranteed to have the same row stride and pixel stride |
| * (in particular, |
| * {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()} |
| * == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and |
| * {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()} |
| * == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()}; |
| * ).</p> |
| * |
| * <p>For example, the {@link android.media.Image} object can provide data |
| * in this format from a {@link android.media.MediaCodec} |
| * through {@link android.media.MediaCodec#getOutputImage} object.</p> |
| * |
| * @see android.media.Image |
| * @see android.media.MediaCodec |
| */ |
| public static final int YUV_444_888 = 0x28; |
| |
| /** |
| * <p>Multi-plane Android RGB format</p> |
| * |
| * <p>This format is a generic RGB format, capable of describing most RGB formats, |
| * with 8 bits per color sample.</p> |
| * |
| * <p>Images in this format are always represented by three separate buffers |
| * of data, one for each color plane. Additional information always |
| * accompanies the buffers, describing the row stride and the pixel stride |
| * for each plane.</p> |
| * |
| * <p>The order of planes in the array returned by |
| * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that |
| * plane #0 is always R (red), plane #1 is always G (green), and plane #2 is always B |
| * (blue).</p> |
| * |
| * <p>All three planes are guaranteed to have the same row strides and pixel strides.</p> |
| * |
| * <p>For example, the {@link android.media.Image} object can provide data |
| * in this format from a {@link android.media.MediaCodec} |
| * through {@link android.media.MediaCodec#getOutputImage} object.</p> |
| * |
| * @see android.media.Image |
| * @see android.media.MediaCodec |
| */ |
| public static final int FLEX_RGB_888 = 0x29; |
| |
| /** |
| * <p>Multi-plane Android RGBA format</p> |
| * |
| * <p>This format is a generic RGBA format, capable of describing most RGBA formats, |
| * with 8 bits per color sample.</p> |
| * |
| * <p>Images in this format are always represented by four separate buffers |
| * of data, one for each color plane. Additional information always |
| * accompanies the buffers, describing the row stride and the pixel stride |
| * for each plane.</p> |
| * |
| * <p>The order of planes in the array returned by |
| * {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that |
| * plane #0 is always R (red), plane #1 is always G (green), plane #2 is always B (blue), |
| * and plane #3 is always A (alpha). This format may represent pre-multiplied or |
| * non-premultiplied alpha.</p> |
| * |
| * <p>All four planes are guaranteed to have the same row strides and pixel strides.</p> |
| * |
| * <p>For example, the {@link android.media.Image} object can provide data |
| * in this format from a {@link android.media.MediaCodec} |
| * through {@link android.media.MediaCodec#getOutputImage} object.</p> |
| * |
| * @see android.media.Image |
| * @see android.media.MediaCodec |
| */ |
| public static final int FLEX_RGBA_8888 = 0x2A; |
| |
| /** |
| * <p>General raw camera sensor image format, usually representing a |
| * single-channel Bayer-mosaic image. Each pixel color sample is stored with |
| * 16 bits of precision.</p> |
| * |
| * <p>The layout of the color mosaic, the maximum and minimum encoding |
| * values of the raw pixel data, the color space of the image, and all other |
| * needed information to interpret a raw sensor image must be queried from |
| * the {@link android.hardware.camera2.CameraDevice} which produced the |
| * image.</p> |
| */ |
| public static final int RAW_SENSOR = 0x20; |
| |
| /** |
| * <p>Private raw camera sensor image format, a single channel image with |
| * implementation dependent pixel layout.</p> |
| * |
| * <p>RAW_PRIVATE is a format for unprocessed raw image buffers coming from an |
| * image sensor. The actual structure of buffers of this format is |
| * implementation-dependent.</p> |
| * |
| */ |
| public static final int RAW_PRIVATE = 0x24; |
| |
| /** |
| * <p> |
| * Android 10-bit raw format |
| * </p> |
| * <p> |
| * This is a single-plane, 10-bit per pixel, densely packed (in each row), |
| * unprocessed format, usually representing raw Bayer-pattern images coming |
| * from an image sensor. |
| * </p> |
| * <p> |
| * In an image buffer with this format, starting from the first pixel of |
| * each row, each 4 consecutive pixels are packed into 5 bytes (40 bits). |
| * Each one of the first 4 bytes contains the top 8 bits of each pixel, The |
| * fifth byte contains the 2 least significant bits of the 4 pixels, the |
| * exact layout data for each 4 consecutive pixels is illustrated below |
| * ({@code Pi[j]} stands for the jth bit of the ith pixel): |
| * </p> |
| * <table> |
| * <thead> |
| * <tr> |
| * <th align="center"></th> |
| * <th align="center">bit 7</th> |
| * <th align="center">bit 6</th> |
| * <th align="center">bit 5</th> |
| * <th align="center">bit 4</th> |
| * <th align="center">bit 3</th> |
| * <th align="center">bit 2</th> |
| * <th align="center">bit 1</th> |
| * <th align="center">bit 0</th> |
| * </tr> |
| * </thead> <tbody> |
| * <tr> |
| * <td align="center">Byte 0:</td> |
| * <td align="center">P0[9]</td> |
| * <td align="center">P0[8]</td> |
| * <td align="center">P0[7]</td> |
| * <td align="center">P0[6]</td> |
| * <td align="center">P0[5]</td> |
| * <td align="center">P0[4]</td> |
| * <td align="center">P0[3]</td> |
| * <td align="center">P0[2]</td> |
| * </tr> |
| * <tr> |
| * <td align="center">Byte 1:</td> |
| * <td align="center">P1[9]</td> |
| * <td align="center">P1[8]</td> |
| * <td align="center">P1[7]</td> |
| * <td align="center">P1[6]</td> |
| * <td align="center">P1[5]</td> |
| * <td align="center">P1[4]</td> |
| * <td align="center">P1[3]</td> |
| * <td align="center">P1[2]</td> |
| * </tr> |
| * <tr> |
| * <td align="center">Byte 2:</td> |
| * <td align="center">P2[9]</td> |
| * <td align="center">P2[8]</td> |
| * <td align="center">P2[7]</td> |
| * <td align="center">P2[6]</td> |
| * <td align="center">P2[5]</td> |
| * <td align="center">P2[4]</td> |
| * <td align="center">P2[3]</td> |
| * <td align="center">P2[2]</td> |
| * </tr> |
| * <tr> |
| * <td align="center">Byte 3:</td> |
| * <td align="center">P3[9]</td> |
| * <td align="center">P3[8]</td> |
| * <td align="center">P3[7]</td> |
| * <td align="center">P3[6]</td> |
| * <td align="center">P3[5]</td> |
| * <td align="center">P3[4]</td> |
| * <td align="center">P3[3]</td> |
| * <td align="center">P3[2]</td> |
| * </tr> |
| * <tr> |
| * <td align="center">Byte 4:</td> |
| * <td align="center">P3[1]</td> |
| * <td align="center">P3[0]</td> |
| * <td align="center">P2[1]</td> |
| * <td align="center">P2[0]</td> |
| * <td align="center">P1[1]</td> |
| * <td align="center">P1[0]</td> |
| * <td align="center">P0[1]</td> |
| * <td align="center">P0[0]</td> |
| * </tr> |
| * </tbody> |
| * </table> |
| * <p> |
| * This format assumes |
| * <ul> |
| * <li>a width multiple of 4 pixels</li> |
| * <li>an even height</li> |
| * </ul> |
| * </p> |
| * |
| * <pre>size = row stride * height</pre> where the row stride is in <em>bytes</em>, |
| * not pixels. |
| * |
| * <p> |
| * Since this is a densely packed format, the pixel stride is always 0. The |
| * application must use the pixel data layout defined in above table to |
| * access each row data. When row stride is equal to {@code width * (10 / 8)}, there |
| * will be no padding bytes at the end of each row, the entire image data is |
| * densely packed. When stride is larger than {@code width * (10 / 8)}, padding |
| * bytes will be present at the end of each row. |
| * </p> |
| * <p> |
| * For example, the {@link android.media.Image} object can provide data in |
| * this format from a {@link android.hardware.camera2.CameraDevice} (if |
| * supported) through a {@link android.media.ImageReader} object. The |
| * {@link android.media.Image#getPlanes() Image#getPlanes()} will return a |
| * single plane containing the pixel data. The pixel stride is always 0 in |
| * {@link android.media.Image.Plane#getPixelStride()}, and the |
| * {@link android.media.Image.Plane#getRowStride()} describes the vertical |
| * neighboring pixel distance (in bytes) between adjacent rows. |
| * </p> |
| * |
| * @see android.media.Image |
| * @see android.media.ImageReader |
| * @see android.hardware.camera2.CameraDevice |
| */ |
| public static final int RAW10 = 0x25; |
| |
| /** |
| * <p> |
| * Android 12-bit raw format |
| * </p> |
| * <p> |
| * This is a single-plane, 12-bit per pixel, densely packed (in each row), |
| * unprocessed format, usually representing raw Bayer-pattern images coming |
| * from an image sensor. |
| * </p> |
| * <p> |
| * In an image buffer with this format, starting from the first pixel of each |
| * row, each two consecutive pixels are packed into 3 bytes (24 bits). The first |
| * and second byte contains the top 8 bits of first and second pixel. The third |
| * byte contains the 4 least significant bits of the two pixels, the exact layout |
| * data for each two consecutive pixels is illustrated below (Pi[j] stands for |
| * the jth bit of the ith pixel): |
| * </p> |
| * <table> |
| * <thead> |
| * <tr> |
| * <th align="center"></th> |
| * <th align="center">bit 7</th> |
| * <th align="center">bit 6</th> |
| * <th align="center">bit 5</th> |
| * <th align="center">bit 4</th> |
| * <th align="center">bit 3</th> |
| * <th align="center">bit 2</th> |
| * <th align="center">bit 1</th> |
| * <th align="center">bit 0</th> |
| * </tr> |
| * </thead> <tbody> |
| * <tr> |
| * <td align="center">Byte 0:</td> |
| * <td align="center">P0[11]</td> |
| * <td align="center">P0[10]</td> |
| * <td align="center">P0[ 9]</td> |
| * <td align="center">P0[ 8]</td> |
| * <td align="center">P0[ 7]</td> |
| * <td align="center">P0[ 6]</td> |
| * <td align="center">P0[ 5]</td> |
| * <td align="center">P0[ 4]</td> |
| * </tr> |
| * <tr> |
| * <td align="center">Byte 1:</td> |
| * <td align="center">P1[11]</td> |
| * <td align="center">P1[10]</td> |
| * <td align="center">P1[ 9]</td> |
| * <td align="center">P1[ 8]</td> |
| * <td align="center">P1[ 7]</td> |
| * <td align="center">P1[ 6]</td> |
| * <td align="center">P1[ 5]</td> |
| * <td align="center">P1[ 4]</td> |
| * </tr> |
| * <tr> |
| * <td align="center">Byte 2:</td> |
| * <td align="center">P1[ 3]</td> |
| * <td align="center">P1[ 2]</td> |
| * <td align="center">P1[ 1]</td> |
| * <td align="center">P1[ 0]</td> |
| * <td align="center">P0[ 3]</td> |
| * <td align="center">P0[ 2]</td> |
| * <td align="center">P0[ 1]</td> |
| * <td align="center">P0[ 0]</td> |
| * </tr> |
| * </tbody> |
| * </table> |
| * <p> |
| * This format assumes |
| * <ul> |
| * <li>a width multiple of 4 pixels</li> |
| * <li>an even height</li> |
| * </ul> |
| * </p> |
| * |
| * <pre>size = row stride * height</pre> where the row stride is in <em>bytes</em>, |
| * not pixels. |
| * |
| * <p> |
| * Since this is a densely packed format, the pixel stride is always 0. The |
| * application must use the pixel data layout defined in above table to |
| * access each row data. When row stride is equal to {@code width * (12 / 8)}, there |
| * will be no padding bytes at the end of each row, the entire image data is |
| * densely packed. When stride is larger than {@code width * (12 / 8)}, padding |
| * bytes will be present at the end of each row. |
| * </p> |
| * <p> |
| * For example, the {@link android.media.Image} object can provide data in |
| * this format from a {@link android.hardware.camera2.CameraDevice} (if |
| * supported) through a {@link android.media.ImageReader} object. The |
| * {@link android.media.Image#getPlanes() Image#getPlanes()} will return a |
| * single plane containing the pixel data. The pixel stride is always 0 in |
| * {@link android.media.Image.Plane#getPixelStride()}, and the |
| * {@link android.media.Image.Plane#getRowStride()} describes the vertical |
| * neighboring pixel distance (in bytes) between adjacent rows. |
| * </p> |
| * |
| * @see android.media.Image |
| * @see android.media.ImageReader |
| * @see android.hardware.camera2.CameraDevice |
| */ |
| public static final int RAW12 = 0x26; |
| |
| /** |
| * <p>Android dense depth image format.</p> |
| * |
| * <p>Each pixel is 16 bits, representing a depth ranging measurement from a depth camera or |
| * similar sensor. The 16-bit sample consists of a confidence value and the actual ranging |
| * measurement.</p> |
| * |
| * <p>The confidence value is an estimate of correctness for this sample. It is encoded in the |
| * 3 most significant bits of the sample, with a value of 0 representing 100% confidence, a |
| * value of 1 representing 0% confidence, a value of 2 representing 1/7, a value of 3 |
| * representing 2/7, and so on.</p> |
| * |
| * <p>As an example, the following sample extracts the range and confidence from the first pixel |
| * of a DEPTH16-format {@link android.media.Image}, and converts the confidence to a |
| * floating-point value between 0 and 1.f inclusive, with 1.f representing maximum confidence: |
| * |
| * <pre> |
| * ShortBuffer shortDepthBuffer = img.getPlanes()[0].getBuffer().asShortBuffer(); |
| * short depthSample = shortDepthBuffer.get() |
| * short depthRange = (short) (depthSample & 0x1FFF); |
| * short depthConfidence = (short) ((depthSample >> 13) & 0x7); |
| * float depthPercentage = depthConfidence == 0 ? 1.f : (depthConfidence - 1) / 7.f; |
| * </pre> |
| * </p> |
| * |
| * <p>This format assumes |
| * <ul> |
| * <li>an even width</li> |
| * <li>an even height</li> |
| * <li>a horizontal stride multiple of 16 pixels</li> |
| * </ul> |
| * </p> |
| * |
| * <pre> y_size = stride * height </pre> |
| * |
| * When produced by a camera, the units for the range are millimeters. |
| */ |
| public static final int DEPTH16 = 0x44363159; |
| |
| /** |
| * Android sparse depth point cloud format. |
| * |
| * <p>A variable-length list of 3D points plus a confidence value, with each point represented |
| * by four floats; first the X, Y, Z position coordinates, and then the confidence value.</p> |
| * |
| * <p>The number of points is {@code (size of the buffer in bytes) / 16}. |
| * |
| * <p>The coordinate system and units of the position values depend on the source of the point |
| * cloud data. The confidence value is between 0.f and 1.f, inclusive, with 0 representing 0% |
| * confidence and 1.f representing 100% confidence in the measured position values.</p> |
| * |
| * <p>As an example, the following code extracts the first depth point in a DEPTH_POINT_CLOUD |
| * format {@link android.media.Image}: |
| * <pre> |
| * FloatBuffer floatDepthBuffer = img.getPlanes()[0].getBuffer().asFloatBuffer(); |
| * float x = floatDepthBuffer.get(); |
| * float y = floatDepthBuffer.get(); |
| * float z = floatDepthBuffer.get(); |
| * float confidence = floatDepthBuffer.get(); |
| * </pre> |
| * |
| * For camera devices that support the |
| * {@link android.hardware.camera2.CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES_DEPTH_OUTPUT DEPTH_OUTPUT} |
| * capability, DEPTH_POINT_CLOUD coordinates have units of meters, and the coordinate system is |
| * defined by the camera's pose transforms: |
| * {@link android.hardware.camera2.CameraCharacteristics#LENS_POSE_TRANSLATION} and |
| * {@link android.hardware.camera2.CameraCharacteristics#LENS_POSE_ROTATION}. That means the origin is |
| * the optical center of the camera device, and the positive Z axis points along the camera's optical axis, |
| * toward the scene. |
| */ |
| public static final int DEPTH_POINT_CLOUD = 0x101; |
| |
| /** |
| * Unprocessed implementation-dependent raw |
| * depth measurements, opaque with 16 bit |
| * samples. |
| * |
| * @hide |
| */ |
| public static final int RAW_DEPTH = 0x1002; |
| |
| /** |
| * Unprocessed implementation-dependent raw |
| * depth measurements, opaque with 10 bit |
| * samples and device specific bit layout. |
| * |
| * @hide |
| */ |
| public static final int RAW_DEPTH10 = 0x1003; |
| |
| /** |
| * Android private opaque image format. |
| * <p> |
| * The choices of the actual format and pixel data layout are entirely up to |
| * the device-specific and framework internal implementations, and may vary |
| * depending on use cases even for the same device. The buffers of this |
| * format can be produced by components like |
| * {@link android.media.ImageWriter ImageWriter} , and interpreted correctly |
| * by consumers like {@link android.hardware.camera2.CameraDevice |
| * CameraDevice} based on the device/framework private information. However, |
| * these buffers are not directly accessible to the application. |
| * </p> |
| * <p> |
| * When an {@link android.media.Image Image} of this format is obtained from |
| * an {@link android.media.ImageReader ImageReader} or |
| * {@link android.media.ImageWriter ImageWriter}, the |
| * {@link android.media.Image#getPlanes() getPlanes()} method will return an |
| * empty {@link android.media.Image.Plane Plane} array. |
| * </p> |
| * <p> |
| * If a buffer of this format is to be used as an OpenGL ES texture, the |
| * framework will assume that sampling the texture will always return an |
| * alpha value of 1.0 (i.e. the buffer contains only opaque pixel values). |
| * </p> |
| */ |
| public static final int PRIVATE = 0x22; |
| |
| /** |
| * Compressed HEIC format. |
| * |
| * <p>This format defines the HEIC brand of High Efficiency Image File |
| * Format as described in ISO/IEC 23008-12.</p> |
| */ |
| public static final int HEIC = 0x48454946; |
| |
| /** |
| * Use this function to retrieve the number of bits per pixel of an |
| * ImageFormat. |
| * |
| * @param format |
| * @return the number of bits per pixel of the given format or -1 if the |
| * format doesn't exist or is not supported. |
| */ |
| public static int getBitsPerPixel(@Format int format) { |
| switch (format) { |
| case RGB_565: |
| return 16; |
| case NV16: |
| return 16; |
| case YUY2: |
| return 16; |
| case YV12: |
| return 12; |
| case Y8: |
| return 8; |
| case Y16: |
| case DEPTH16: |
| return 16; |
| case NV21: |
| return 12; |
| case YUV_420_888: |
| return 12; |
| case YUV_422_888: |
| return 16; |
| case YUV_444_888: |
| return 24; |
| case FLEX_RGB_888: |
| return 24; |
| case FLEX_RGBA_8888: |
| return 32; |
| case RAW_DEPTH: |
| case RAW_SENSOR: |
| return 16; |
| case YCBCR_P010: |
| return 24; |
| case YCBCR_P210: |
| return 32; |
| case RAW_DEPTH10: |
| case RAW10: |
| return 10; |
| case RAW12: |
| return 12; |
| } |
| return -1; |
| } |
| |
| /** |
| * Determine whether or not this is a public-visible {@code format}. |
| * |
| * <p>In particular, {@code @hide} formats will return {@code false}.</p> |
| * |
| * <p>Any other formats (including UNKNOWN) will return {@code false}.</p> |
| * |
| * @param format an integer format |
| * @return a boolean |
| * |
| * @hide |
| */ |
| public static boolean isPublicFormat(@Format int format) { |
| switch (format) { |
| case RGB_565: |
| case NV16: |
| case YUY2: |
| case YV12: |
| case JPEG: |
| case NV21: |
| case YUV_420_888: |
| case YUV_422_888: |
| case YUV_444_888: |
| case YCBCR_P010: |
| case FLEX_RGB_888: |
| case FLEX_RGBA_8888: |
| case RAW_SENSOR: |
| case RAW_PRIVATE: |
| case RAW10: |
| case RAW12: |
| case DEPTH16: |
| case DEPTH_POINT_CLOUD: |
| case PRIVATE: |
| case RAW_DEPTH: |
| case RAW_DEPTH10: |
| case Y8: |
| case DEPTH_JPEG: |
| case HEIC: |
| case JPEG_R: |
| return true; |
| } |
| if (android.media.codec.Flags.p210FormatSupport() && format == YCBCR_P210) { |
| return true; |
| } |
| return false; |
| } |
| } |