906 lines
34 KiB
Java
906 lines
34 KiB
Java
![]() |
/*
|
||
|
* Copyright (C) 2010 The Android Open Source Project
|
||
|
*
|
||
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||
|
* you may not use this file except in compliance with the License.
|
||
|
* You may obtain a copy of the License at
|
||
|
*
|
||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||
|
*
|
||
|
* Unless required by applicable law or agreed to in writing, software
|
||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||
|
* See the License for the specific language governing permissions and
|
||
|
* limitations under the License.
|
||
|
*/
|
||
|
|
||
|
package android.graphics;
|
||
|
|
||
|
import android.annotation.IntDef;
|
||
|
|
||
|
import java.lang.annotation.Retention;
|
||
|
import java.lang.annotation.RetentionPolicy;
|
||
|
|
||
|
public class ImageFormat {
|
||
|
/** @hide */
|
||
|
@Retention(RetentionPolicy.SOURCE)
|
||
|
@IntDef(value = {
|
||
|
UNKNOWN,
|
||
|
/*
|
||
|
* Since some APIs accept either ImageFormat or PixelFormat (and the two
|
||
|
* enums do not overlap since they're both partial versions of the
|
||
|
* internal format enum), add PixelFormat values here so linting
|
||
|
* tools won't complain when method arguments annotated with
|
||
|
* ImageFormat are provided with PixelFormat values.
|
||
|
*/
|
||
|
PixelFormat.RGBA_8888,
|
||
|
PixelFormat.RGBX_8888,
|
||
|
PixelFormat.RGB_888,
|
||
|
RGB_565,
|
||
|
YV12,
|
||
|
Y8,
|
||
|
Y16,
|
||
|
YCBCR_P010,
|
||
|
NV16,
|
||
|
NV21,
|
||
|
YUY2,
|
||
|
JPEG,
|
||
|
DEPTH_JPEG,
|
||
|
YUV_420_888,
|
||
|
YUV_422_888,
|
||
|
YUV_444_888,
|
||
|
FLEX_RGB_888,
|
||
|
FLEX_RGBA_8888,
|
||
|
RAW_SENSOR,
|
||
|
RAW_PRIVATE,
|
||
|
RAW10,
|
||
|
RAW12,
|
||
|
DEPTH16,
|
||
|
DEPTH_POINT_CLOUD,
|
||
|
RAW_DEPTH,
|
||
|
RAW_DEPTH10,
|
||
|
PRIVATE,
|
||
|
HEIC,
|
||
|
JPEG_R
|
||
|
})
|
||
|
public @interface Format {
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* these constants are chosen to be binary compatible with their previous
|
||
|
* location in PixelFormat.java
|
||
|
*/
|
||
|
|
||
|
public static final int UNKNOWN = 0;
|
||
|
|
||
|
/**
|
||
|
* RGB format used for pictures encoded as RGB_565. See
|
||
|
* {@link android.hardware.Camera.Parameters#setPictureFormat(int)}.
|
||
|
*/
|
||
|
public static final int RGB_565 = 4;
|
||
|
|
||
|
/**
|
||
|
* <p>Android YUV format.</p>
|
||
|
*
|
||
|
* <p>This format is exposed to software decoders and applications.</p>
|
||
|
*
|
||
|
* <p>YV12 is a 4:2:0 YCrCb planar format comprised of a WxH Y plane followed
|
||
|
* by (W/2) x (H/2) Cr and Cb planes.</p>
|
||
|
*
|
||
|
* <p>This format assumes
|
||
|
* <ul>
|
||
|
* <li>an even width</li>
|
||
|
* <li>an even height</li>
|
||
|
* <li>a horizontal stride multiple of 16 pixels</li>
|
||
|
* <li>a vertical stride equal to the height</li>
|
||
|
* </ul>
|
||
|
* </p>
|
||
|
*
|
||
|
* <pre> y_size = stride * height
|
||
|
* c_stride = ALIGN(stride/2, 16)
|
||
|
* c_size = c_stride * height/2
|
||
|
* size = y_size + c_size * 2
|
||
|
* cr_offset = y_size
|
||
|
* cb_offset = y_size + c_size</pre>
|
||
|
*
|
||
|
* <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is
|
||
|
* recommended for YUV output instead.</p>
|
||
|
*
|
||
|
* <p>For the older camera API, this format is guaranteed to be supported for
|
||
|
* {@link android.hardware.Camera} preview images since API level 12; for earlier API versions,
|
||
|
* check {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}.
|
||
|
*
|
||
|
* <p>Note that for camera preview callback use (see
|
||
|
* {@link android.hardware.Camera#setPreviewCallback}), the
|
||
|
* <var>stride</var> value is the smallest possible; that is, it is equal
|
||
|
* to:
|
||
|
*
|
||
|
* <pre>stride = ALIGN(width, 16)</pre>
|
||
|
*
|
||
|
* @see android.hardware.Camera.Parameters#setPreviewCallback
|
||
|
* @see android.hardware.Camera.Parameters#setPreviewFormat
|
||
|
* @see android.hardware.Camera.Parameters#getSupportedPreviewFormats
|
||
|
* </p>
|
||
|
*/
|
||
|
public static final int YV12 = 0x32315659;
|
||
|
|
||
|
/**
|
||
|
* <p>Android Y8 format.</p>
|
||
|
*
|
||
|
* <p>Y8 is a YUV planar format comprised of a WxH Y plane only, with each pixel
|
||
|
* being represented by 8 bits. It is equivalent to just the Y plane from {@link #YV12}
|
||
|
* format.</p>
|
||
|
*
|
||
|
* <p>This format assumes
|
||
|
* <ul>
|
||
|
* <li>an even width</li>
|
||
|
* <li>an even height</li>
|
||
|
* <li>a horizontal stride multiple of 16 pixels</li>
|
||
|
* </ul>
|
||
|
* </p>
|
||
|
*
|
||
|
* <pre> size = stride * height </pre>
|
||
|
*
|
||
|
* <p>For example, the {@link android.media.Image} object can provide data
|
||
|
* in this format from a {@link android.hardware.camera2.CameraDevice} (if
|
||
|
* supported) through a {@link android.media.ImageReader} object. The
|
||
|
* {@link android.media.Image#getPlanes() Image#getPlanes()} will return a
|
||
|
* single plane containing the pixel data. The pixel stride is always 1 in
|
||
|
* {@link android.media.Image.Plane#getPixelStride()}, and the
|
||
|
* {@link android.media.Image.Plane#getRowStride()} describes the vertical
|
||
|
* neighboring pixel distance (in bytes) between adjacent rows.</p>
|
||
|
*
|
||
|
* @see android.media.Image
|
||
|
* @see android.media.ImageReader
|
||
|
* @see android.hardware.camera2.CameraDevice
|
||
|
*/
|
||
|
public static final int Y8 = 0x20203859;
|
||
|
|
||
|
/**
|
||
|
* <p>Android Y16 format.</p>
|
||
|
*
|
||
|
* Y16 is a YUV planar format comprised of a WxH Y plane, with each pixel
|
||
|
* being represented by 16 bits. It is just like {@link #Y8}, but has 16
|
||
|
* bits per pixel (little endian).</p>
|
||
|
*
|
||
|
* <p>This format assumes
|
||
|
* <ul>
|
||
|
* <li>an even width</li>
|
||
|
* <li>an even height</li>
|
||
|
* <li>a horizontal stride multiple of 16 pixels</li>
|
||
|
* </ul>
|
||
|
* </p>
|
||
|
*
|
||
|
* <pre> y_size = stride * height </pre>
|
||
|
*
|
||
|
* <p>For example, the {@link android.media.Image} object can provide data
|
||
|
* in this format from a {@link android.hardware.camera2.CameraDevice}
|
||
|
* through a {@link android.media.ImageReader} object if this format is
|
||
|
* supported by {@link android.hardware.camera2.CameraDevice}.</p>
|
||
|
*
|
||
|
* @see android.media.Image
|
||
|
* @see android.media.ImageReader
|
||
|
* @see android.hardware.camera2.CameraDevice
|
||
|
*
|
||
|
* @hide
|
||
|
*/
|
||
|
public static final int Y16 = 0x20363159;
|
||
|
|
||
|
/**
|
||
|
* <p>Android YUV P010 format.</p>
|
||
|
*
|
||
|
* P010 is a 4:2:0 YCbCr semiplanar format comprised of a WxH Y plane
|
||
|
* followed by a Wx(H/2) CbCr plane. Each sample is represented by a 16-bit
|
||
|
* little-endian value, with the lower 6 bits set to zero.
|
||
|
*
|
||
|
* <p>For example, the {@link android.media.Image} object can provide data
|
||
|
* in this format from a {@link android.hardware.camera2.CameraDevice}
|
||
|
* through a {@link android.media.ImageReader} object if this format is
|
||
|
* supported by {@link android.hardware.camera2.CameraDevice}.</p>
|
||
|
*
|
||
|
* @see android.media.Image
|
||
|
* @see android.media.ImageReader
|
||
|
* @see android.hardware.camera2.CameraDevice
|
||
|
*
|
||
|
*/
|
||
|
public static final int YCBCR_P010 = 0x36;
|
||
|
|
||
|
/**
|
||
|
* YCbCr format, used for video.
|
||
|
*
|
||
|
* <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is
|
||
|
* recommended for YUV output instead.</p>
|
||
|
*
|
||
|
* <p>Whether this format is supported by the old camera API can be determined by
|
||
|
* {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}.</p>
|
||
|
*
|
||
|
*/
|
||
|
public static final int NV16 = 0x10;
|
||
|
|
||
|
/**
|
||
|
* YCrCb format used for images, which uses the NV21 encoding format.
|
||
|
*
|
||
|
* <p>This is the default format
|
||
|
* for {@link android.hardware.Camera} preview images, when not otherwise set with
|
||
|
* {@link android.hardware.Camera.Parameters#setPreviewFormat(int)}.</p>
|
||
|
*
|
||
|
* <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is
|
||
|
* recommended for YUV output instead.</p>
|
||
|
*/
|
||
|
public static final int NV21 = 0x11;
|
||
|
|
||
|
/**
|
||
|
* YCbCr format used for images, which uses YUYV (YUY2) encoding format.
|
||
|
*
|
||
|
* <p>For the {@link android.hardware.camera2} API, the {@link #YUV_420_888} format is
|
||
|
* recommended for YUV output instead.</p>
|
||
|
*
|
||
|
* <p>This is an alternative format for {@link android.hardware.Camera} preview images. Whether
|
||
|
* this format is supported by the camera hardware can be determined by
|
||
|
* {@link android.hardware.Camera.Parameters#getSupportedPreviewFormats()}.</p>
|
||
|
*/
|
||
|
public static final int YUY2 = 0x14;
|
||
|
|
||
|
/**
|
||
|
* Compressed JPEG format.
|
||
|
*
|
||
|
* <p>This format is always supported as an output format for the
|
||
|
* {@link android.hardware.camera2} API, and as a picture format for the older
|
||
|
* {@link android.hardware.Camera} API</p>
|
||
|
*/
|
||
|
public static final int JPEG = 0x100;
|
||
|
|
||
|
/**
|
||
|
* Depth augmented compressed JPEG format.
|
||
|
*
|
||
|
* <p>JPEG compressed main image along with XMP embedded depth metadata
|
||
|
* following ISO 16684-1:2011(E).</p>
|
||
|
*/
|
||
|
public static final int DEPTH_JPEG = 0x69656963;
|
||
|
|
||
|
/**
|
||
|
* Compressed JPEG format that includes an embedded recovery map.
|
||
|
*
|
||
|
* <p>JPEG compressed main image along with embedded recovery map following the
|
||
|
* <a href="https://developer.android.com/guide/topics/media/hdr-image-format">Ultra HDR
|
||
|
* Image format specification</a>.</p>
|
||
|
*/
|
||
|
public static final int JPEG_R = 0x1005;
|
||
|
|
||
|
/**
|
||
|
* <p>Multi-plane Android YUV 420 format</p>
|
||
|
*
|
||
|
* <p>This format is a generic YCbCr format, capable of describing any 4:2:0
|
||
|
* chroma-subsampled planar or semiplanar buffer (but not fully interleaved),
|
||
|
* with 8 bits per color sample.</p>
|
||
|
*
|
||
|
* <p>Images in this format are always represented by three separate buffers
|
||
|
* of data, one for each color plane. Additional information always
|
||
|
* accompanies the buffers, describing the row stride and the pixel stride
|
||
|
* for each plane.</p>
|
||
|
*
|
||
|
* <p>The order of planes in the array returned by
|
||
|
* {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
|
||
|
* plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).</p>
|
||
|
*
|
||
|
* <p>The Y-plane is guaranteed not to be interleaved with the U/V planes
|
||
|
* (in particular, pixel stride is always 1 in
|
||
|
* {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}).</p>
|
||
|
*
|
||
|
* <p>The U/V planes are guaranteed to have the same row stride and pixel stride
|
||
|
* (in particular,
|
||
|
* {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()}
|
||
|
* == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and
|
||
|
* {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()}
|
||
|
* == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()};
|
||
|
* ).</p>
|
||
|
*
|
||
|
* <p>For example, the {@link android.media.Image} object can provide data
|
||
|
* in this format from a {@link android.hardware.camera2.CameraDevice}
|
||
|
* through a {@link android.media.ImageReader} object.</p>
|
||
|
*
|
||
|
* @see android.media.Image
|
||
|
* @see android.media.ImageReader
|
||
|
* @see android.hardware.camera2.CameraDevice
|
||
|
*/
|
||
|
public static final int YUV_420_888 = 0x23;
|
||
|
|
||
|
/**
|
||
|
* <p>Multi-plane Android YUV 422 format</p>
|
||
|
*
|
||
|
* <p>This format is a generic YCbCr format, capable of describing any 4:2:2
|
||
|
* chroma-subsampled (planar, semiplanar or interleaved) format,
|
||
|
* with 8 bits per color sample.</p>
|
||
|
*
|
||
|
* <p>Images in this format are always represented by three separate buffers
|
||
|
* of data, one for each color plane. Additional information always
|
||
|
* accompanies the buffers, describing the row stride and the pixel stride
|
||
|
* for each plane.</p>
|
||
|
*
|
||
|
* <p>The order of planes in the array returned by
|
||
|
* {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
|
||
|
* plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).</p>
|
||
|
*
|
||
|
* <p>In contrast to the {@link #YUV_420_888} format, the Y-plane may have a pixel
|
||
|
* stride greater than 1 in
|
||
|
* {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}.</p>
|
||
|
*
|
||
|
* <p>The U/V planes are guaranteed to have the same row stride and pixel stride
|
||
|
* (in particular,
|
||
|
* {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()}
|
||
|
* == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and
|
||
|
* {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()}
|
||
|
* == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()};
|
||
|
* ).</p>
|
||
|
*
|
||
|
* <p>For example, the {@link android.media.Image} object can provide data
|
||
|
* in this format from a {@link android.media.MediaCodec}
|
||
|
* through {@link android.media.MediaCodec#getOutputImage} object.</p>
|
||
|
*
|
||
|
* @see android.media.Image
|
||
|
* @see android.media.MediaCodec
|
||
|
*/
|
||
|
public static final int YUV_422_888 = 0x27;
|
||
|
|
||
|
/**
|
||
|
* <p>Multi-plane Android YUV 444 format</p>
|
||
|
*
|
||
|
* <p>This format is a generic YCbCr format, capable of describing any 4:4:4
|
||
|
* (planar, semiplanar or interleaved) format,
|
||
|
* with 8 bits per color sample.</p>
|
||
|
*
|
||
|
* <p>Images in this format are always represented by three separate buffers
|
||
|
* of data, one for each color plane. Additional information always
|
||
|
* accompanies the buffers, describing the row stride and the pixel stride
|
||
|
* for each plane.</p>
|
||
|
*
|
||
|
* <p>The order of planes in the array returned by
|
||
|
* {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
|
||
|
* plane #0 is always Y, plane #1 is always U (Cb), and plane #2 is always V (Cr).</p>
|
||
|
*
|
||
|
* <p>In contrast to the {@link #YUV_420_888} format, the Y-plane may have a pixel
|
||
|
* stride greater than 1 in
|
||
|
* {@link android.media.Image.Plane#getPixelStride() yPlane.getPixelStride()}.</p>
|
||
|
*
|
||
|
* <p>The U/V planes are guaranteed to have the same row stride and pixel stride
|
||
|
* (in particular,
|
||
|
* {@link android.media.Image.Plane#getRowStride() uPlane.getRowStride()}
|
||
|
* == {@link android.media.Image.Plane#getRowStride() vPlane.getRowStride()} and
|
||
|
* {@link android.media.Image.Plane#getPixelStride() uPlane.getPixelStride()}
|
||
|
* == {@link android.media.Image.Plane#getPixelStride() vPlane.getPixelStride()};
|
||
|
* ).</p>
|
||
|
*
|
||
|
* <p>For example, the {@link android.media.Image} object can provide data
|
||
|
* in this format from a {@link android.media.MediaCodec}
|
||
|
* through {@link android.media.MediaCodec#getOutputImage} object.</p>
|
||
|
*
|
||
|
* @see android.media.Image
|
||
|
* @see android.media.MediaCodec
|
||
|
*/
|
||
|
public static final int YUV_444_888 = 0x28;
|
||
|
|
||
|
/**
|
||
|
* <p>Multi-plane Android RGB format</p>
|
||
|
*
|
||
|
* <p>This format is a generic RGB format, capable of describing most RGB formats,
|
||
|
* with 8 bits per color sample.</p>
|
||
|
*
|
||
|
* <p>Images in this format are always represented by three separate buffers
|
||
|
* of data, one for each color plane. Additional information always
|
||
|
* accompanies the buffers, describing the row stride and the pixel stride
|
||
|
* for each plane.</p>
|
||
|
*
|
||
|
* <p>The order of planes in the array returned by
|
||
|
* {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
|
||
|
* plane #0 is always R (red), plane #1 is always G (green), and plane #2 is always B
|
||
|
* (blue).</p>
|
||
|
*
|
||
|
* <p>All three planes are guaranteed to have the same row strides and pixel strides.</p>
|
||
|
*
|
||
|
* <p>For example, the {@link android.media.Image} object can provide data
|
||
|
* in this format from a {@link android.media.MediaCodec}
|
||
|
* through {@link android.media.MediaCodec#getOutputImage} object.</p>
|
||
|
*
|
||
|
* @see android.media.Image
|
||
|
* @see android.media.MediaCodec
|
||
|
*/
|
||
|
public static final int FLEX_RGB_888 = 0x29;
|
||
|
|
||
|
/**
|
||
|
* <p>Multi-plane Android RGBA format</p>
|
||
|
*
|
||
|
* <p>This format is a generic RGBA format, capable of describing most RGBA formats,
|
||
|
* with 8 bits per color sample.</p>
|
||
|
*
|
||
|
* <p>Images in this format are always represented by four separate buffers
|
||
|
* of data, one for each color plane. Additional information always
|
||
|
* accompanies the buffers, describing the row stride and the pixel stride
|
||
|
* for each plane.</p>
|
||
|
*
|
||
|
* <p>The order of planes in the array returned by
|
||
|
* {@link android.media.Image#getPlanes() Image#getPlanes()} is guaranteed such that
|
||
|
* plane #0 is always R (red), plane #1 is always G (green), plane #2 is always B (blue),
|
||
|
* and plane #3 is always A (alpha). This format may represent pre-multiplied or
|
||
|
* non-premultiplied alpha.</p>
|
||
|
*
|
||
|
* <p>All four planes are guaranteed to have the same row strides and pixel strides.</p>
|
||
|
*
|
||
|
* <p>For example, the {@link android.media.Image} object can provide data
|
||
|
* in this format from a {@link android.media.MediaCodec}
|
||
|
* through {@link android.media.MediaCodec#getOutputImage} object.</p>
|
||
|
*
|
||
|
* @see android.media.Image
|
||
|
* @see android.media.MediaCodec
|
||
|
*/
|
||
|
public static final int FLEX_RGBA_8888 = 0x2A;
|
||
|
|
||
|
/**
|
||
|
* <p>General raw camera sensor image format, usually representing a
|
||
|
* single-channel Bayer-mosaic image. Each pixel color sample is stored with
|
||
|
* 16 bits of precision.</p>
|
||
|
*
|
||
|
* <p>The layout of the color mosaic, the maximum and minimum encoding
|
||
|
* values of the raw pixel data, the color space of the image, and all other
|
||
|
* needed information to interpret a raw sensor image must be queried from
|
||
|
* the {@link android.hardware.camera2.CameraDevice} which produced the
|
||
|
* image.</p>
|
||
|
*/
|
||
|
public static final int RAW_SENSOR = 0x20;
|
||
|
|
||
|
/**
|
||
|
* <p>Private raw camera sensor image format, a single channel image with
|
||
|
* implementation dependent pixel layout.</p>
|
||
|
*
|
||
|
* <p>RAW_PRIVATE is a format for unprocessed raw image buffers coming from an
|
||
|
* image sensor. The actual structure of buffers of this format is
|
||
|
* implementation-dependent.</p>
|
||
|
*
|
||
|
*/
|
||
|
public static final int RAW_PRIVATE = 0x24;
|
||
|
|
||
|
/**
|
||
|
* <p>
|
||
|
* Android 10-bit raw format
|
||
|
* </p>
|
||
|
* <p>
|
||
|
* This is a single-plane, 10-bit per pixel, densely packed (in each row),
|
||
|
* unprocessed format, usually representing raw Bayer-pattern images coming
|
||
|
* from an image sensor.
|
||
|
* </p>
|
||
|
* <p>
|
||
|
* In an image buffer with this format, starting from the first pixel of
|
||
|
* each row, each 4 consecutive pixels are packed into 5 bytes (40 bits).
|
||
|
* Each one of the first 4 bytes contains the top 8 bits of each pixel, The
|
||
|
* fifth byte contains the 2 least significant bits of the 4 pixels, the
|
||
|
* exact layout data for each 4 consecutive pixels is illustrated below
|
||
|
* ({@code Pi[j]} stands for the jth bit of the ith pixel):
|
||
|
* </p>
|
||
|
* <table>
|
||
|
* <thead>
|
||
|
* <tr>
|
||
|
* <th align="center"></th>
|
||
|
* <th align="center">bit 7</th>
|
||
|
* <th align="center">bit 6</th>
|
||
|
* <th align="center">bit 5</th>
|
||
|
* <th align="center">bit 4</th>
|
||
|
* <th align="center">bit 3</th>
|
||
|
* <th align="center">bit 2</th>
|
||
|
* <th align="center">bit 1</th>
|
||
|
* <th align="center">bit 0</th>
|
||
|
* </tr>
|
||
|
* </thead> <tbody>
|
||
|
* <tr>
|
||
|
* <td align="center">Byte 0:</td>
|
||
|
* <td align="center">P0[9]</td>
|
||
|
* <td align="center">P0[8]</td>
|
||
|
* <td align="center">P0[7]</td>
|
||
|
* <td align="center">P0[6]</td>
|
||
|
* <td align="center">P0[5]</td>
|
||
|
* <td align="center">P0[4]</td>
|
||
|
* <td align="center">P0[3]</td>
|
||
|
* <td align="center">P0[2]</td>
|
||
|
* </tr>
|
||
|
* <tr>
|
||
|
* <td align="center">Byte 1:</td>
|
||
|
* <td align="center">P1[9]</td>
|
||
|
* <td align="center">P1[8]</td>
|
||
|
* <td align="center">P1[7]</td>
|
||
|
* <td align="center">P1[6]</td>
|
||
|
* <td align="center">P1[5]</td>
|
||
|
* <td align="center">P1[4]</td>
|
||
|
* <td align="center">P1[3]</td>
|
||
|
* <td align="center">P1[2]</td>
|
||
|
* </tr>
|
||
|
* <tr>
|
||
|
* <td align="center">Byte 2:</td>
|
||
|
* <td align="center">P2[9]</td>
|
||
|
* <td align="center">P2[8]</td>
|
||
|
* <td align="center">P2[7]</td>
|
||
|
* <td align="center">P2[6]</td>
|
||
|
* <td align="center">P2[5]</td>
|
||
|
* <td align="center">P2[4]</td>
|
||
|
* <td align="center">P2[3]</td>
|
||
|
* <td align="center">P2[2]</td>
|
||
|
* </tr>
|
||
|
* <tr>
|
||
|
* <td align="center">Byte 3:</td>
|
||
|
* <td align="center">P3[9]</td>
|
||
|
* <td align="center">P3[8]</td>
|
||
|
* <td align="center">P3[7]</td>
|
||
|
* <td align="center">P3[6]</td>
|
||
|
* <td align="center">P3[5]</td>
|
||
|
* <td align="center">P3[4]</td>
|
||
|
* <td align="center">P3[3]</td>
|
||
|
* <td align="center">P3[2]</td>
|
||
|
* </tr>
|
||
|
* <tr>
|
||
|
* <td align="center">Byte 4:</td>
|
||
|
* <td align="center">P3[1]</td>
|
||
|
* <td align="center">P3[0]</td>
|
||
|
* <td align="center">P2[1]</td>
|
||
|
* <td align="center">P2[0]</td>
|
||
|
* <td align="center">P1[1]</td>
|
||
|
* <td align="center">P1[0]</td>
|
||
|
* <td align="center">P0[1]</td>
|
||
|
* <td align="center">P0[0]</td>
|
||
|
* </tr>
|
||
|
* </tbody>
|
||
|
* </table>
|
||
|
* <p>
|
||
|
* This format assumes
|
||
|
* <ul>
|
||
|
* <li>a width multiple of 4 pixels</li>
|
||
|
* <li>an even height</li>
|
||
|
* </ul>
|
||
|
* </p>
|
||
|
*
|
||
|
* <pre>size = row stride * height</pre> where the row stride is in <em>bytes</em>,
|
||
|
* not pixels.
|
||
|
*
|
||
|
* <p>
|
||
|
* Since this is a densely packed format, the pixel stride is always 0. The
|
||
|
* application must use the pixel data layout defined in above table to
|
||
|
* access each row data. When row stride is equal to {@code width * (10 / 8)}, there
|
||
|
* will be no padding bytes at the end of each row, the entire image data is
|
||
|
* densely packed. When stride is larger than {@code width * (10 / 8)}, padding
|
||
|
* bytes will be present at the end of each row.
|
||
|
* </p>
|
||
|
* <p>
|
||
|
* For example, the {@link android.media.Image} object can provide data in
|
||
|
* this format from a {@link android.hardware.camera2.CameraDevice} (if
|
||
|
* supported) through a {@link android.media.ImageReader} object. The
|
||
|
* {@link android.media.Image#getPlanes() Image#getPlanes()} will return a
|
||
|
* single plane containing the pixel data. The pixel stride is always 0 in
|
||
|
* {@link android.media.Image.Plane#getPixelStride()}, and the
|
||
|
* {@link android.media.Image.Plane#getRowStride()} describes the vertical
|
||
|
* neighboring pixel distance (in bytes) between adjacent rows.
|
||
|
* </p>
|
||
|
*
|
||
|
* @see android.media.Image
|
||
|
* @see android.media.ImageReader
|
||
|
* @see android.hardware.camera2.CameraDevice
|
||
|
*/
|
||
|
public static final int RAW10 = 0x25;
|
||
|
|
||
|
/**
|
||
|
* <p>
|
||
|
* Android 12-bit raw format
|
||
|
* </p>
|
||
|
* <p>
|
||
|
* This is a single-plane, 12-bit per pixel, densely packed (in each row),
|
||
|
* unprocessed format, usually representing raw Bayer-pattern images coming
|
||
|
* from an image sensor.
|
||
|
* </p>
|
||
|
* <p>
|
||
|
* In an image buffer with this format, starting from the first pixel of each
|
||
|
* row, each two consecutive pixels are packed into 3 bytes (24 bits). The first
|
||
|
* and second byte contains the top 8 bits of first and second pixel. The third
|
||
|
* byte contains the 4 least significant bits of the two pixels, the exact layout
|
||
|
* data for each two consecutive pixels is illustrated below (Pi[j] stands for
|
||
|
* the jth bit of the ith pixel):
|
||
|
* </p>
|
||
|
* <table>
|
||
|
* <thead>
|
||
|
* <tr>
|
||
|
* <th align="center"></th>
|
||
|
* <th align="center">bit 7</th>
|
||
|
* <th align="center">bit 6</th>
|
||
|
* <th align="center">bit 5</th>
|
||
|
* <th align="center">bit 4</th>
|
||
|
* <th align="center">bit 3</th>
|
||
|
* <th align="center">bit 2</th>
|
||
|
* <th align="center">bit 1</th>
|
||
|
* <th align="center">bit 0</th>
|
||
|
* </tr>
|
||
|
* </thead> <tbody>
|
||
|
* <tr>
|
||
|
* <td align="center">Byte 0:</td>
|
||
|
* <td align="center">P0[11]</td>
|
||
|
* <td align="center">P0[10]</td>
|
||
|
* <td align="center">P0[ 9]</td>
|
||
|
* <td align="center">P0[ 8]</td>
|
||
|
* <td align="center">P0[ 7]</td>
|
||
|
* <td align="center">P0[ 6]</td>
|
||
|
* <td align="center">P0[ 5]</td>
|
||
|
* <td align="center">P0[ 4]</td>
|
||
|
* </tr>
|
||
|
* <tr>
|
||
|
* <td align="center">Byte 1:</td>
|
||
|
* <td align="center">P1[11]</td>
|
||
|
* <td align="center">P1[10]</td>
|
||
|
* <td align="center">P1[ 9]</td>
|
||
|
* <td align="center">P1[ 8]</td>
|
||
|
* <td align="center">P1[ 7]</td>
|
||
|
* <td align="center">P1[ 6]</td>
|
||
|
* <td align="center">P1[ 5]</td>
|
||
|
* <td align="center">P1[ 4]</td>
|
||
|
* </tr>
|
||
|
* <tr>
|
||
|
* <td align="center">Byte 2:</td>
|
||
|
* <td align="center">P1[ 3]</td>
|
||
|
* <td align="center">P1[ 2]</td>
|
||
|
* <td align="center">P1[ 1]</td>
|
||
|
* <td align="center">P1[ 0]</td>
|
||
|
* <td align="center">P0[ 3]</td>
|
||
|
* <td align="center">P0[ 2]</td>
|
||
|
* <td align="center">P0[ 1]</td>
|
||
|
* <td align="center">P0[ 0]</td>
|
||
|
* </tr>
|
||
|
* </tbody>
|
||
|
* </table>
|
||
|
* <p>
|
||
|
* This format assumes
|
||
|
* <ul>
|
||
|
* <li>a width multiple of 4 pixels</li>
|
||
|
* <li>an even height</li>
|
||
|
* </ul>
|
||
|
* </p>
|
||
|
*
|
||
|
* <pre>size = row stride * height</pre> where the row stride is in <em>bytes</em>,
|
||
|
* not pixels.
|
||
|
*
|
||
|
* <p>
|
||
|
* Since this is a densely packed format, the pixel stride is always 0. The
|
||
|
* application must use the pixel data layout defined in above table to
|
||
|
* access each row data. When row stride is equal to {@code width * (12 / 8)}, there
|
||
|
* will be no padding bytes at the end of each row, the entire image data is
|
||
|
* densely packed. When stride is larger than {@code width * (12 / 8)}, padding
|
||
|
* bytes will be present at the end of each row.
|
||
|
* </p>
|
||
|
* <p>
|
||
|
* For example, the {@link android.media.Image} object can provide data in
|
||
|
* this format from a {@link android.hardware.camera2.CameraDevice} (if
|
||
|
* supported) through a {@link android.media.ImageReader} object. The
|
||
|
* {@link android.media.Image#getPlanes() Image#getPlanes()} will return a
|
||
|
* single plane containing the pixel data. The pixel stride is always 0 in
|
||
|
* {@link android.media.Image.Plane#getPixelStride()}, and the
|
||
|
* {@link android.media.Image.Plane#getRowStride()} describes the vertical
|
||
|
* neighboring pixel distance (in bytes) between adjacent rows.
|
||
|
* </p>
|
||
|
*
|
||
|
* @see android.media.Image
|
||
|
* @see android.media.ImageReader
|
||
|
* @see android.hardware.camera2.CameraDevice
|
||
|
*/
|
||
|
public static final int RAW12 = 0x26;
|
||
|
|
||
|
/**
|
||
|
* <p>Android dense depth image format.</p>
|
||
|
*
|
||
|
* <p>Each pixel is 16 bits, representing a depth ranging measurement from a depth camera or
|
||
|
* similar sensor. The 16-bit sample consists of a confidence value and the actual ranging
|
||
|
* measurement.</p>
|
||
|
*
|
||
|
* <p>The confidence value is an estimate of correctness for this sample. It is encoded in the
|
||
|
* 3 most significant bits of the sample, with a value of 0 representing 100% confidence, a
|
||
|
* value of 1 representing 0% confidence, a value of 2 representing 1/7, a value of 3
|
||
|
* representing 2/7, and so on.</p>
|
||
|
*
|
||
|
* <p>As an example, the following sample extracts the range and confidence from the first pixel
|
||
|
* of a DEPTH16-format {@link android.media.Image}, and converts the confidence to a
|
||
|
* floating-point value between 0 and 1.f inclusive, with 1.f representing maximum confidence:
|
||
|
*
|
||
|
* <pre>
|
||
|
* ShortBuffer shortDepthBuffer = img.getPlanes()[0].getBuffer().asShortBuffer();
|
||
|
* short depthSample = shortDepthBuffer.get()
|
||
|
* short depthRange = (short) (depthSample & 0x1FFF);
|
||
|
* short depthConfidence = (short) ((depthSample >> 13) & 0x7);
|
||
|
* float depthPercentage = depthConfidence == 0 ? 1.f : (depthConfidence - 1) / 7.f;
|
||
|
* </pre>
|
||
|
* </p>
|
||
|
*
|
||
|
* <p>This format assumes
|
||
|
* <ul>
|
||
|
* <li>an even width</li>
|
||
|
* <li>an even height</li>
|
||
|
* <li>a horizontal stride multiple of 16 pixels</li>
|
||
|
* </ul>
|
||
|
* </p>
|
||
|
*
|
||
|
* <pre> y_size = stride * height </pre>
|
||
|
*
|
||
|
* When produced by a camera, the units for the range are millimeters.
|
||
|
*/
|
||
|
public static final int DEPTH16 = 0x44363159;
|
||
|
|
||
|
/**
|
||
|
* Android sparse depth point cloud format.
|
||
|
*
|
||
|
* <p>A variable-length list of 3D points plus a confidence value, with each point represented
|
||
|
* by four floats; first the X, Y, Z position coordinates, and then the confidence value.</p>
|
||
|
*
|
||
|
* <p>The number of points is {@code (size of the buffer in bytes) / 16}.
|
||
|
*
|
||
|
* <p>The coordinate system and units of the position values depend on the source of the point
|
||
|
* cloud data. The confidence value is between 0.f and 1.f, inclusive, with 0 representing 0%
|
||
|
* confidence and 1.f representing 100% confidence in the measured position values.</p>
|
||
|
*
|
||
|
* <p>As an example, the following code extracts the first depth point in a DEPTH_POINT_CLOUD
|
||
|
* format {@link android.media.Image}:
|
||
|
* <pre>
|
||
|
* FloatBuffer floatDepthBuffer = img.getPlanes()[0].getBuffer().asFloatBuffer();
|
||
|
* float x = floatDepthBuffer.get();
|
||
|
* float y = floatDepthBuffer.get();
|
||
|
* float z = floatDepthBuffer.get();
|
||
|
* float confidence = floatDepthBuffer.get();
|
||
|
* </pre>
|
||
|
*
|
||
|
* For camera devices that support the
|
||
|
* {@link android.hardware.camera2.CameraCharacteristics#REQUEST_AVAILABLE_CAPABILITIES_DEPTH_OUTPUT DEPTH_OUTPUT}
|
||
|
* capability, DEPTH_POINT_CLOUD coordinates have units of meters, and the coordinate system is
|
||
|
* defined by the camera's pose transforms:
|
||
|
* {@link android.hardware.camera2.CameraCharacteristics#LENS_POSE_TRANSLATION} and
|
||
|
* {@link android.hardware.camera2.CameraCharacteristics#LENS_POSE_ROTATION}. That means the origin is
|
||
|
* the optical center of the camera device, and the positive Z axis points along the camera's optical axis,
|
||
|
* toward the scene.
|
||
|
*/
|
||
|
public static final int DEPTH_POINT_CLOUD = 0x101;
|
||
|
|
||
|
/**
|
||
|
* Unprocessed implementation-dependent raw
|
||
|
* depth measurements, opaque with 16 bit
|
||
|
* samples.
|
||
|
*
|
||
|
* @hide
|
||
|
*/
|
||
|
public static final int RAW_DEPTH = 0x1002;
|
||
|
|
||
|
/**
|
||
|
* Unprocessed implementation-dependent raw
|
||
|
* depth measurements, opaque with 10 bit
|
||
|
* samples and device specific bit layout.
|
||
|
*
|
||
|
* @hide
|
||
|
*/
|
||
|
public static final int RAW_DEPTH10 = 0x1003;
|
||
|
|
||
|
/**
|
||
|
* Android private opaque image format.
|
||
|
* <p>
|
||
|
* The choices of the actual format and pixel data layout are entirely up to
|
||
|
* the device-specific and framework internal implementations, and may vary
|
||
|
* depending on use cases even for the same device. The buffers of this
|
||
|
* format can be produced by components like
|
||
|
* {@link android.media.ImageWriter ImageWriter} , and interpreted correctly
|
||
|
* by consumers like {@link android.hardware.camera2.CameraDevice
|
||
|
* CameraDevice} based on the device/framework private information. However,
|
||
|
* these buffers are not directly accessible to the application.
|
||
|
* </p>
|
||
|
* <p>
|
||
|
* When an {@link android.media.Image Image} of this format is obtained from
|
||
|
* an {@link android.media.ImageReader ImageReader} or
|
||
|
* {@link android.media.ImageWriter ImageWriter}, the
|
||
|
* {@link android.media.Image#getPlanes() getPlanes()} method will return an
|
||
|
* empty {@link android.media.Image.Plane Plane} array.
|
||
|
* </p>
|
||
|
* <p>
|
||
|
* If a buffer of this format is to be used as an OpenGL ES texture, the
|
||
|
* framework will assume that sampling the texture will always return an
|
||
|
* alpha value of 1.0 (i.e. the buffer contains only opaque pixel values).
|
||
|
* </p>
|
||
|
*/
|
||
|
public static final int PRIVATE = 0x22;
|
||
|
|
||
|
/**
|
||
|
* Compressed HEIC format.
|
||
|
*
|
||
|
* <p>This format defines the HEIC brand of High Efficiency Image File
|
||
|
* Format as described in ISO/IEC 23008-12.</p>
|
||
|
*/
|
||
|
public static final int HEIC = 0x48454946;
|
||
|
|
||
|
/**
|
||
|
* Use this function to retrieve the number of bits per pixel of an
|
||
|
* ImageFormat.
|
||
|
*
|
||
|
* @param format
|
||
|
* @return the number of bits per pixel of the given format or -1 if the
|
||
|
* format doesn't exist or is not supported.
|
||
|
*/
|
||
|
public static int getBitsPerPixel(@Format int format) {
|
||
|
switch (format) {
|
||
|
case RGB_565:
|
||
|
return 16;
|
||
|
case NV16:
|
||
|
return 16;
|
||
|
case YUY2:
|
||
|
return 16;
|
||
|
case YV12:
|
||
|
return 12;
|
||
|
case Y8:
|
||
|
return 8;
|
||
|
case Y16:
|
||
|
case DEPTH16:
|
||
|
return 16;
|
||
|
case NV21:
|
||
|
return 12;
|
||
|
case YUV_420_888:
|
||
|
return 12;
|
||
|
case YUV_422_888:
|
||
|
return 16;
|
||
|
case YUV_444_888:
|
||
|
return 24;
|
||
|
case FLEX_RGB_888:
|
||
|
return 24;
|
||
|
case FLEX_RGBA_8888:
|
||
|
return 32;
|
||
|
case RAW_DEPTH:
|
||
|
case RAW_SENSOR:
|
||
|
return 16;
|
||
|
case YCBCR_P010:
|
||
|
return 24;
|
||
|
case RAW_DEPTH10:
|
||
|
case RAW10:
|
||
|
return 10;
|
||
|
case RAW12:
|
||
|
return 12;
|
||
|
}
|
||
|
return -1;
|
||
|
}
|
||
|
|
||
|
/**
|
||
|
* Determine whether or not this is a public-visible {@code format}.
|
||
|
*
|
||
|
* <p>In particular, {@code @hide} formats will return {@code false}.</p>
|
||
|
*
|
||
|
* <p>Any other formats (including UNKNOWN) will return {@code false}.</p>
|
||
|
*
|
||
|
* @param format an integer format
|
||
|
* @return a boolean
|
||
|
*
|
||
|
* @hide
|
||
|
*/
|
||
|
public static boolean isPublicFormat(@Format int format) {
|
||
|
switch (format) {
|
||
|
case RGB_565:
|
||
|
case NV16:
|
||
|
case YUY2:
|
||
|
case YV12:
|
||
|
case JPEG:
|
||
|
case NV21:
|
||
|
case YUV_420_888:
|
||
|
case YUV_422_888:
|
||
|
case YUV_444_888:
|
||
|
case YCBCR_P010:
|
||
|
case FLEX_RGB_888:
|
||
|
case FLEX_RGBA_8888:
|
||
|
case RAW_SENSOR:
|
||
|
case RAW_PRIVATE:
|
||
|
case RAW10:
|
||
|
case RAW12:
|
||
|
case DEPTH16:
|
||
|
case DEPTH_POINT_CLOUD:
|
||
|
case PRIVATE:
|
||
|
case RAW_DEPTH:
|
||
|
case RAW_DEPTH10:
|
||
|
case Y8:
|
||
|
case DEPTH_JPEG:
|
||
|
case HEIC:
|
||
|
case JPEG_R:
|
||
|
return true;
|
||
|
}
|
||
|
|
||
|
return false;
|
||
|
}
|
||
|
}
|