/* * Copyright (C) 2012 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package android.media; import static android.media.codec.Flags.FLAG_IN_PROCESS_SW_AUDIO_CODEC; import static android.media.codec.Flags.FLAG_REGION_OF_INTEREST; import static com.android.media.codec.flags.Flags.FLAG_CODEC_IMPORTANCE; import static com.android.media.codec.flags.Flags.FLAG_LARGE_AUDIO_FRAME; import android.annotation.FlaggedApi; import android.annotation.IntDef; import android.annotation.NonNull; import android.annotation.Nullable; import android.compat.annotation.UnsupportedAppUsage; import android.graphics.Rect; import android.text.TextUtils; import java.lang.annotation.Retention; import java.lang.annotation.RetentionPolicy; import java.nio.ByteBuffer; import java.nio.ByteOrder; import java.util.AbstractSet; import java.util.HashMap; import java.util.Iterator; import java.util.List; import java.util.Map; import java.util.Set; import java.util.stream.Collectors; /** * Encapsulates the information describing the format of media data, be it audio or video, as * well as optional feature metadata. *

* The format of the media data is specified as key/value pairs. Keys are strings. Values can * be integer, long, float, String or ByteBuffer. *

* The feature metadata is specificed as string/boolean pairs. *

* Keys common to all audio/video formats, all keys not marked optional are mandatory: * * * * * * * * * * *
NameValue TypeDescription
{@link #KEY_MIME}StringThe type of the format.
{@link #KEY_CODECS_STRING}Stringoptional, the RFC 6381 codecs string of the MediaFormat
{@link #KEY_MAX_INPUT_SIZE}Integeroptional, maximum size of a buffer of input data
{@link #KEY_PIXEL_ASPECT_RATIO_WIDTH}Integeroptional, the pixel aspect ratio width
{@link #KEY_PIXEL_ASPECT_RATIO_HEIGHT}Integeroptional, the pixel aspect ratio height
{@link #KEY_BIT_RATE}Integerencoder-only, desired bitrate in bits/second
{@link #KEY_DURATION}longthe duration of the content (in microseconds)
* * Video formats have the following keys: * * * * * * * * * * * * * * * *
NameValue TypeDescription
{@link #KEY_WIDTH}Integer
{@link #KEY_HEIGHT}Integer
{@link #KEY_COLOR_FORMAT}Integerset by the user * for encoders, readable in the output format of decoders
{@link #KEY_FRAME_RATE}Integer or Floatrequired for encoders, * optional for decoders
{@link #KEY_CAPTURE_RATE}Integer
{@link #KEY_I_FRAME_INTERVAL}Integer (or Float)encoder-only, * time-interval between key frames. * Float support added in {@link android.os.Build.VERSION_CODES#N_MR1}
{@link #KEY_INTRA_REFRESH_PERIOD}Integerencoder-only, optional
{@link #KEY_LATENCY}Integerencoder-only, optional
{@link #KEY_MAX_WIDTH}Integerdecoder-only, optional, max-resolution width
{@link #KEY_MAX_HEIGHT}Integerdecoder-only, optional, max-resolution height
{@link #KEY_REPEAT_PREVIOUS_FRAME_AFTER}Longencoder in surface-mode * only, optional
{@link #KEY_PUSH_BLANK_BUFFERS_ON_STOP}Integer(1)decoder rendering * to a surface only, optional
{@link #KEY_TEMPORAL_LAYERING}Stringencoder only, optional, * temporal-layering schema
* Specify both {@link #KEY_MAX_WIDTH} and {@link #KEY_MAX_HEIGHT} to enable * adaptive playback (seamless resolution change) for a video decoder that * supports it ({@link MediaCodecInfo.CodecCapabilities#FEATURE_AdaptivePlayback}). * The values are used as hints for the codec: they are the maximum expected * resolution to prepare for. Depending on codec support, preparing for larger * maximum resolution may require more memory even if that resolution is never * reached. These fields have no effect for codecs that do not support adaptive * playback.

* * Audio formats have the following keys: * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
NameValue TypeDescription
{@link #KEY_CHANNEL_COUNT}Integer
{@link #KEY_SAMPLE_RATE}Integer
{@link #KEY_PCM_ENCODING}Integeroptional
{@link #KEY_IS_ADTS}Integeroptional, if decoding AAC audio content, setting this key to 1 indicates that each audio frame is prefixed by the ADTS header.
{@link #KEY_AAC_PROFILE}Integerencoder-only, optional, if content is AAC audio, specifies the desired profile.
{@link #KEY_AAC_SBR_MODE}Integerencoder-only, optional, if content is AAC audio, specifies the desired SBR mode.
{@link #KEY_AAC_DRC_TARGET_REFERENCE_LEVEL}Integerdecoder-only, optional, if content is AAC audio, specifies the target reference level.
{@link #KEY_AAC_ENCODED_TARGET_LEVEL}Integerdecoder-only, optional, if content is AAC audio, specifies the target reference level used at encoder.
{@link #KEY_AAC_DRC_BOOST_FACTOR}Integerdecoder-only, optional, if content is AAC audio, specifies the DRC boost factor.
{@link #KEY_AAC_DRC_ATTENUATION_FACTOR}Integerdecoder-only, optional, if content is AAC audio, specifies the DRC attenuation factor.
{@link #KEY_AAC_DRC_HEAVY_COMPRESSION}Integerdecoder-only, optional, if content is AAC audio, specifies whether to use heavy compression.
{@link #KEY_AAC_MAX_OUTPUT_CHANNEL_COUNT}Integerdecoder-only, optional, if content is AAC audio, specifies the maximum number of channels the decoder outputs.
{@link #KEY_AAC_DRC_EFFECT_TYPE}Integerdecoder-only, optional, if content is AAC audio, specifies the MPEG-D DRC effect type to use.
{@link #KEY_AAC_DRC_OUTPUT_LOUDNESS}Integerdecoder-only, optional, if content is AAC audio, returns the DRC output loudness.
{@link #KEY_AAC_DRC_ALBUM_MODE}Integerdecoder-only, optional, if content is AAC audio, specifies the whether MPEG-D DRC Album Mode is active or not.
{@link #KEY_CHANNEL_MASK}Integeroptional, a mask of audio channel assignments
{@link #KEY_ENCODER_DELAY}Integeroptional, the number of frames to trim from the start of the decoded audio stream.
{@link #KEY_ENCODER_PADDING}Integeroptional, the number of frames to trim from the end of the decoded audio stream.
{@link #KEY_FLAC_COMPRESSION_LEVEL}Integerencoder-only, optional, if content is FLAC audio, specifies the desired compression level.
{@link #KEY_MPEGH_PROFILE_LEVEL_INDICATION}Integerdecoder-only, optional, if content is MPEG-H audio, * specifies the profile and level of the stream.
{@link #KEY_MPEGH_COMPATIBLE_SETS}ByteBufferdecoder-only, optional, if content is MPEG-H audio, * specifies the compatible sets (profile and level) of the stream.
{@link #KEY_MPEGH_REFERENCE_CHANNEL_LAYOUT}Integerdecoder-only, optional, if content is MPEG-H audio, * specifies the preferred reference channel layout of the stream.
{@link #KEY_MAX_BUFFER_BATCH_OUTPUT_SIZE}Integeroptional, used with * large audio frame support, specifies max size of output buffer in bytes.
{@link #KEY_BUFFER_BATCH_THRESHOLD_OUTPUT_SIZE}Integeroptional, * used with large audio frame support, specifies threshold output size in bytes.
* * Subtitle formats have the following keys: * * * * *
{@link #KEY_MIME}StringThe type of the format.
{@link #KEY_LANGUAGE}StringThe language of the content.
{@link #KEY_CAPTION_SERVICE_NUMBER}intoptional, the closed-caption service or channel number.
* * Image formats have the following keys: * * * * * * * * * *
{@link #KEY_MIME}StringThe type of the format.
{@link #KEY_WIDTH}Integer
{@link #KEY_HEIGHT}Integer
{@link #KEY_COLOR_FORMAT}Integerset by the user * for encoders, readable in the output format of decoders
{@link #KEY_TILE_WIDTH}Integerrequired if the image has grid
{@link #KEY_TILE_HEIGHT}Integerrequired if the image has grid
{@link #KEY_GRID_ROWS}Integerrequired if the image has grid
{@link #KEY_GRID_COLUMNS}Integerrequired if the image has grid
*/ public final class MediaFormat { public static final String MIMETYPE_VIDEO_VP8 = "video/x-vnd.on2.vp8"; public static final String MIMETYPE_VIDEO_VP9 = "video/x-vnd.on2.vp9"; public static final String MIMETYPE_VIDEO_AV1 = "video/av01"; public static final String MIMETYPE_VIDEO_AVC = "video/avc"; public static final String MIMETYPE_VIDEO_HEVC = "video/hevc"; public static final String MIMETYPE_VIDEO_MPEG4 = "video/mp4v-es"; public static final String MIMETYPE_VIDEO_H263 = "video/3gpp"; public static final String MIMETYPE_VIDEO_MPEG2 = "video/mpeg2"; public static final String MIMETYPE_VIDEO_RAW = "video/raw"; public static final String MIMETYPE_VIDEO_DOLBY_VISION = "video/dolby-vision"; public static final String MIMETYPE_VIDEO_SCRAMBLED = "video/scrambled"; public static final String MIMETYPE_AUDIO_AMR_NB = "audio/3gpp"; public static final String MIMETYPE_AUDIO_AMR_WB = "audio/amr-wb"; public static final String MIMETYPE_AUDIO_MPEG = "audio/mpeg"; public static final String MIMETYPE_AUDIO_AAC = "audio/mp4a-latm"; public static final String MIMETYPE_AUDIO_QCELP = "audio/qcelp"; public static final String MIMETYPE_AUDIO_VORBIS = "audio/vorbis"; public static final String MIMETYPE_AUDIO_OPUS = "audio/opus"; public static final String MIMETYPE_AUDIO_G711_ALAW = "audio/g711-alaw"; public static final String MIMETYPE_AUDIO_G711_MLAW = "audio/g711-mlaw"; public static final String MIMETYPE_AUDIO_RAW = "audio/raw"; public static final String MIMETYPE_AUDIO_FLAC = "audio/flac"; public static final String MIMETYPE_AUDIO_MSGSM = "audio/gsm"; public static final String MIMETYPE_AUDIO_AC3 = "audio/ac3"; public static final String MIMETYPE_AUDIO_EAC3 = "audio/eac3"; public static final String MIMETYPE_AUDIO_EAC3_JOC = "audio/eac3-joc"; public static final String MIMETYPE_AUDIO_AC4 = "audio/ac4"; public static final String MIMETYPE_AUDIO_SCRAMBLED = "audio/scrambled"; /** MIME type for MPEG-H Audio single stream */ public static final String MIMETYPE_AUDIO_MPEGH_MHA1 = "audio/mha1"; /** MIME type for MPEG-H Audio single stream, encapsulated in MHAS */ public static final String MIMETYPE_AUDIO_MPEGH_MHM1 = "audio/mhm1"; /** MIME type for DTS Digital Surround (up to 5.1 channels) audio stream, aka DTS-CA. */ public static final String MIMETYPE_AUDIO_DTS = "audio/vnd.dts"; /** * MIME type for DTS HD (up to 7.1 channels) audio stream. * With codec profile DTS_HDProfileHRA represents DTS HD High Resolution Audio. * With codec profile DTS_HDProfileMA represents DTS HD Master Audio. * With codec profile DTS_HDProfileLBR represents DTS Express. */ public static final String MIMETYPE_AUDIO_DTS_HD = "audio/vnd.dts.hd"; /** * MIME type for DTS UHD (object-based) audio stream, aka DTS:X. * With codec profile DTS_UHDProfileP1 represents DTS-UHD P1. * With codec profile DTS_UHDProfileP2 represents DTS-UHD P2. */ public static final String MIMETYPE_AUDIO_DTS_UHD = "audio/vnd.dts.uhd"; /** MIME type for Dynamic Resolution Adaptation (DRA) audio stream. */ public static final String MIMETYPE_AUDIO_DRA = "audio/vnd.dra"; /** MIME type for Dolby Metadata-enhanced Audio Transmission (MAT) audio stream. */ public static final String MIMETYPE_AUDIO_DOLBY_MAT = "audio/vnd.dolby.mat"; /** MIME type for Dolby TrueHD audio format, based on Meridian Lossless Packing (MLP). */ public static final String MIMETYPE_AUDIO_DOLBY_TRUEHD = "audio/vnd.dolby.mlp"; /** * MIME type for AAC Low Complexity (LC) audio stream. Uses the scheme defined by * RFC 6381 with OTI of MPEG-4 (40) and AOT of AAC LC (2) from ISO/IEC 14496-3. */ public static final String MIMETYPE_AUDIO_AAC_LC = "audio/mp4a.40.02"; /** * MIME type for HE-AAC v1 (LC + SBR) audio stream. Uses the scheme defined by * RFC 6381 with OTI of MPEG-4 (40) and AOT of AAC SBR (5) from ISO/IEC 14496-3. */ public static final String MIMETYPE_AUDIO_AAC_HE_V1 = "audio/mp4a.40.05"; /** * MIME type for HE-AAC v2 (LC + SBR + PS) audio stream. Uses the scheme defined by * RFC 6381 with OTI of MPEG-4 (40) and AOT of PS (29) from ISO/IEC 14496-3. */ public static final String MIMETYPE_AUDIO_AAC_HE_V2 = "audio/mp4a.40.29"; /** * MIME type for AAC Enhanced Low Delay (ELD) audio stream. Uses the scheme defined by * RFC 6381 with OTI of MPEG-4 (40) and AOT of ELD (39) from ISO/IEC 14496-3. */ public static final String MIMETYPE_AUDIO_AAC_ELD = "audio/mp4a.40.39"; /** * MIME type for AAC XHE audio stream. Uses the scheme defined by * RFC 6381 with OTI of MPEG-4 (40) and AOT of USAC (42) from ISO/IEC 14496-3. */ public static final String MIMETYPE_AUDIO_AAC_XHE = "audio/mp4a.40.42"; /** * MIME type for MPEG-H Baseline (BL) Profile L3 audio stream. Uses the scheme defined by * RFC 6381 with mpegh3daProfileLevelIndication for main profile/L3 (0x3) from ISO/IEC 23008-3. */ public static final String MIMETYPE_AUDIO_MPEGH_BL_L3 = "audio/mhm1.03"; /** * MIME type for MPEG-H Baseline (BL) Profile L4 audio stream. Uses the scheme defined by * RFC 6381 with mpegh3daProfileLevelIndication for main profile/L4 (0x4) from ISO/IEC 23008-3. */ public static final String MIMETYPE_AUDIO_MPEGH_BL_L4 = "audio/mhm1.04"; /** * MIME type for MPEG-H Low Complexity (LC) L3 audio stream. Uses the scheme defined by * RFC 6381 with mpegh3daProfileLevelIndication for LC profile/L3 (0xD) from ISO/IEC 23008-3. */ public static final String MIMETYPE_AUDIO_MPEGH_LC_L3 = "audio/mhm1.0d"; /** * MIME type for MPEG-H Low Complexity (LC) L4 audio stream. Uses the scheme defined by * RFC 6381 with mpegh3daProfileLevelIndication for LC profile/L4 (0xE) from ISO/IEC 23008-3. */ public static final String MIMETYPE_AUDIO_MPEGH_LC_L4 = "audio/mhm1.0e"; /** * MIME type for the IEC61937 audio stream encapsulation. This type isn't defined by IANA. */ public static final String MIMETYPE_AUDIO_IEC61937 = "audio/x-iec61937"; /** * MIME type for HEIF still image data encoded in HEVC. * * To decode such an image, {@link MediaCodec} decoder for * {@link #MIMETYPE_VIDEO_HEVC} shall be used. The client needs to form * the correct {@link #MediaFormat} based on additional information in * the track format (shown in the next paragraph), and send it to * {@link MediaCodec#configure}. * * The track's MediaFormat will come with {@link #KEY_WIDTH} and * {@link #KEY_HEIGHT} keys, which describes the width and height * of the image. If the image doesn't contain grid (i.e. none of * {@link #KEY_TILE_WIDTH}, {@link #KEY_TILE_HEIGHT}, * {@link #KEY_GRID_ROWS}, {@link #KEY_GRID_COLUMNS} are present), the * track will contain a single sample of coded data for the entire image, * and the image width and height should be used to set up the decoder. * * If the image does come with grid, each sample from the track will * contain one tile in the grid, of which the size is described by * {@link #KEY_TILE_WIDTH} and {@link #KEY_TILE_HEIGHT}. This size * (instead of {@link #KEY_WIDTH} and {@link #KEY_HEIGHT}) should be * used to set up the decoder. The track contains {@link #KEY_GRID_ROWS} * by {@link #KEY_GRID_COLUMNS} samples in row-major, top-row first, * left-to-right order. The output image should be reconstructed by * first tiling the decoding results of the tiles in the correct order, * then trimming (before rotation is applied) on the bottom and right * side, if the tiled area is larger than the image width and height. */ public static final String MIMETYPE_IMAGE_ANDROID_HEIC = "image/vnd.android.heic"; /** * MIME type for AVIF still image data encoded in AV1. * * To decode such an image, {@link MediaCodec} decoder for * {@link #MIMETYPE_VIDEO_AV1} shall be used. The client needs to form * the correct {@link #MediaFormat} based on additional information in * the track format (shown in the next paragraph), and send it to * {@link MediaCodec#configure}. * * The track's MediaFormat will come with {@link #KEY_WIDTH} and * {@link #KEY_HEIGHT} keys, which describes the width and height * of the image. If the image doesn't contain grid (i.e. none of * {@link #KEY_TILE_WIDTH}, {@link #KEY_TILE_HEIGHT}, * {@link #KEY_GRID_ROWS}, {@link #KEY_GRID_COLUMNS} are present), the * track will contain a single sample of coded data for the entire image, * and the image width and height should be used to set up the decoder. * * If the image does come with grid, each sample from the track will * contain one tile in the grid, of which the size is described by * {@link #KEY_TILE_WIDTH} and {@link #KEY_TILE_HEIGHT}. This size * (instead of {@link #KEY_WIDTH} and {@link #KEY_HEIGHT}) should be * used to set up the decoder. The track contains {@link #KEY_GRID_ROWS} * by {@link #KEY_GRID_COLUMNS} samples in row-major, top-row first, * left-to-right order. The output image should be reconstructed by * first tiling the decoding results of the tiles in the correct order, * then trimming (before rotation is applied) on the bottom and right * side, if the tiled area is larger than the image width and height. */ public static final String MIMETYPE_IMAGE_AVIF = "image/avif"; /** * MIME type for WebVTT subtitle data. */ public static final String MIMETYPE_TEXT_VTT = "text/vtt"; /** * MIME type for SubRip (SRT) container. */ public static final String MIMETYPE_TEXT_SUBRIP = "application/x-subrip"; /** * MIME type for CEA-608 closed caption data. */ public static final String MIMETYPE_TEXT_CEA_608 = "text/cea-608"; /** * MIME type for CEA-708 closed caption data. */ public static final String MIMETYPE_TEXT_CEA_708 = "text/cea-708"; @UnsupportedAppUsage private Map mMap; /** * A key describing the log session ID for MediaCodec. The log session ID is a random 32-byte * hexadecimal string that is used to associate metrics from multiple media codec instances * to the same playback or recording session. The value is created as * {@link android.media.metrics.LogSessionId LogSessionId}. Sessions are created in * {@link android.media.metrics.MediaMetricsManager MediaMetricsManager}. * The associated value is a string. */ public static final String LOG_SESSION_ID = "log-session-id"; /** * A key describing the mime type of the MediaFormat. * The associated value is a string. */ public static final String KEY_MIME = "mime"; /** * A key describing the codecs string of the MediaFormat. See RFC 6381 section 3.2 for the * syntax of the value. The value does not hold {@link MediaCodec}-exposed codec names. * The associated value is a string. * * @see MediaParser.TrackData#mediaFormat */ public static final String KEY_CODECS_STRING = "codecs-string"; /** * An optional key describing the low latency decoding mode. This is an optional parameter * that applies only to decoders. If enabled, the decoder doesn't hold input and output * data more than required by the codec standards. * The associated value is an integer (0 or 1): 1 when low-latency decoding is enabled, * 0 otherwise. The default value is 0. */ public static final String KEY_LOW_LATENCY = "low-latency"; /** * A key describing the language of the content, using either ISO 639-1 * or 639-2/T codes. The associated value is a string. */ public static final String KEY_LANGUAGE = "language"; /** * A key describing the closed caption service number. For CEA-608 caption tracks, holds the * channel number. For CEA-708, holds the service number. * The associated value is an int. */ public static final String KEY_CAPTION_SERVICE_NUMBER = "caption-service-number"; /** * A key describing the sample rate of an audio format. * The associated value is an integer */ public static final String KEY_SAMPLE_RATE = "sample-rate"; /** * A key describing the number of channels in an audio format. * The associated value is an integer */ public static final String KEY_CHANNEL_COUNT = "channel-count"; /** * A key describing the width of the content in a video format. * The associated value is an integer */ public static final String KEY_WIDTH = "width"; /** * A key describing the height of the content in a video format. * The associated value is an integer */ public static final String KEY_HEIGHT = "height"; /** * A key describing the bottom-coordinate (y) of the crop rectangle. * This is the bottom-most row included in the crop frame, * where row indices start at 0. * Additional information on the crop rectangle semantics can be found at * {@link android.media.MediaCodec}. */ public static final String KEY_CROP_BOTTOM = "crop-bottom"; /** * A key describing the left-coordinate (x) of the crop rectangle. * This is the left-most column included in the crop frame, * where column indices start at 0. * Additional information on the crop rectangle semantics can be found at * {@link android.media.MediaCodec}. */ public static final String KEY_CROP_LEFT = "crop-left"; /** * A key describing the right-coordinate (x) of the crop rectangle. * This is the right-most column included in the crop frame, * where column indices start at 0. * Additional information on the crop rectangle semantics can be found at * {@link android.media.MediaCodec}. */ public static final String KEY_CROP_RIGHT = "crop-right"; /** * A key describing the top-coordinate (y) of the crop rectangle. * This is the top-most row included in the crop frame, * where row indices start at 0. * Additional information on the crop rectangle semantics can be found at * {@link android.media.MediaCodec}. */ public static final String KEY_CROP_TOP = "crop-top"; /** * A key describing the maximum expected width of the content in a video * decoder format, in case there are resolution changes in the video content. * The associated value is an integer */ public static final String KEY_MAX_WIDTH = "max-width"; /** * A key describing the maximum expected height of the content in a video * decoder format, in case there are resolution changes in the video content. * The associated value is an integer */ public static final String KEY_MAX_HEIGHT = "max-height"; /** A key describing the maximum size in bytes of a buffer of data * described by this MediaFormat. * The associated value is an integer */ public static final String KEY_MAX_INPUT_SIZE = "max-input-size"; /** * A key describing the maximum output buffer size in bytes when using * large buffer mode containing multiple access units. * * When not-set - codec functions with one access-unit per frame. * When set less than the size of two access-units - will make codec * operate in single access-unit per output frame. * When set to a value too big - The component or the framework will * override this value to a reasonable max size not exceeding typical * 10 seconds of data (device dependent) when set to a value larger than * that. The value final value used will be returned in the output format. * * The associated value is an integer * * @see FEATURE_MultipleFrames */ @FlaggedApi(FLAG_LARGE_AUDIO_FRAME) public static final String KEY_BUFFER_BATCH_MAX_OUTPUT_SIZE = "buffer-batch-max-output-size"; /** * A key describing the threshold output size in bytes when using large buffer * mode containing multiple access units. * * This is an optional parameter. * * If not set - the component can set this to a reasonable value. * If set larger than max size, the components will * clip this setting to maximum buffer batching output size. * * The component will return a partial output buffer if the output buffer reaches or * surpass this limit. * * Threshold size should be always less or equal to KEY_MAX_BUFFER_BATCH_OUTPUT_SIZE. * The final setting of this value as determined by the component will be returned * in the output format * * The associated value is an integer * * @see FEATURE_MultipleFrames */ @FlaggedApi(FLAG_LARGE_AUDIO_FRAME) public static final String KEY_BUFFER_BATCH_THRESHOLD_OUTPUT_SIZE = "buffer-batch-threshold-output-size"; /** * A key describing the pixel aspect ratio width. * The associated value is an integer */ public static final String KEY_PIXEL_ASPECT_RATIO_WIDTH = "sar-width"; /** * A key describing the pixel aspect ratio height. * The associated value is an integer */ public static final String KEY_PIXEL_ASPECT_RATIO_HEIGHT = "sar-height"; /** * A key describing the average bitrate in bits/sec. * The associated value is an integer */ public static final String KEY_BIT_RATE = "bitrate"; /** * A key describing the hardware AV sync id. * The associated value is an integer * * See android.media.tv.tuner.Tuner#getAvSyncHwId. */ public static final String KEY_HARDWARE_AV_SYNC_ID = "hw-av-sync-id"; /** * A key describing the max bitrate in bits/sec. * This is usually over a one-second sliding window (e.g. over any window of one second). * The associated value is an integer * @hide */ public static final String KEY_MAX_BIT_RATE = "max-bitrate"; /** * A key describing the color format of the content in a video format. * Constants are declared in {@link android.media.MediaCodecInfo.CodecCapabilities}. */ public static final String KEY_COLOR_FORMAT = "color-format"; /** * A key describing the frame rate of a video format in frames/sec. *

* The associated value is normally an integer when the value is used by the platform, * but video codecs also accept float configuration values. * Specifically, {@link MediaExtractor#getTrackFormat MediaExtractor} provides an integer * value corresponding to the frame rate information of the track if specified and non-zero. * Otherwise, this key is not present. {@link MediaCodec#configure MediaCodec} accepts both * float and integer values. *

* This represents the desired operating frame rate if the * {@link #KEY_OPERATING_RATE} is not present and {@link #KEY_PRIORITY} is {@code 0} * (realtime). Otherwise, this is just informational. *

* For video encoders this value corresponds to the intended frame rate (the rate at which * the application intends to send frames to the encoder, as calculated by the buffer * timestamps, and not from the actual real-time rate that the frames are sent to * the encoder). Encoders use this hint for rate control, specifically for the initial * frames, as encoders are expected to support variable frame rate (for rate control) based * on the actual {@link MediaCodec.BufferInfo#presentationTimeUs buffer timestamps} of * subsequent frames. *

* This key is not used in the {@code MediaCodec} * {@link MediaCodec#getInputFormat input}/{@link MediaCodec#getOutputFormat output} formats, * nor by {@link MediaMuxer#addTrack MediaMuxer}. */ public static final String KEY_FRAME_RATE = "frame-rate"; /** * A key describing the width (in pixels) of each tile of the content in a * {@link #MIMETYPE_IMAGE_ANDROID_HEIC} / {@link #MIMETYPE_IMAGE_AVIF} track. * The associated value is an integer. * * Refer to {@link #MIMETYPE_IMAGE_ANDROID_HEIC} / {@link #MIMETYPE_IMAGE_AVIF} on decoding * instructions of such tracks. * * @see #KEY_TILE_HEIGHT * @see #KEY_GRID_ROWS * @see #KEY_GRID_COLUMNS */ public static final String KEY_TILE_WIDTH = "tile-width"; /** * A key describing the height (in pixels) of each tile of the content in a * {@link #MIMETYPE_IMAGE_ANDROID_HEIC} / {@link #MIMETYPE_IMAGE_AVIF} track. * The associated value is an integer. * * Refer to {@link #MIMETYPE_IMAGE_ANDROID_HEIC} / {@link #MIMETYPE_IMAGE_AVIF} on decoding * instructions of such tracks. * * @see #KEY_TILE_WIDTH * @see #KEY_GRID_ROWS * @see #KEY_GRID_COLUMNS */ public static final String KEY_TILE_HEIGHT = "tile-height"; /** * A key describing the number of grid rows in the content in a * {@link #MIMETYPE_IMAGE_ANDROID_HEIC} / {@link #MIMETYPE_IMAGE_AVIF} track. * The associated value is an integer. * * Refer to {@link #MIMETYPE_IMAGE_ANDROID_HEIC} / {@link #MIMETYPE_IMAGE_AVIF} on decoding * instructions of such tracks. * * @see #KEY_TILE_WIDTH * @see #KEY_TILE_HEIGHT * @see #KEY_GRID_COLUMNS */ public static final String KEY_GRID_ROWS = "grid-rows"; /** * A key describing the number of grid columns in the content in a * {@link #MIMETYPE_IMAGE_ANDROID_HEIC} / {@link #MIMETYPE_IMAGE_AVIF} track. * The associated value is an integer. * * Refer to {@link #MIMETYPE_IMAGE_ANDROID_HEIC} / {@link #MIMETYPE_IMAGE_AVIF} on decoding * instructions of such tracks. * * @see #KEY_TILE_WIDTH * @see #KEY_TILE_HEIGHT * @see #KEY_GRID_ROWS */ public static final String KEY_GRID_COLUMNS = "grid-cols"; /** * A key describing the raw audio sample encoding/format. * *

The associated value is an integer, using one of the * {@link AudioFormat}.ENCODING_PCM_ values.

* *

This is an optional key for audio decoders and encoders specifying the * desired raw audio sample format during {@link MediaCodec#configure * MediaCodec.configure(…)} call. Use {@link MediaCodec#getInputFormat * MediaCodec.getInput}/{@link MediaCodec#getOutputFormat OutputFormat(…)} * to confirm the actual format. For the PCM decoder this key specifies both * input and output sample encodings.

* *

This key is also used by {@link MediaExtractor} to specify the sample * format of audio data, if it is specified.

* *

If this key is missing, the raw audio sample format is signed 16-bit short.

*/ public static final String KEY_PCM_ENCODING = "pcm-encoding"; /** * A key describing the capture rate of a video format in frames/sec. *

* When capture rate is different than the frame rate, it means that the * video is acquired at a different rate than the playback, which produces * slow motion or timelapse effect during playback. Application can use the * value of this key to tell the relative speed ratio between capture and * playback rates when the video was recorded. *

*

* The associated value is an integer or a float. *

*/ public static final String KEY_CAPTURE_RATE = "capture-rate"; /** * A key for retrieving the slow-motion marker information associated with a video track. *

* The associated value is a ByteBuffer in {@link ByteOrder#BIG_ENDIAN} * (networking order) of the following format: *

*
     *     float(32) playbackRate;
     *     unsigned int(32) numMarkers;
     *     for (i = 0;i < numMarkers; i++) {
     *         int(64) timestampUs;
     *         float(32) speedRatio;
     *     }
* The meaning of each field is as follows: * * * * * * * * * * * * * * * * * * *
playbackRateThe frame rate at which the playback should happen (or the flattened * clip should be).
numMarkersThe number of slow-motion markers that follows.
timestampUsThe starting point of a new segment.
speedRatioThe playback speed for that segment. The playback speed is a floating * point number, indicating how fast the time progresses relative to that * written in the container. (Eg. 4.0 means time goes 4x as fast, which * makes 30fps become 120fps.)
*

* The following constraints apply to the timestampUs of the markers: *

*
  • The timestampUs shall be monotonically increasing.
  • *
  • The timestampUs shall fall within the time span of the video track.
  • *
  • The first timestampUs should match that of the first video sample.
  • */ public static final String KEY_SLOW_MOTION_MARKERS = "slow-motion-markers"; /** * A key describing the frequency of key frames expressed in seconds between key frames. *

    * This key is used by video encoders. * A negative value means no key frames are requested after the first frame. * A zero value means a stream containing all key frames is requested. *

    * Most video encoders will convert this value of the number of non-key-frames between * key-frames, using the {@linkplain #KEY_FRAME_RATE frame rate} information; therefore, * if the actual frame rate differs (e.g. input frames are dropped or the frame rate * changes), the time interval between key frames will not be the * configured value. *

    * The associated value is an integer (or float since * {@link android.os.Build.VERSION_CODES#N_MR1}). */ public static final String KEY_I_FRAME_INTERVAL = "i-frame-interval"; /** * An optional key describing the period of intra refresh in frames. This is an * optional parameter that applies only to video encoders. If encoder supports it * ({@link MediaCodecInfo.CodecCapabilities#FEATURE_IntraRefresh}), the whole * frame is completely refreshed after the specified period. Also for each frame, * a fix subset of macroblocks must be intra coded which leads to more constant bitrate * than inserting a key frame. This key is recommended for video streaming applications * as it provides low-delay and good error-resilience. This key is ignored if the * video encoder does not support the intra refresh feature. Use the output format to * verify that this feature was enabled. * The associated value is an integer. */ public static final String KEY_INTRA_REFRESH_PERIOD = "intra-refresh-period"; /** * An optional key describing whether encoders prepend headers to sync frames (e.g. * SPS and PPS to IDR frames for H.264). This is an optional parameter that applies only * to video encoders. A video encoder may not support this feature; the component will fail * to configure in that case. For other components, this key is ignored. * * The value is an integer, with 1 indicating to prepend headers to every sync frames, * or 0 otherwise. The default value is 0. */ public static final String KEY_PREPEND_HEADER_TO_SYNC_FRAMES = "prepend-sps-pps-to-idr-frames"; /** * A key describing the temporal layering schema. This is an optional parameter * that applies only to video encoders. Use {@link MediaCodec#getOutputFormat} * after {@link MediaCodec#configure configure} to query if the encoder supports * the desired schema. Supported values are {@code webrtc.vp8.N-layer}, * {@code android.generic.N}, {@code android.generic.N+M} and {@code none}, where * {@code N} denotes the total number of non-bidirectional layers (which must be at least 1) * and {@code M} denotes the total number of bidirectional layers (which must be non-negative). *

    {@code android.generic.*} schemas have been added in {@link * android.os.Build.VERSION_CODES#N_MR1}. *

    * The encoder may support fewer temporal layers, in which case the output format * will contain the configured schema. If the encoder does not support temporal * layering, the output format will not have an entry with this key. * The associated value is a string. */ public static final String KEY_TEMPORAL_LAYERING = "ts-schema"; /** * A key describing the stride of the video bytebuffer layout. * Stride (or row increment) is the difference between the index of a pixel * and that of the pixel directly underneath. For YUV 420 formats, the * stride corresponds to the Y plane; the stride of the U and V planes can * be calculated based on the color format, though it is generally undefined * and depends on the device and release. * The associated value is an integer, representing number of bytes. */ public static final String KEY_STRIDE = "stride"; /** * A key describing the plane height of a multi-planar (YUV) video bytebuffer layout. * Slice height (or plane height/vertical stride) is the number of rows that must be skipped * to get from the top of the Y plane to the top of the U plane in the bytebuffer. In essence * the offset of the U plane is sliceHeight * stride. The height of the U/V planes * can be calculated based on the color format, though it is generally undefined * and depends on the device and release. * The associated value is an integer, representing number of rows. */ public static final String KEY_SLICE_HEIGHT = "slice-height"; /** * Applies only when configuring a video encoder in "surface-input" mode. * The associated value is a long and gives the time in microseconds * after which the frame previously submitted to the encoder will be * repeated (once) if no new frame became available since. */ public static final String KEY_REPEAT_PREVIOUS_FRAME_AFTER = "repeat-previous-frame-after"; /** * Instruct the video encoder in "surface-input" mode to drop excessive * frames from the source, so that the input frame rate to the encoder * does not exceed the specified fps. * * The associated value is a float, representing the max frame rate to * feed the encoder at. * */ public static final String KEY_MAX_FPS_TO_ENCODER = "max-fps-to-encoder"; /** * Instruct the video encoder in "surface-input" mode to limit the gap of * timestamp between any two adjacent frames fed to the encoder to the * specified amount (in micro-second). * * The associated value is a long int. When positive, it represents the max * timestamp gap between two adjacent frames fed to the encoder. When negative, * the absolute value represents a fixed timestamp gap between any two adjacent * frames fed to the encoder. Note that this will also apply even when the * original timestamp goes backward in time. Under normal conditions, such frames * would be dropped and not sent to the encoder. * * The output timestamp will be restored to the original timestamp and will * not be affected. * * This is used in some special scenarios where input frames arrive sparingly * but it's undesirable to allocate more bits to any single frame, or when it's * important to ensure all frames are captured (rather than captured in the * correct order). * */ public static final String KEY_MAX_PTS_GAP_TO_ENCODER = "max-pts-gap-to-encoder"; /** * If specified when configuring a video encoder that's in "surface-input" * mode, it will instruct the encoder to put the surface source in suspended * state when it's connected. No video frames will be accepted until a resume * operation (see {@link MediaCodec#PARAMETER_KEY_SUSPEND}), optionally with * timestamp specified via {@link MediaCodec#PARAMETER_KEY_SUSPEND_TIME}, is * received. * * The value is an integer, with 1 indicating to create with the surface * source suspended, or 0 otherwise. The default value is 0. * * If this key is not set or set to 0, the surface source will accept buffers * as soon as it's connected to the encoder (although they may not be encoded * immediately). This key can be used when the client wants to prepare the * encoder session in advance, but do not want to accept buffers immediately. */ public static final String KEY_CREATE_INPUT_SURFACE_SUSPENDED = "create-input-buffers-suspended"; /** * If specified when configuring a video decoder rendering to a surface, * causes the decoder to output "blank", i.e. black frames to the surface * when stopped to clear out any previously displayed contents. * The associated value is an integer of value 1. */ public static final String KEY_PUSH_BLANK_BUFFERS_ON_STOP = "push-blank-buffers-on-shutdown"; /** * A key describing the duration (in microseconds) of the content. * The associated value is a long. */ public static final String KEY_DURATION = "durationUs"; /** * A key mapping to a value of 1 if the content is AAC audio and * audio frames are prefixed with an ADTS header. * The associated value is an integer (0 or 1). * This key is only supported when _decoding_ content, it cannot * be used to configure an encoder to emit ADTS output. */ public static final String KEY_IS_ADTS = "is-adts"; /** * A key describing the channel composition of audio content. This mask * is composed of bits drawn from channel mask definitions in {@link android.media.AudioFormat}. * The associated value is an integer. */ public static final String KEY_CHANNEL_MASK = "channel-mask"; /** * A key describing the maximum number of channels that can be output by an audio decoder. * By default, the decoder will output the same number of channels as present in the encoded * stream, if supported. Set this value to limit the number of output channels, and use * the downmix information in the stream, if available. *

    Values larger than the number of channels in the content to decode behave like the number * of channels in the content (if applicable), for instance passing 99 for a 5.1 audio stream * behaves like passing 6. *

    This key is only used during decoding. */ public static final String KEY_MAX_OUTPUT_CHANNEL_COUNT = "max-output-channel-count"; /** * A key describing the number of frames to trim from the start of the decoded audio stream. * The associated value is an integer. */ public static final String KEY_ENCODER_DELAY = "encoder-delay"; /** * A key describing the number of frames to trim from the end of the decoded audio stream. * The associated value is an integer. */ public static final String KEY_ENCODER_PADDING = "encoder-padding"; /** * A key describing the AAC profile to be used (AAC audio formats only). * Constants are declared in {@link android.media.MediaCodecInfo.CodecProfileLevel}. */ public static final String KEY_AAC_PROFILE = "aac-profile"; /** * A key describing the AAC SBR mode to be used (AAC audio formats only). * The associated value is an integer and can be set to following values: *

    * Note: If this key is not defined the default SRB mode for the desired AAC profile will * be used. *

    This key is only used during encoding. */ public static final String KEY_AAC_SBR_MODE = "aac-sbr-mode"; /** * A key describing the maximum number of channels that can be output by the AAC decoder. * By default, the decoder will output the same number of channels as present in the encoded * stream, if supported. Set this value to limit the number of output channels, and use * the downmix information in the stream, if available. *

    Values larger than the number of channels in the content to decode behave just * like the actual channel count of the content (e.g. passing 99 for the decoding of 5.1 content * will behave like using 6). *

    This key is only used during decoding. * @deprecated Use the non-AAC-specific key {@link #KEY_MAX_OUTPUT_CHANNEL_COUNT} instead */ public static final String KEY_AAC_MAX_OUTPUT_CHANNEL_COUNT = "aac-max-output-channel_count"; /** * A key describing the Target Reference Level (Target Loudness). *

    For normalizing loudness across program items, a gain is applied to the audio output so * that the output loudness matches the Target Reference Level. The gain is derived as the * difference between the Target Reference Level and the Program Reference Level (Program * Loudness). The latter can be given in the bitstream and indicates the actual loudness value * of the program item.

    *

    The Target Reference Level controls loudness normalization for both MPEG-4 DRC and * MPEG-D DRC. *

    The value is given as an integer value between * 40 and 127, and is calculated as -4 * Target Reference Level in LKFS. * Therefore, it represents the range of -10 to -31.75 LKFS. *

    For MPEG-4 DRC, a value of -1 switches off loudness normalization and DRC processing.

    *

    For MPEG-D DRC, a value of -1 switches off loudness normalization only. For DRC processing * options of MPEG-D DRC, see {@link #KEY_AAC_DRC_EFFECT_TYPE}

    *

    The default value on mobile devices is 64 (-16 LKFS). *

    This key is only used during decoding. */ public static final String KEY_AAC_DRC_TARGET_REFERENCE_LEVEL = "aac-target-ref-level"; /** * A key describing for selecting the DRC effect type for MPEG-D DRC. * The supported values are defined in ISO/IEC 23003-4:2015 and are described as follows: * * * * * * * * * * *
    ValueEffect
    -1Off
    0None
    1Late night
    2Noisy environment
    3Limited playback range
    4Low playback level
    5Dialog enhancement
    6General compression
    *

    The value -1 (Off) disables DRC processing, while loudness normalization may still be * active and dependent on {@link #KEY_AAC_DRC_TARGET_REFERENCE_LEVEL}.
    * The value 0 (None) automatically enables DRC processing if necessary to prevent signal * clipping
    * The value 6 (General compression) can be used for enabling MPEG-D DRC without particular * DRC effect type request.
    * The default DRC effect type is 3 ("Limited playback range") on mobile devices. *

    This key is only used during decoding. */ public static final String KEY_AAC_DRC_EFFECT_TYPE = "aac-drc-effect-type"; /** * A key describing the target reference level that was assumed at the encoder for * calculation of attenuation gains for clipping prevention. *

    If it is known, this information can be provided as an integer value between * 0 and 127, which is calculated as -4 * Encoded Target Level in LKFS. * If the Encoded Target Level is unknown, the value can be set to -1. *

    The default value is -1 (unknown). *

    The value is ignored when heavy compression (see {@link #KEY_AAC_DRC_HEAVY_COMPRESSION}) * or MPEG-D DRC is used. *

    This key is only used during decoding. */ public static final String KEY_AAC_ENCODED_TARGET_LEVEL = "aac-encoded-target-level"; /** * A key describing the boost factor allowing to adapt the dynamics of the output to the * actual listening requirements. This relies on DRC gain sequences that can be transmitted in * the encoded bitstream to be able to reduce the dynamics of the output signal upon request. * This factor enables the user to select how much of the gains are applied. *

    Positive gains (boost) and negative gains (attenuation, see * {@link #KEY_AAC_DRC_ATTENUATION_FACTOR}) can be controlled separately for a better match * to different use-cases. *

    Typically, attenuation gains are sent for loud signal segments, and boost gains are sent * for soft signal segments. If the output is listened to in a noisy environment, for example, * the boost factor is used to enable the positive gains, i.e. to amplify soft signal segments * beyond the noise floor. But for listening late at night, the attenuation * factor is used to enable the negative gains, to prevent loud signal from surprising * the listener. In applications which generally need a low dynamic range, both the boost factor * and the attenuation factor are used in order to enable all DRC gains. *

    In order to prevent clipping, it is also recommended to apply the attenuation gains * in case of a downmix and/or loudness normalization to high target reference levels. *

    Both the boost and the attenuation factor parameters are given as integer values * between 0 and 127, representing the range of the factor of 0 (i.e. don't apply) * to 1 (i.e. fully apply boost/attenuation gains respectively). *

    The default value is 127 (fully apply boost DRC gains). *

    This key is only used during decoding. */ public static final String KEY_AAC_DRC_BOOST_FACTOR = "aac-drc-boost-level"; /** * A key describing the attenuation factor allowing to adapt the dynamics of the output to the * actual listening requirements. * See {@link #KEY_AAC_DRC_BOOST_FACTOR} for a description of the role of this attenuation * factor and the value range. *

    The default value is 127 (fully apply attenuation DRC gains). *

    This key is only used during decoding. */ public static final String KEY_AAC_DRC_ATTENUATION_FACTOR = "aac-drc-cut-level"; /** * A key describing the selection of the heavy compression profile for MPEG-4 DRC. *

    Two separate DRC gain sequences can be transmitted in one bitstream: light compression * and heavy compression. When selecting the application of the heavy compression, one of * the sequences is selected: *

    * Note that heavy compression doesn't offer the features of scaling of DRC gains * (see {@link #KEY_AAC_DRC_BOOST_FACTOR} and {@link #KEY_AAC_DRC_ATTENUATION_FACTOR} for the * boost and attenuation factors), and frequency-selective (multiband) DRC. * Light compression usually contains clipping prevention for stereo downmixing while heavy * compression, if additionally provided in the bitstream, is usually stronger, and contains * clipping prevention for stereo and mono downmixing. *

    The default is 1 (heavy compression). *

    This key is only used during decoding. */ public static final String KEY_AAC_DRC_HEAVY_COMPRESSION = "aac-drc-heavy-compression"; /** * A key to retrieve the output loudness of a decoded bitstream. *

    If loudness normalization is active, the value corresponds to the Target Reference Level * (see {@link #KEY_AAC_DRC_TARGET_REFERENCE_LEVEL}).
    * If loudness normalization is not active, the value corresponds to the loudness metadata * given in the bitstream. *

    The value is retrieved with getInteger() and is given as an integer value between 0 and * 231. It is calculated as -4 * Output Loudness in LKFS. Therefore, it represents the range of * 0 to -57.75 LKFS. *

    A value of -1 indicates that no loudness metadata is present in the bitstream. *

    Loudness metadata can originate from MPEG-4 DRC or MPEG-D DRC. *

    This key is only used during decoding. */ public static final String KEY_AAC_DRC_OUTPUT_LOUDNESS = "aac-drc-output-loudness"; /** * A key describing the album mode for MPEG-D DRC as defined in ISO/IEC 23003-4. *

    The associated value is an integer and can be set to following values: * * * * *
    ValueAlbum Mode
    0disabled
    1enabled
    *

    Disabled album mode leads to application of gain sequences for fading in and out, if * provided in the bitstream. Enabled album mode makes use of dedicated album loudness * information, if provided in the bitstream. *

    The default value is 0 (album mode disabled). *

    This key is only used during decoding. */ public static final String KEY_AAC_DRC_ALBUM_MODE = "aac-drc-album-mode"; /** * A key describing the FLAC compression level to be used (FLAC audio format only). * The associated value is an integer ranging from 0 (fastest, least compression) * to 8 (slowest, most compression). */ public static final String KEY_FLAC_COMPRESSION_LEVEL = "flac-compression-level"; /** * A key describing the MPEG-H stream profile-level indication. * * See ISO_IEC_23008-3;2019 MHADecoderConfigurationRecord mpegh3daProfileLevelIndication. */ public static final String KEY_MPEGH_PROFILE_LEVEL_INDICATION = "mpegh-profile-level-indication"; /** * A key describing the MPEG-H stream compatible sets. * * See FDAmd_2 of ISO_IEC_23008-3;2019 MHAProfileAndLevelCompatibilitySetBox. */ public static final String KEY_MPEGH_COMPATIBLE_SETS = "mpegh-compatible-sets"; /** * A key describing the MPEG-H stream reference channel layout. * * See ISO_IEC_23008-3;2019 MHADecoderConfigurationRecord referenceChannelLayout * and ISO_IEC_23001‐8 ChannelConfiguration value. */ public static final String KEY_MPEGH_REFERENCE_CHANNEL_LAYOUT = "mpegh-reference-channel-layout"; /** * A key describing the encoding complexity. * The associated value is an integer. These values are device and codec specific, * but lower values generally result in faster and/or less power-hungry encoding. * * @see MediaCodecInfo.EncoderCapabilities#getComplexityRange() */ public static final String KEY_COMPLEXITY = "complexity"; /** * A key describing the desired encoding quality. * The associated value is an integer. This key is only supported for encoders * that are configured in constant-quality mode. These values are device and * codec specific, but lower values generally result in more efficient * (smaller-sized) encoding. * * @see MediaCodecInfo.EncoderCapabilities#getQualityRange() */ public static final String KEY_QUALITY = "quality"; /** * A key describing the desired codec priority. *

    * The associated value is an integer. Higher value means lower priority. *

    * Currently, only two levels are supported:
    * 0: realtime priority - meaning that the codec shall support the given * performance configuration (e.g. framerate) at realtime. This should * only be used by media playback, capture, and possibly by realtime * communication scenarios if best effort performance is not suitable.
    * 1: non-realtime priority (best effort). *

    * This is a hint used at codec configuration and resource planning - to understand * the realtime requirements of the application; however, due to the nature of * media components, performance is not guaranteed. * */ public static final String KEY_PRIORITY = "priority"; /** * A key describing the desired operating frame rate for video or sample rate for audio * that the codec will need to operate at. *

    * The associated value is an integer or a float representing frames-per-second or * samples-per-second *

    * This is used for cases like high-speed/slow-motion video capture, where the video encoder * format contains the target playback rate (e.g. 30fps), but the component must be able to * handle the high operating capture rate (e.g. 240fps). *

    * This rate will be used by codec for resource planning and setting the operating points. * */ public static final String KEY_OPERATING_RATE = "operating-rate"; /** * A key describing the desired profile to be used by an encoder. *

    * The associated value is an integer. * Constants are declared in {@link MediaCodecInfo.CodecProfileLevel}. * This key is used as a hint, and is only supported for codecs * that specify a profile. When configuring profile, encoder configuration * may fail if other parameters are not compatible with the desired * profile or if the desired profile is not supported, but it may also * fail silently (where the encoder ends up using a different, compatible profile.) *

    * It is recommended that the profile is set for all encoders. For more information, see * the Encoder Profiles section of the {@link MediaCodec} API reference. *

    * Note: Codecs are free to use all the available * coding tools at the specified profile, but may ultimately choose to not do so. *

    * Note: When configuring video encoders, profile (if set) must be * set together with {@link #KEY_LEVEL level}. * * @see MediaCodecInfo.CodecCapabilities#profileLevels */ public static final String KEY_PROFILE = "profile"; /** * A key describing the desired profile to be used by an encoder. *

    * The associated value is an integer. * Constants are declared in {@link MediaCodecInfo.CodecProfileLevel}. * This key is used as a further hint when specifying a desired profile, * and is only supported for codecs that specify a level. *

    * This key is ignored if the {@link #KEY_PROFILE profile} is not specified. * Otherwise, the value should be a level compatible with the configured encoding * parameters. *

    * Note: This key cannot be used to constrain the encoder's * output to a maximum encoding level. Encoders are free to target a different * level if the configured encoding parameters dictate it. Nevertheless, * encoders shall use (and encode) a level sufficient to decode the generated * bitstream, though they may exceed the (Video) Buffering Verifier limits for * that encoded level. * * @see MediaCodecInfo.CodecCapabilities#profileLevels */ public static final String KEY_LEVEL = "level"; /** * An optional key describing the desired encoder latency in frames. This is an optional * parameter that applies only to video encoders. If encoder supports it, it should ouput * at least one output frame after being queued the specified number of frames. This key * is ignored if the video encoder does not support the latency feature. Use the output * format to verify that this feature was enabled and the actual value used by the encoder. *

    * If the key is not specified, the default latency will be implenmentation specific. * The associated value is an integer. */ public static final String KEY_LATENCY = "latency"; /** * An optional key describing the maximum number of non-display-order coded frames. * This is an optional parameter that applies only to video encoders. Application should * check the value for this key in the output format to see if codec will produce * non-display-order coded frames. If encoder supports it, the output frames' order will be * different from the display order and each frame's display order could be retrived from * {@link MediaCodec.BufferInfo#presentationTimeUs}. Before API level 27, application may * receive non-display-order coded frames even though the application did not request it. * Note: Application should not rearrange the frames to display order before feeding them * to {@link MediaMuxer#writeSampleData}. *

    * The default value is 0. */ public static final String KEY_OUTPUT_REORDER_DEPTH = "output-reorder-depth"; /** * A key describing the desired clockwise rotation on an output surface. * This key is only used when the codec is configured using an output surface. * The associated value is an integer, representing degrees. Supported values * are 0, 90, 180 or 270. This is an optional field; if not specified, rotation * defaults to 0. * * @see MediaCodecInfo.CodecCapabilities#profileLevels */ public static final String KEY_ROTATION = "rotation-degrees"; /** * A key describing the desired bitrate mode to be used by an encoder. * Constants are declared in {@link MediaCodecInfo.EncoderCapabilities}. * * @see MediaCodecInfo.EncoderCapabilities#isBitrateModeSupported(int) */ public static final String KEY_BITRATE_MODE = "bitrate-mode"; /** * A key describing the maximum Quantization Parameter allowed for encoding video. * This key applies to all three video picture types (I, P, and B). * The value is used directly for picture type I; a per-mime formula is used * to calculate the value for the remaining picture types. * * This calculation can be avoided by directly specifying values for each picture type * using the type-specific keys {@link #KEY_VIDEO_QP_I_MAX}, {@link #KEY_VIDEO_QP_P_MAX}, * and {@link #KEY_VIDEO_QP_B_MAX}. * * The associated value is an integer. */ public static final String KEY_VIDEO_QP_MAX = "video-qp-max"; /** * A key describing the minimum Quantization Parameter allowed for encoding video. * This key applies to all three video frame types (I, P, and B). * The value is used directly for picture type I; a per-mime formula is used * to calculate the value for the remaining picture types. * * This calculation can be avoided by directly specifying values for each picture type * using the type-specific keys {@link #KEY_VIDEO_QP_I_MIN}, {@link #KEY_VIDEO_QP_P_MIN}, * and {@link #KEY_VIDEO_QP_B_MIN}. * * The associated value is an integer. */ public static final String KEY_VIDEO_QP_MIN = "video-qp-min"; /** * A key describing the maximum Quantization Parameter allowed for encoding video. * This value applies to video I-frames. * * The associated value is an integer. */ public static final String KEY_VIDEO_QP_I_MAX = "video-qp-i-max"; /** * A key describing the minimum Quantization Parameter allowed for encoding video. * This value applies to video I-frames. * * The associated value is an integer. */ public static final String KEY_VIDEO_QP_I_MIN = "video-qp-i-min"; /** * A key describing the maximum Quantization Parameter allowed for encoding video. * This value applies to video P-frames. * * The associated value is an integer. */ public static final String KEY_VIDEO_QP_P_MAX = "video-qp-p-max"; /** * A key describing the minimum Quantization Parameter allowed for encoding video. * This value applies to video P-frames. * * The associated value is an integer. */ public static final String KEY_VIDEO_QP_P_MIN = "video-qp-p-min"; /** * A key describing the maximum Quantization Parameter allowed for encoding video. * This value applies to video B-frames. * * The associated value is an integer. */ public static final String KEY_VIDEO_QP_B_MAX = "video-qp-b-max"; /** * A key describing the minimum Quantization Parameter allowed for encoding video. * This value applies to video B-frames. * * The associated value is an integer. */ public static final String KEY_VIDEO_QP_B_MIN = "video-qp-b-min"; /** * A key describing the level of encoding statistics information emitted from video encoder. * * The associated value is an integer. */ public static final String KEY_VIDEO_ENCODING_STATISTICS_LEVEL = "video-encoding-statistics-level"; /** * Encoding Statistics Level None. * Encoder generates no information about Encoding statistics. */ public static final int VIDEO_ENCODING_STATISTICS_LEVEL_NONE = 0; /** * Encoding Statistics Level 1. * Encoder generates {@link MediaFormat#KEY_PICTURE_TYPE} and * {@link MediaFormat#KEY_VIDEO_QP_AVERAGE} for each frame. */ public static final int VIDEO_ENCODING_STATISTICS_LEVEL_1 = 1; /** @hide */ @IntDef({ VIDEO_ENCODING_STATISTICS_LEVEL_NONE, VIDEO_ENCODING_STATISTICS_LEVEL_1, }) @Retention(RetentionPolicy.SOURCE) public @interface VideoEncodingStatisticsLevel {} /** * A key describing the per-frame average block QP (Quantization Parameter). * This is a part of a video 'Encoding Statistics' export feature. * This value is emitted from video encoder for a video frame. * The average value is rounded to the nearest integer value. * * The associated value is an integer. */ public static final String KEY_VIDEO_QP_AVERAGE = "video-qp-average"; /** * A key describing the picture type of the encoded frame. * This is a part of a video 'Encoding Statistics' export feature. * This value is emitted from video encoder for a video frame. * * The associated value is an integer. */ public static final String KEY_PICTURE_TYPE = "picture-type"; /** Picture Type is unknown. */ public static final int PICTURE_TYPE_UNKNOWN = 0; /** Picture Type is I Frame. */ public static final int PICTURE_TYPE_I = 1; /** Picture Type is P Frame. */ public static final int PICTURE_TYPE_P = 2; /** Picture Type is B Frame. */ public static final int PICTURE_TYPE_B = 3; /** @hide */ @IntDef({ PICTURE_TYPE_UNKNOWN, PICTURE_TYPE_I, PICTURE_TYPE_P, PICTURE_TYPE_B, }) @Retention(RetentionPolicy.SOURCE) public @interface PictureType {} /** * A key describing the audio session ID of the AudioTrack associated * to a tunneled video codec. * The associated value is an integer. * * @see MediaCodecInfo.CodecCapabilities#FEATURE_TunneledPlayback */ public static final String KEY_AUDIO_SESSION_ID = "audio-session-id"; /** * A key describing the audio hardware sync ID of the AudioTrack associated * to a tunneled video codec. The associated value is an integer. * * @hide * * @see MediaCodecInfo.CodecCapabilities#FEATURE_TunneledPlayback * @see AudioManager#getAudioHwSyncForSession */ public static final String KEY_AUDIO_HW_SYNC = "audio-hw-sync"; /** * A key for boolean AUTOSELECT behavior for the track. Tracks with AUTOSELECT=true * are considered when automatically selecting a track without specific user * choice, based on the current locale. * This is currently only used for subtitle tracks, when the user selected * 'Default' for the captioning locale. * The associated value is an integer, where non-0 means TRUE. This is an optional * field; if not specified, AUTOSELECT defaults to TRUE. */ public static final String KEY_IS_AUTOSELECT = "is-autoselect"; /** * A key for boolean DEFAULT behavior for the track. The track with DEFAULT=true is * selected in the absence of a specific user choice. * This is currently used in two scenarios: * 1) for subtitle tracks, when the user selected 'Default' for the captioning locale. * 2) for a {@link #MIMETYPE_IMAGE_ANDROID_HEIC} / {@link #MIMETYPE_IMAGE_AVIF} track, * indicating the image is the primary item in the file. * * The associated value is an integer, where non-0 means TRUE. This is an optional * field; if not specified, DEFAULT is considered to be FALSE. */ public static final String KEY_IS_DEFAULT = "is-default"; /** * A key for the FORCED field for subtitle tracks. True if it is a * forced subtitle track. Forced subtitle tracks are essential for the * content and are shown even when the user turns off Captions. They * are used for example to translate foreign/alien dialogs or signs. * The associated value is an integer, where non-0 means TRUE. This is an * optional field; if not specified, FORCED defaults to FALSE. */ public static final String KEY_IS_FORCED_SUBTITLE = "is-forced-subtitle"; /** * A key describing the number of haptic channels in an audio format. * The associated value is an integer. */ public static final String KEY_HAPTIC_CHANNEL_COUNT = "haptic-channel-count"; /** @hide */ public static final String KEY_IS_TIMED_TEXT = "is-timed-text"; // The following color aspect values must be in sync with the ones in HardwareAPI.h. /** * An optional key describing the color primaries, white point and * luminance factors for video content. * * The associated value is an integer: 0 if unspecified, or one of the * COLOR_STANDARD_ values. */ public static final String KEY_COLOR_STANDARD = "color-standard"; /** BT.709 color chromacity coordinates with KR = 0.2126, KB = 0.0722. */ public static final int COLOR_STANDARD_BT709 = 1; /** BT.601 625 color chromacity coordinates with KR = 0.299, KB = 0.114. */ public static final int COLOR_STANDARD_BT601_PAL = 2; /** BT.601 525 color chromacity coordinates with KR = 0.299, KB = 0.114. */ public static final int COLOR_STANDARD_BT601_NTSC = 4; /** BT.2020 color chromacity coordinates with KR = 0.2627, KB = 0.0593. */ public static final int COLOR_STANDARD_BT2020 = 6; /** @hide */ @IntDef({ COLOR_STANDARD_BT709, COLOR_STANDARD_BT601_PAL, COLOR_STANDARD_BT601_NTSC, COLOR_STANDARD_BT2020, }) @Retention(RetentionPolicy.SOURCE) public @interface ColorStandard {} /** * An optional key describing the opto-electronic transfer function used * for the video content. * * The associated value is an integer: 0 if unspecified, or one of the * COLOR_TRANSFER_ values. */ public static final String KEY_COLOR_TRANSFER = "color-transfer"; /** Linear transfer characteristic curve. */ public static final int COLOR_TRANSFER_LINEAR = 1; /** SMPTE 170M transfer characteristic curve used by BT.601/BT.709/BT.2020. This is the curve * used by most non-HDR video content. */ public static final int COLOR_TRANSFER_SDR_VIDEO = 3; /** SMPTE ST 2084 transfer function. This is used by some HDR video content. */ public static final int COLOR_TRANSFER_ST2084 = 6; /** ARIB STD-B67 hybrid-log-gamma transfer function. This is used by some HDR video content. */ public static final int COLOR_TRANSFER_HLG = 7; /** @hide */ @IntDef({ COLOR_TRANSFER_LINEAR, COLOR_TRANSFER_SDR_VIDEO, COLOR_TRANSFER_ST2084, COLOR_TRANSFER_HLG, }) @Retention(RetentionPolicy.SOURCE) public @interface ColorTransfer {} /** * An optional key describing the range of the component values of the video content. * * The associated value is an integer: 0 if unspecified, or one of the * COLOR_RANGE_ values. */ public static final String KEY_COLOR_RANGE = "color-range"; /** Limited range. Y component values range from 16 to 235 for 8-bit content. * Cr, Cy values range from 16 to 240 for 8-bit content. * This is the default for video content. */ public static final int COLOR_RANGE_LIMITED = 2; /** Full range. Y, Cr and Cb component values range from 0 to 255 for 8-bit content. */ public static final int COLOR_RANGE_FULL = 1; /** @hide */ @IntDef({ COLOR_RANGE_LIMITED, COLOR_RANGE_FULL, }) @Retention(RetentionPolicy.SOURCE) public @interface ColorRange {} /** * An optional key describing the static metadata of HDR (high-dynamic-range) video content. * * The associated value is a ByteBuffer. This buffer contains the raw contents of the * Static Metadata Descriptor (including the descriptor ID) of an HDMI Dynamic Range and * Mastering InfoFrame as defined by CTA-861.3. This key must be provided to video decoders * for HDR video content unless this information is contained in the bitstream and the video * decoder supports an HDR-capable profile. This key must be provided to video encoders for * HDR video content. */ public static final String KEY_HDR_STATIC_INFO = "hdr-static-info"; /** * An optional key describing the HDR10+ metadata of the video content. * * The associated value is a ByteBuffer containing HDR10+ metadata conforming to the * user_data_registered_itu_t_t35() syntax of SEI message for ST 2094-40. This key will * be present on: *

    * - The formats of output buffers of a decoder configured for HDR10+ profiles (such as * {@link MediaCodecInfo.CodecProfileLevel#VP9Profile2HDR10Plus}, {@link * MediaCodecInfo.CodecProfileLevel#VP9Profile3HDR10Plus} or {@link * MediaCodecInfo.CodecProfileLevel#HEVCProfileMain10HDR10Plus}), or *

    * - The formats of output buffers of an encoder configured for an HDR10+ profiles that * uses out-of-band metadata (such as {@link * MediaCodecInfo.CodecProfileLevel#VP9Profile2HDR10Plus} or {@link * MediaCodecInfo.CodecProfileLevel#VP9Profile3HDR10Plus}). * * @see MediaCodec#PARAMETER_KEY_HDR10_PLUS_INFO */ public static final String KEY_HDR10_PLUS_INFO = "hdr10-plus-info"; /** * An optional key describing the opto-electronic transfer function * requested for the output video content. * * The associated value is an integer: 0 if unspecified, or one of the * COLOR_TRANSFER_ values. When unspecified the component will not touch the * video content; otherwise the component will tone-map the raw video frame * to match the requested transfer function. * * After configure, component's input format will contain this key to note * whether the request is supported or not. If the value in the input format * is the same as the requested value, the request is supported. The value * is set to 0 if unsupported. */ public static final String KEY_COLOR_TRANSFER_REQUEST = "color-transfer-request"; /** * A key describing a unique ID for the content of a media track. * *

    This key is used by {@link MediaExtractor}. Some extractors provide multiple encodings * of the same track (e.g. float audio tracks for FLAC and WAV may be expressed as two * tracks via MediaExtractor: a normal PCM track for backward compatibility, and a float PCM * track for added fidelity. Similarly, Dolby Vision extractor may provide a baseline SDR * version of a DV track.) This key can be used to identify which MediaExtractor tracks refer * to the same underlying content. *

    * * The associated value is an integer. */ public static final String KEY_TRACK_ID = "track-id"; /** * A key describing the system id of the conditional access system used to scramble * a media track. *

    * This key is set by {@link MediaExtractor} if the track is scrambled with a conditional * access system, regardless of the presence of a valid {@link MediaCas} object. *

    * The associated value is an integer. * @hide */ public static final String KEY_CA_SYSTEM_ID = "ca-system-id"; /** * A key describing the {@link MediaCas.Session} object associated with a media track. *

    * This key is set by {@link MediaExtractor} if the track is scrambled with a conditional * access system, after it receives a valid {@link MediaCas} object. *

    * The associated value is a ByteBuffer. * @hide */ public static final String KEY_CA_SESSION_ID = "ca-session-id"; /** * A key describing the private data in the CA_descriptor associated with a media track. *

    * This key is set by {@link MediaExtractor} if the track is scrambled with a conditional * access system, before it receives a valid {@link MediaCas} object. *

    * The associated value is a ByteBuffer. * @hide */ public static final String KEY_CA_PRIVATE_DATA = "ca-private-data"; /** * A key describing the maximum number of B frames between I or P frames, * to be used by a video encoder. * The associated value is an integer. The default value is 0, which means * that no B frames are allowed. Note that non-zero value does not guarantee * B frames; it's up to the encoder to decide. */ public static final String KEY_MAX_B_FRAMES = "max-bframes"; /** * A key for applications to opt out of allowing * a Surface to discard undisplayed/unconsumed frames * as means to catch up after falling behind. * This value is an integer. * The value 0 indicates the surface is not allowed to drop frames. * The value 1 indicates the surface is allowed to drop frames. * * {@link MediaCodec} describes the semantics. */ public static final String KEY_ALLOW_FRAME_DROP = "allow-frame-drop"; /** * A key describing the desired codec importance for the application. *

    * The associated value is a positive integer including zero. * Higher value means lesser importance. *

    * The resource manager may use the codec importance, along with other factors * when reclaiming codecs from an application. * The specifics of reclaim policy is device dependent, but specifying the codec importance, * will allow the resource manager to prioritize reclaiming less important codecs * (assigned higher values) from the (reclaim) requesting application first. * So, the codec importance is only relevant within the context of that application. *

    * The codec importance can be set: *

    * Any change/update in codec importance is guaranteed upon the completion of the function call * that sets the codec importance. So, in case of concurrent codec operations, * make sure to wait for the change in codec importance, before using another codec. * Note that unless specified, by default the codecs will have highest importance (of value 0). * */ @FlaggedApi(FLAG_CODEC_IMPORTANCE) public static final String KEY_IMPORTANCE = "importance"; /** @hide */ @IntDef(flag = true, prefix = {"FLAG_SECURITY_MODEL_"}, value = { FLAG_SECURITY_MODEL_SANDBOXED, FLAG_SECURITY_MODEL_MEMORY_SAFE, FLAG_SECURITY_MODEL_TRUSTED_CONTENT_ONLY, }) @Retention(RetentionPolicy.SOURCE) public @interface SecurityModelFlag {} /** * Flag for {@link MediaCodecInfo#SECURITY_MODEL_SANDBOXED}. */ @FlaggedApi(FLAG_IN_PROCESS_SW_AUDIO_CODEC) public static final int FLAG_SECURITY_MODEL_SANDBOXED = (1 << MediaCodecInfo.SECURITY_MODEL_SANDBOXED); /** * Flag for {@link MediaCodecInfo#SECURITY_MODEL_MEMORY_SAFE}. */ @FlaggedApi(FLAG_IN_PROCESS_SW_AUDIO_CODEC) public static final int FLAG_SECURITY_MODEL_MEMORY_SAFE = (1 << MediaCodecInfo.SECURITY_MODEL_MEMORY_SAFE); /** * Flag for {@link MediaCodecInfo#SECURITY_MODEL_TRUSTED_CONTENT_ONLY}. */ @FlaggedApi(FLAG_IN_PROCESS_SW_AUDIO_CODEC) public static final int FLAG_SECURITY_MODEL_TRUSTED_CONTENT_ONLY = (1 << MediaCodecInfo.SECURITY_MODEL_TRUSTED_CONTENT_ONLY); /** * A key describing the requested security model as flags. *

    * The associated value is a flag of the following values: * {@link FLAG_SECURITY_MODEL_SANDBOXED}, * {@link FLAG_SECURITY_MODEL_MEMORY_SAFE}, * {@link FLAG_SECURITY_MODEL_TRUSTED_CONTENT_ONLY}. The default value is * {@link FLAG_SECURITY_MODEL_SANDBOXED}. *

    * When passed to {@link MediaCodecList#findDecoderForFormat} or * {@link MediaCodecList#findEncoderForFormat}, MediaCodecList filters * the security model of the codecs according to this flag value. *

    * When passed to {@link MediaCodec#configure}, MediaCodec verifies * the security model matches the flag value passed, and throws * {@link java.lang.IllegalArgumentException} if the model does not match. *

    * @see MediaCodecInfo#getSecurityModel * @see MediaCodecList#findDecoderForFormat * @see MediaCodecList#findEncoderForFormat */ @FlaggedApi(FLAG_IN_PROCESS_SW_AUDIO_CODEC) public static final String KEY_SECURITY_MODEL = "security-model"; /** * QpOffsetRect constitutes the metadata required for encoding a region of interest in an * image or a video frame. The region of interest is represented by a rectangle. The four * integer coordinates of the rectangle are stored in fields left, top, right, bottom. * Note that the right and bottom coordinates are exclusive. * This is paired with a suggestive qp offset information that is to be used during encoding * of the blocks belonging to the to the box. */ @FlaggedApi(FLAG_REGION_OF_INTEREST) public static final class QpOffsetRect { private Rect mContour; private int mQpOffset; /** * Create a new region of interest with the specified coordinates and qpOffset. Note: no * range checking is performed, so the caller must ensure that left >= 0, left <= right, * top >= 0 and top <= bottom. Note that the right and bottom coordinates are exclusive. * * @param contour Rectangle specifying the region of interest * @param qpOffset qpOffset to be used for the blocks in the specified rectangle */ public QpOffsetRect(@NonNull Rect contour, int qpOffset) { mContour = contour; mQpOffset = qpOffset; } /** * Update the region of interest information with the specified coordinates and qpOffset * * @param contour Rectangle specifying the region of interest * @param qpOffset qpOffset to be used for the blocks in the specified rectangle */ public void set(@NonNull Rect contour, int qpOffset) { mContour = contour; mQpOffset = qpOffset; } /** * @return Return a string representation of qpOffsetRect in a compact form. * Helper function to insert key {@link #PARAMETER_KEY_QP_OFFSET_RECTS} in MediaFormat */ @NonNull public String flattenToString() { return TextUtils.formatSimple("%d,%d-%d,%d=%d;", mContour.top, mContour.left, mContour.bottom, mContour.right, mQpOffset); } /** * @return Return a string representation of qpOffsetRect in a compact form. * Helper function to insert key {@link #PARAMETER_KEY_QP_OFFSET_RECTS} in MediaFormat */ @NonNull public static String flattenToString(@NonNull List qpOffsetRects) { StringBuilder builder = new StringBuilder(); for (QpOffsetRect qpOffsetRect : qpOffsetRects) { builder.append(qpOffsetRect.flattenToString()); } return builder.toString(); } } /* package private */ MediaFormat(@NonNull Map map) { mMap = map; } /** * Creates an empty MediaFormat */ public MediaFormat() { mMap = new HashMap(); } @UnsupportedAppUsage /* package private */ Map getMap() { return mMap; } /** * Returns true iff a key of the given name exists in the format. */ public final boolean containsKey(@NonNull String name) { return mMap.containsKey(name); } /** * Returns true iff a feature of the given name exists in the format. */ public final boolean containsFeature(@NonNull String name) { return mMap.containsKey(KEY_FEATURE_ + name); } public static final int TYPE_NULL = 0; public static final int TYPE_INTEGER = 1; public static final int TYPE_LONG = 2; public static final int TYPE_FLOAT = 3; public static final int TYPE_STRING = 4; public static final int TYPE_BYTE_BUFFER = 5; /** @hide */ @IntDef({ TYPE_NULL, TYPE_INTEGER, TYPE_LONG, TYPE_FLOAT, TYPE_STRING, TYPE_BYTE_BUFFER }) @Retention(RetentionPolicy.SOURCE) public @interface Type {} /** * Returns the value type for a key. If the key does not exist, it returns TYPE_NULL. */ public final @Type int getValueTypeForKey(@NonNull String name) { Object value = mMap.get(name); if (value == null) { return TYPE_NULL; } else if (value instanceof Integer) { return TYPE_INTEGER; } else if (value instanceof Long) { return TYPE_LONG; } else if (value instanceof Float) { return TYPE_FLOAT; } else if (value instanceof String) { return TYPE_STRING; } else if (value instanceof ByteBuffer) { return TYPE_BYTE_BUFFER; } throw new RuntimeException("invalid value for key"); } /** * A key prefix used together with a {@link MediaCodecInfo.CodecCapabilities} * feature name describing a required or optional feature for a codec capabilities * query. * The associated value is an integer, where non-0 value means the feature is * requested to be present, while 0 value means the feature is requested to be not * present. * @see MediaCodecList#findDecoderForFormat * @see MediaCodecList#findEncoderForFormat * @see MediaCodecInfo.CodecCapabilities#isFormatSupported * * @hide */ public static final String KEY_FEATURE_ = "feature-"; /** * Returns the value of a numeric key. This is provided as a convenience method for keys * that may take multiple numeric types, such as {@link #KEY_FRAME_RATE}, or {@link * #KEY_I_FRAME_INTERVAL}. * * @return null if the key does not exist or the stored value for the key is null * @throws ClassCastException if the stored value for the key is ByteBuffer or String */ public final @Nullable Number getNumber(@NonNull String name) { return (Number) mMap.get(name); } /** * Returns the value of a numeric key, or the default value if the key is missing. * * @return defaultValue if the key does not exist or the stored value for the key is null * @throws ClassCastException if the stored value for the key is ByteBuffer or String */ public final @NonNull Number getNumber(@NonNull String name, @NonNull Number defaultValue) { Number ret = getNumber(name); return ret == null ? defaultValue : ret; } /** * Returns the value of an integer key. * * @throws NullPointerException if the key does not exist or the stored value for the key is * null * @throws ClassCastException if the stored value for the key is long, float, ByteBuffer or * String */ public final int getInteger(@NonNull String name) { return (int) mMap.get(name); } /** * Returns the value of an integer key, or the default value if the key is missing. * * @return defaultValue if the key does not exist or the stored value for the key is null * @throws ClassCastException if the stored value for the key is long, float, ByteBuffer or * String */ public final int getInteger(@NonNull String name, int defaultValue) { try { return getInteger(name); } catch (NullPointerException e) { /* no such field or field is null */ return defaultValue; } } /** * Returns the value of a long key. * * @throws NullPointerException if the key does not exist or the stored value for the key is * null * @throws ClassCastException if the stored value for the key is int, float, ByteBuffer or * String */ public final long getLong(@NonNull String name) { return (long) mMap.get(name); } /** * Returns the value of a long key, or the default value if the key is missing. * * @return defaultValue if the key does not exist or the stored value for the key is null * @throws ClassCastException if the stored value for the key is int, float, ByteBuffer or * String */ public final long getLong(@NonNull String name, long defaultValue) { try { return getLong(name); } catch (NullPointerException e) { /* no such field or field is null */ return defaultValue; } } /** * Returns the value of a float key. * * @throws NullPointerException if the key does not exist or the stored value for the key is * null * @throws ClassCastException if the stored value for the key is int, long, ByteBuffer or * String */ public final float getFloat(@NonNull String name) { return (float) mMap.get(name); } /** * Returns the value of a float key, or the default value if the key is missing. * * @return defaultValue if the key does not exist or the stored value for the key is null * @throws ClassCastException if the stored value for the key is int, long, ByteBuffer or * String */ public final float getFloat(@NonNull String name, float defaultValue) { Object value = mMap.get(name); return value != null ? (float) value : defaultValue; } /** * Returns the value of a string key. * * @return null if the key does not exist or the stored value for the key is null * @throws ClassCastException if the stored value for the key is int, long, float or ByteBuffer */ public final @Nullable String getString(@NonNull String name) { return (String)mMap.get(name); } /** * Returns the value of a string key, or the default value if the key is missing. * * @return defaultValue if the key does not exist or the stored value for the key is null * @throws ClassCastException if the stored value for the key is int, long, float or ByteBuffer */ public final @NonNull String getString(@NonNull String name, @NonNull String defaultValue) { String ret = getString(name); return ret == null ? defaultValue : ret; } /** * Returns the value of a ByteBuffer key. * * @return null if the key does not exist or the stored value for the key is null * @throws ClassCastException if the stored value for the key is int, long, float or String */ public final @Nullable ByteBuffer getByteBuffer(@NonNull String name) { return (ByteBuffer)mMap.get(name); } /** * Returns the value of a ByteBuffer key, or the default value if the key is missing. * * @return defaultValue if the key does not exist or the stored value for the key is null * @throws ClassCastException if the stored value for the key is int, long, float or String */ public final @NonNull ByteBuffer getByteBuffer( @NonNull String name, @NonNull ByteBuffer defaultValue) { ByteBuffer ret = getByteBuffer(name); return ret == null ? defaultValue : ret; } /** * Returns whether a feature is to be enabled ({@code true}) or disabled * ({@code false}). * * @param feature the name of a {@link MediaCodecInfo.CodecCapabilities} feature. * * @throws IllegalArgumentException if the feature was neither set to be enabled * nor to be disabled. */ public boolean getFeatureEnabled(@NonNull String feature) { Integer enabled = (Integer)mMap.get(KEY_FEATURE_ + feature); if (enabled == null) { throw new IllegalArgumentException("feature is not specified"); } return enabled != 0; } /** * Sets the value of an integer key. */ public final void setInteger(@NonNull String name, int value) { mMap.put(name, value); } /** * Sets the value of a long key. */ public final void setLong(@NonNull String name, long value) { mMap.put(name, value); } /** * Sets the value of a float key. */ public final void setFloat(@NonNull String name, float value) { mMap.put(name, value); } /** * Sets the value of a string key. *

    * If value is {@code null}, it sets a null value that behaves similarly to a missing key. * This could be used prior to API level {@link android os.Build.VERSION_CODES#Q} to effectively * remove a key. */ public final void setString(@NonNull String name, @Nullable String value) { mMap.put(name, value); } /** * Sets the value of a ByteBuffer key. *

    * If value is {@code null}, it sets a null value that behaves similarly to a missing key. * This could be used prior to API level {@link android os.Build.VERSION_CODES#Q} to effectively * remove a key. */ public final void setByteBuffer(@NonNull String name, @Nullable ByteBuffer bytes) { mMap.put(name, bytes); } /** * Removes a value of a given key if present. Has no effect if the key is not present. */ public final void removeKey(@NonNull String name) { // exclude feature mappings if (!name.startsWith(KEY_FEATURE_)) { mMap.remove(name); } } /** * Removes a given feature setting if present. Has no effect if the feature setting is not * present. */ public final void removeFeature(@NonNull String name) { mMap.remove(KEY_FEATURE_ + name); } /** * A Partial set view for a portion of the keys in a MediaFormat object. * * This class is needed as we want to return a portion of the actual format keys in getKeys() * and another portion of the keys in getFeatures(), and still allow the view properties. */ private abstract class FilteredMappedKeySet extends AbstractSet { private Set mKeys; // Returns true if this set should include this key abstract protected boolean keepKey(String key); // Maps a key from the underlying key set into its new value in this key set abstract protected String mapKeyToItem(String key); // Maps a key from this key set into its original value in the underlying key set abstract protected String mapItemToKey(String item); public FilteredMappedKeySet() { mKeys = mMap.keySet(); } // speed up contains and remove from abstract implementation (that would iterate // over each element) @Override public boolean contains(Object o) { if (o instanceof String) { String key = mapItemToKey((String)o); return keepKey(key) && mKeys.contains(key); } return false; } @Override public boolean remove(Object o) { if (o instanceof String) { String key = mapItemToKey((String)o); if (keepKey(key) && mKeys.remove(key)) { mMap.remove(key); return true; } } return false; } private class KeyIterator implements Iterator { Iterator mIterator; String mLast; public KeyIterator() { // We must create a copy of the filtered stream, as remove operation has to modify // the underlying data structure (mMap), so the iterator's operation is undefined. // Use a list as it is likely less memory consuming than the other alternative: set. mIterator = mKeys.stream().filter(k -> keepKey(k)).collect(Collectors.toList()).iterator(); } @Override public boolean hasNext() { return mIterator.hasNext(); } @Override public String next() { mLast = mIterator.next(); return mapKeyToItem(mLast); } @Override public void remove() { mIterator.remove(); mMap.remove(mLast); } } @Override public Iterator iterator() { return new KeyIterator(); } @Override public int size() { return (int) mKeys.stream().filter(this::keepKey).count(); } } /** * A Partial set view for a portion of the keys in a MediaFormat object for keys that * don't start with a prefix, such as "feature-" */ private class UnprefixedKeySet extends FilteredMappedKeySet { private String mPrefix; public UnprefixedKeySet(String prefix) { super(); mPrefix = prefix; } protected boolean keepKey(String key) { return !key.startsWith(mPrefix); } protected String mapKeyToItem(String key) { return key; } protected String mapItemToKey(String item) { return item; } } /** * A Partial set view for a portion of the keys in a MediaFormat object for keys that * start with a prefix, such as "feature-", with the prefix removed */ private class PrefixedKeySetWithPrefixRemoved extends FilteredMappedKeySet { private String mPrefix; private int mPrefixLength; public PrefixedKeySetWithPrefixRemoved(String prefix) { super(); mPrefix = prefix; mPrefixLength = prefix.length(); } protected boolean keepKey(String key) { return key.startsWith(mPrefix); } protected String mapKeyToItem(String key) { return key.substring(mPrefixLength); } protected String mapItemToKey(String item) { return mPrefix + item; } } /** * Returns a {@link java.util.Set Set} view of the keys contained in this MediaFormat. * * The set is backed by the MediaFormat object, so changes to the format are reflected in the * set, and vice-versa. If the format is modified while an iteration over the set is in progress * (except through the iterator's own remove operation), the results of the iteration are * undefined. The set supports element removal, which removes the corresponding mapping from the * format, via the Iterator.remove, Set.remove, removeAll, retainAll, and clear operations. * It does not support the add or addAll operations. */ public final @NonNull java.util.Set getKeys() { return new UnprefixedKeySet(KEY_FEATURE_); } /** * Returns a {@link java.util.Set Set} view of the features contained in this MediaFormat. * * The set is backed by the MediaFormat object, so changes to the format are reflected in the * set, and vice-versa. If the format is modified while an iteration over the set is in progress * (except through the iterator's own remove operation), the results of the iteration are * undefined. The set supports element removal, which removes the corresponding mapping from the * format, via the Iterator.remove, Set.remove, removeAll, retainAll, and clear operations. * It does not support the add or addAll operations. */ public final @NonNull java.util.Set getFeatures() { return new PrefixedKeySetWithPrefixRemoved(KEY_FEATURE_); } /** * Create a copy of a media format object. */ public MediaFormat(@NonNull MediaFormat other) { this(); mMap.putAll(other.mMap); } /** * Sets whether a feature is to be enabled ({@code true}) or disabled * ({@code false}). * * If {@code enabled} is {@code true}, the feature is requested to be present. * Otherwise, the feature is requested to be not present. * * @param feature the name of a {@link MediaCodecInfo.CodecCapabilities} feature. * * @see MediaCodecList#findDecoderForFormat * @see MediaCodecList#findEncoderForFormat * @see MediaCodecInfo.CodecCapabilities#isFormatSupported */ public void setFeatureEnabled(@NonNull String feature, boolean enabled) { setInteger(KEY_FEATURE_ + feature, enabled ? 1 : 0); } /** * Creates a minimal audio format. * @param mime The mime type of the content. * @param sampleRate The sampling rate of the content. * @param channelCount The number of audio channels in the content. */ public static final @NonNull MediaFormat createAudioFormat( @NonNull String mime, int sampleRate, int channelCount) { MediaFormat format = new MediaFormat(); format.setString(KEY_MIME, mime); format.setInteger(KEY_SAMPLE_RATE, sampleRate); format.setInteger(KEY_CHANNEL_COUNT, channelCount); return format; } /** * Creates a minimal subtitle format. * @param mime The mime type of the content. * @param language The language of the content, using either ISO 639-1 or 639-2/T * codes. Specify null or "und" if language information is only included * in the content. (This will also work if there are multiple language * tracks in the content.) */ public static final @NonNull MediaFormat createSubtitleFormat( @NonNull String mime, String language) { MediaFormat format = new MediaFormat(); format.setString(KEY_MIME, mime); format.setString(KEY_LANGUAGE, language); return format; } /** * Creates a minimal video format. * @param mime The mime type of the content. * @param width The width of the content (in pixels) * @param height The height of the content (in pixels) */ public static final @NonNull MediaFormat createVideoFormat( @NonNull String mime, int width, int height) { MediaFormat format = new MediaFormat(); format.setString(KEY_MIME, mime); format.setInteger(KEY_WIDTH, width); format.setInteger(KEY_HEIGHT, height); return format; } @Override public @NonNull String toString() { return mMap.toString(); } }