diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-19 00:47:55 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-19 00:47:55 +0000 |
commit | 26a029d407be480d791972afb5975cf62c9360a6 (patch) | |
tree | f435a8308119effd964b339f76abb83a57c29483 /dom/media/MediaInfo.h | |
parent | Initial commit. (diff) | |
download | firefox-26a029d407be480d791972afb5975cf62c9360a6.tar.xz firefox-26a029d407be480d791972afb5975cf62c9360a6.zip |
Adding upstream version 124.0.1.upstream/124.0.1
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'dom/media/MediaInfo.h')
-rw-r--r-- | dom/media/MediaInfo.h | 752 |
1 files changed, 752 insertions, 0 deletions
diff --git a/dom/media/MediaInfo.h b/dom/media/MediaInfo.h new file mode 100644 index 0000000000..abe5dfd9da --- /dev/null +++ b/dom/media/MediaInfo.h @@ -0,0 +1,752 @@ +/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*- */ +/* vim: set ts=8 sts=2 et sw=2 tw=80: */ +/* This Source Code Form is subject to the terms of the Mozilla Public + * License, v. 2.0. If a copy of the MPL was not distributed with this + * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ +#if !defined(MediaInfo_h) +# define MediaInfo_h + +# include "mozilla/UniquePtr.h" +# include "mozilla/RefPtr.h" +# include "mozilla/Variant.h" +# include "nsTHashMap.h" +# include "nsString.h" +# include "nsTArray.h" +# include "AudioConfig.h" +# include "ImageTypes.h" +# include "MediaData.h" +# include "TimeUnits.h" +# include "mozilla/gfx/Point.h" // for gfx::IntSize +# include "mozilla/gfx/Rect.h" // for gfx::IntRect +# include "mozilla/gfx/Types.h" // for gfx::ColorDepth + +namespace mozilla { + +class AudioInfo; +class VideoInfo; +class TextInfo; + +class MetadataTag { + public: + MetadataTag(const nsACString& aKey, const nsACString& aValue) + : mKey(aKey), mValue(aValue) {} + nsCString mKey; + nsCString mValue; + bool operator==(const MetadataTag& rhs) const { + return mKey == rhs.mKey && mValue == rhs.mValue; + } +}; + +using MetadataTags = nsTHashMap<nsCStringHashKey, nsCString>; + +// Start codec specific data structs. If modifying these remember to also +// modify the MediaIPCUtils so that any new members are sent across IPC. + +// Generic types, we should prefer a specific type when we can. + +// Generic empty type. Prefer to use a specific type but not populate members +// if possible, as that helps with type checking. +struct NoCodecSpecificData { + bool operator==(const NoCodecSpecificData& rhs) const { return true; } +}; + +// Generic binary blob type. Prefer not to use this structure. It's here to ease +// the transition to codec specific structures in the code. +struct AudioCodecSpecificBinaryBlob { + bool operator==(const AudioCodecSpecificBinaryBlob& rhs) const { + return *mBinaryBlob == *rhs.mBinaryBlob; + } + + RefPtr<MediaByteBuffer> mBinaryBlob{new MediaByteBuffer}; +}; + +// End generic types. + +// Audio codec specific data types. + +struct AacCodecSpecificData { + bool operator==(const AacCodecSpecificData& rhs) const { + return *mEsDescriptorBinaryBlob == *rhs.mEsDescriptorBinaryBlob && + *mDecoderConfigDescriptorBinaryBlob == + *rhs.mDecoderConfigDescriptorBinaryBlob; + } + // An explanation for the necessity of handling the encoder delay and the + // padding is available here: + // https://developer.apple.com/library/archive/documentation/QuickTime/QTFF/QTFFAppenG/QTFFAppenG.html + + // The number of frames that should be skipped from the beginning of the + // decoded stream. + uint32_t mEncoderDelayFrames{0}; + + // The total number of frames of the media, that is, excluding the encoder + // delay and the padding of the last packet, that must be discarded. + uint64_t mMediaFrameCount{0}; + + // The bytes of the ES_Descriptor field parsed out of esds box. We store + // this as a blob as some decoders want this. + RefPtr<MediaByteBuffer> mEsDescriptorBinaryBlob{new MediaByteBuffer}; + + // The bytes of the DecoderConfigDescriptor field within the parsed + // ES_Descriptor. This is a subset of the ES_Descriptor, so it is technically + // redundant to store both. However, some decoders expect this binary blob + // instead of the whole ES_Descriptor, so both are stored for convenience + // and clarity (rather than reparsing the ES_Descriptor). + // TODO(bug 1768562): use a Span to track this rather than duplicating data. + RefPtr<MediaByteBuffer> mDecoderConfigDescriptorBinaryBlob{ + new MediaByteBuffer}; +}; + +struct FlacCodecSpecificData { + bool operator==(const FlacCodecSpecificData& rhs) const { + return *mStreamInfoBinaryBlob == *rhs.mStreamInfoBinaryBlob; + } + + // A binary blob of the data from the METADATA_BLOCK_STREAMINFO block + // in the flac header. + // See https://xiph.org/flac/format.html#metadata_block_streaminfo + // Consumers of this data (ffmpeg) take a blob, so we don't parse the data, + // just store the blob. For headerless flac files this will be left empty. + RefPtr<MediaByteBuffer> mStreamInfoBinaryBlob{new MediaByteBuffer}; +}; + +struct Mp3CodecSpecificData { + bool operator==(const Mp3CodecSpecificData& rhs) const { + return mEncoderDelayFrames == rhs.mEncoderDelayFrames && + mEncoderPaddingFrames == rhs.mEncoderPaddingFrames; + } + + // The number of frames that should be skipped from the beginning of the + // decoded stream. + // See https://bugzilla.mozilla.org/show_bug.cgi?id=1566389 for more info. + uint32_t mEncoderDelayFrames{0}; + + // The number of frames that should be skipped from the end of the decoded + // stream. + // See https://bugzilla.mozilla.org/show_bug.cgi?id=1566389 for more info. + uint32_t mEncoderPaddingFrames{0}; +}; + +struct OpusCodecSpecificData { + bool operator==(const OpusCodecSpecificData& rhs) const { + return mContainerCodecDelayFrames == rhs.mContainerCodecDelayFrames && + *mHeadersBinaryBlob == *rhs.mHeadersBinaryBlob; + } + // The codec delay (aka pre-skip) in audio frames. + // See https://tools.ietf.org/html/rfc7845#section-4.2 for more info. + // This member should store the codec delay parsed from the container file. + // In some cases (such as the ogg container), this information is derived + // from the same headers stored in the header blob, making storing this + // separately redundant. However, other containers store the delay in + // addition to the header blob, in which case we can check this container + // delay against the header delay to ensure they're consistent. + int64_t mContainerCodecDelayFrames{-1}; + + // A binary blob of opus header data, specifically the Identification Header. + // See https://datatracker.ietf.org/doc/html/rfc7845.html#section-5.1 + RefPtr<MediaByteBuffer> mHeadersBinaryBlob{new MediaByteBuffer}; +}; + +struct VorbisCodecSpecificData { + bool operator==(const VorbisCodecSpecificData& rhs) const { + return *mHeadersBinaryBlob == *rhs.mHeadersBinaryBlob; + } + + // A binary blob of headers in the 'extradata' format (the format ffmpeg + // expects for packing the extradata field). This is also the format some + // containers use for storing the data. Specifically, this format consists of + // the page_segments field, followed by the segment_table field, followed by + // the three Vorbis header packets, respectively the identification header, + // the comments header, and the setup header, in that order. + // See also https://xiph.org/vorbis/doc/framing.html and + // https://xiph.org/vorbis/doc/Vorbis_I_spec.html#x1-610004.2 + RefPtr<MediaByteBuffer> mHeadersBinaryBlob{new MediaByteBuffer}; +}; + +struct WaveCodecSpecificData { + bool operator==(const WaveCodecSpecificData& rhs) const { return true; } + // Intentionally empty. We don't store any wave specific data, but this + // variant is useful for type checking. +}; + +using AudioCodecSpecificVariant = + mozilla::Variant<NoCodecSpecificData, AudioCodecSpecificBinaryBlob, + AacCodecSpecificData, FlacCodecSpecificData, + Mp3CodecSpecificData, OpusCodecSpecificData, + VorbisCodecSpecificData, WaveCodecSpecificData>; + +// Returns a binary blob representation of the AudioCodecSpecificVariant. This +// does not guarantee that a binary representation exists. Will return an empty +// buffer if no representation exists. Prefer `GetAudioCodecSpecificBlob` which +// asserts if getting a blob is unexpected for a given codec config. +inline already_AddRefed<MediaByteBuffer> ForceGetAudioCodecSpecificBlob( + const AudioCodecSpecificVariant& v) { + return v.match( + [](const NoCodecSpecificData&) { + return RefPtr<MediaByteBuffer>(new MediaByteBuffer).forget(); + }, + [](const AudioCodecSpecificBinaryBlob& binaryBlob) { + return RefPtr<MediaByteBuffer>(binaryBlob.mBinaryBlob).forget(); + }, + [](const AacCodecSpecificData& aacData) { + // We return the mDecoderConfigDescriptor blob here, as it is more + // commonly used by decoders at time of writing than the + // ES_Descriptor data. However, consumers of this data should + // prefer getting one or the other specifically, rather than + // calling this. + return RefPtr<MediaByteBuffer>( + aacData.mDecoderConfigDescriptorBinaryBlob) + .forget(); + }, + [](const FlacCodecSpecificData& flacData) { + return RefPtr<MediaByteBuffer>(flacData.mStreamInfoBinaryBlob).forget(); + }, + [](const Mp3CodecSpecificData&) { + return RefPtr<MediaByteBuffer>(new MediaByteBuffer).forget(); + }, + [](const OpusCodecSpecificData& opusData) { + return RefPtr<MediaByteBuffer>(opusData.mHeadersBinaryBlob).forget(); + }, + [](const VorbisCodecSpecificData& vorbisData) { + return RefPtr<MediaByteBuffer>(vorbisData.mHeadersBinaryBlob).forget(); + }, + [](const WaveCodecSpecificData&) { + return RefPtr<MediaByteBuffer>(new MediaByteBuffer).forget(); + }); +} + +// Same as `ForceGetAudioCodecSpecificBlob` but with extra asserts to ensure +// we're not trying to get a binary blob from codecs where we don't store the +// information as a blob or where a blob is ambiguous. +inline already_AddRefed<MediaByteBuffer> GetAudioCodecSpecificBlob( + const AudioCodecSpecificVariant& v) { + MOZ_ASSERT(!v.is<NoCodecSpecificData>(), + "NoCodecSpecificData shouldn't be used as a blob"); + MOZ_ASSERT(!v.is<AacCodecSpecificData>(), + "AacCodecSpecificData has 2 blobs internally, one should " + "explicitly be selected"); + MOZ_ASSERT(!v.is<Mp3CodecSpecificData>(), + "Mp3CodecSpecificData shouldn't be used as a blob"); + + return ForceGetAudioCodecSpecificBlob(v); +} + +// End audio codec specific data types. + +// End codec specific data structs. + +class TrackInfo { + public: + enum TrackType { kUndefinedTrack, kAudioTrack, kVideoTrack, kTextTrack }; + TrackInfo(TrackType aType, const nsAString& aId, const nsAString& aKind, + const nsAString& aLabel, const nsAString& aLanguage, bool aEnabled, + uint32_t aTrackId) + : mId(aId), + mKind(aKind), + mLabel(aLabel), + mLanguage(aLanguage), + mEnabled(aEnabled), + mTrackId(aTrackId), + mIsRenderedExternally(false), + mType(aType) { + MOZ_COUNT_CTOR(TrackInfo); + } + + // Only used for backward compatibility. Do not use in new code. + void Init(const nsAString& aId, const nsAString& aKind, + const nsAString& aLabel, const nsAString& aLanguage, + bool aEnabled) { + mId = aId; + mKind = aKind; + mLabel = aLabel; + mLanguage = aLanguage; + mEnabled = aEnabled; + } + + // Fields common with MediaTrack object. + nsString mId; + nsString mKind; + nsString mLabel; + nsString mLanguage; + bool mEnabled; + + uint32_t mTrackId; + + nsCString mMimeType; + media::TimeUnit mDuration; + media::TimeUnit mMediaTime; + uint32_t mTimeScale = 0; + CryptoTrack mCrypto; + + CopyableTArray<MetadataTag> mTags; + + // True if the track is gonna be (decrypted)/decoded and + // rendered directly by non-gecko components. + bool mIsRenderedExternally; + + virtual AudioInfo* GetAsAudioInfo() { return nullptr; } + virtual VideoInfo* GetAsVideoInfo() { return nullptr; } + virtual TextInfo* GetAsTextInfo() { return nullptr; } + virtual const AudioInfo* GetAsAudioInfo() const { return nullptr; } + virtual const VideoInfo* GetAsVideoInfo() const { return nullptr; } + virtual const TextInfo* GetAsTextInfo() const { return nullptr; } + + bool IsAudio() const { return !!GetAsAudioInfo(); } + bool IsVideo() const { return !!GetAsVideoInfo(); } + bool IsText() const { return !!GetAsTextInfo(); } + TrackType GetType() const { return mType; } + + nsCString ToString() const; + + bool virtual IsValid() const = 0; + + virtual UniquePtr<TrackInfo> Clone() const = 0; + + MOZ_COUNTED_DTOR_VIRTUAL(TrackInfo) + + protected: + TrackInfo(const TrackInfo& aOther) { + mId = aOther.mId; + mKind = aOther.mKind; + mLabel = aOther.mLabel; + mLanguage = aOther.mLanguage; + mEnabled = aOther.mEnabled; + mTrackId = aOther.mTrackId; + mMimeType = aOther.mMimeType; + mDuration = aOther.mDuration; + mMediaTime = aOther.mMediaTime; + mCrypto = aOther.mCrypto; + mIsRenderedExternally = aOther.mIsRenderedExternally; + mType = aOther.mType; + mTags = aOther.mTags.Clone(); + MOZ_COUNT_CTOR(TrackInfo); + } + bool IsEqualTo(const TrackInfo& rhs) const; + + private: + TrackType mType; +}; + +// String version of track type. +const char* TrackTypeToStr(TrackInfo::TrackType aTrack); + +enum class VideoRotation { + kDegree_0 = 0, + kDegree_90 = 90, + kDegree_180 = 180, + kDegree_270 = 270, +}; + +// Stores info relevant to presenting media frames. +class VideoInfo : public TrackInfo { + public: + VideoInfo() : VideoInfo(-1, -1) {} + + VideoInfo(int32_t aWidth, int32_t aHeight) + : VideoInfo(gfx::IntSize(aWidth, aHeight)) {} + + explicit VideoInfo(const gfx::IntSize& aSize) + : TrackInfo(kVideoTrack, u"2"_ns, u"main"_ns, u""_ns, u""_ns, true, 2), + mDisplay(aSize), + mStereoMode(StereoMode::MONO), + mImage(aSize), + mCodecSpecificConfig(new MediaByteBuffer), + mExtraData(new MediaByteBuffer), + mRotation(VideoRotation::kDegree_0) {} + + VideoInfo(const VideoInfo& aOther) = default; + + bool operator==(const VideoInfo& rhs) const; + + bool IsValid() const override { + return mDisplay.width > 0 && mDisplay.height > 0; + } + + VideoInfo* GetAsVideoInfo() override { return this; } + + const VideoInfo* GetAsVideoInfo() const override { return this; } + + UniquePtr<TrackInfo> Clone() const override { + return MakeUnique<VideoInfo>(*this); + } + + void SetAlpha(bool aAlphaPresent) { mAlphaPresent = aAlphaPresent; } + + bool HasAlpha() const { return mAlphaPresent; } + + gfx::IntRect ImageRect() const { + if (!mImageRect) { + return gfx::IntRect(0, 0, mImage.width, mImage.height); + } + return *mImageRect; + } + + void SetImageRect(const gfx::IntRect& aRect) { mImageRect = Some(aRect); } + void ResetImageRect() { mImageRect.reset(); } + + // Returned the crop rectangle scaled to aWidth/aHeight size relative to + // mImage size. + // If aWidth and aHeight are identical to the original + // mImage.width/mImage.height then the scaling ratio will be 1. This is used + // for when the frame size is different from what the container reports. This + // is legal in WebM, and we will preserve the ratio of the crop rectangle as + // it was reported relative to the picture size reported by the container. + gfx::IntRect ScaledImageRect(int64_t aWidth, int64_t aHeight) const { + if ((aWidth == mImage.width && aHeight == mImage.height) || !mImage.width || + !mImage.height) { + return ImageRect(); + } + + gfx::IntRect imageRect = ImageRect(); + int64_t w = (aWidth * imageRect.Width()) / mImage.width; + int64_t h = (aHeight * imageRect.Height()) / mImage.height; + if (!w || !h) { + return imageRect; + } + + imageRect.x = AssertedCast<int>((imageRect.x * aWidth) / mImage.width); + imageRect.y = AssertedCast<int>((imageRect.y * aHeight) / mImage.height); + imageRect.SetWidth(AssertedCast<int>(w)); + imageRect.SetHeight(AssertedCast<int>(h)); + return imageRect; + } + + VideoRotation ToSupportedRotation(int32_t aDegree) const { + switch (aDegree) { + case 90: + return VideoRotation::kDegree_90; + case 180: + return VideoRotation::kDegree_180; + case 270: + return VideoRotation::kDegree_270; + default: + NS_WARNING_ASSERTION(aDegree == 0, "Invalid rotation degree, ignored"); + return VideoRotation::kDegree_0; + } + } + + nsString ToString() const { + std::array YUVColorSpaceStrings = {"BT601", "BT709", "BT2020", "Identity", + "Default"}; + + std::array ColorDepthStrings = { + "COLOR_8", + "COLOR_10", + "COLOR_12", + "COLOR_16", + }; + + std::array TransferFunctionStrings = { + "BT709", + "SRGB", + "PQ", + "HLG", + }; + + std::array ColorRangeStrings = { + "LIMITED", + "FULL", + }; + + std::array ColorPrimariesStrings = {"Display", + "UNKNOWN" + "SRGB", + "DISPLAY_P3", + "BT601_525", + "BT709", + "BT601_625" + "BT709", + "BT2020"}; + nsString rv; + rv.AppendLiteral(u"VideoInfo: "); + rv.AppendPrintf("display size: %dx%d ", mDisplay.Width(), + mDisplay.Height()); + rv.AppendPrintf("stereo mode: %d", static_cast<int>(mStereoMode)); + rv.AppendPrintf("image size: %dx%d ", mImage.Width(), mImage.Height()); + if (mCodecSpecificConfig) { + rv.AppendPrintf("codec specific config: %zu bytes", + mCodecSpecificConfig->Length()); + } + if (mExtraData) { + rv.AppendPrintf("extra data: %zu bytes", mExtraData->Length()); + } + rv.AppendPrintf("rotation: %d", static_cast<int>(mRotation)); + rv.AppendPrintf("colors: %s", ColorDepthStrings[static_cast<int>(mColorDepth)]); + if (mColorSpace) { + rv.AppendPrintf( + "YUV colorspace: %s ", + YUVColorSpaceStrings[static_cast<int>(mColorSpace.value())]); + } + if (mColorPrimaries) { + rv.AppendPrintf( + "color primaries: %s ", + ColorPrimariesStrings[static_cast<int>(mColorPrimaries.value())]); + } + if (mTransferFunction) { + rv.AppendPrintf( + "transfer function %s ", + TransferFunctionStrings[static_cast<int>(mTransferFunction.value())]); + } + rv.AppendPrintf("color range: %s", ColorRangeStrings[static_cast<int>(mColorRange)]); + if (mImageRect) { + rv.AppendPrintf("image rect: %dx%d", mImageRect->Width(), + mImageRect->Height()); + } + rv.AppendPrintf("alpha present: %s", mAlphaPresent ? "true" : "false"); + if (mFrameRate) { + rv.AppendPrintf("frame rate: %dHz", mFrameRate.value()); + } + + return rv; + } + + // Size in pixels at which the video is rendered. This is after it has + // been scaled by its aspect ratio. + gfx::IntSize mDisplay; + + // Indicates the frame layout for single track stereo videos. + StereoMode mStereoMode; + + // Size of the decoded video's image. + gfx::IntSize mImage; + + RefPtr<MediaByteBuffer> mCodecSpecificConfig; + RefPtr<MediaByteBuffer> mExtraData; + + // Describing how many degrees video frames should be rotated in clock-wise to + // get correct view. + VideoRotation mRotation; + + // Should be 8, 10 or 12. Default value is 8. + gfx::ColorDepth mColorDepth = gfx::ColorDepth::COLOR_8; + + // Matrix coefficients (if specified by the video) imply a colorspace. + Maybe<gfx::YUVColorSpace> mColorSpace; + + // Color primaries are independent from the coefficients. + Maybe<gfx::ColorSpace2> mColorPrimaries; + + // Transfer functions get their own member, which may not be strongly + // correlated to the colorspace. + Maybe<gfx::TransferFunction> mTransferFunction; + + // True indicates no restriction on Y, U, V values (otherwise 16-235 for 8 + // bits etc) + gfx::ColorRange mColorRange = gfx::ColorRange::LIMITED; + + Maybe<int32_t> GetFrameRate() const { return mFrameRate; } + void SetFrameRate(int32_t aRate) { mFrameRate = Some(aRate); } + + private: + friend struct IPC::ParamTraits<VideoInfo>; + + // mImage may be cropped; currently only used with the WebM container. + // If unset, no cropping is to occur. + Maybe<gfx::IntRect> mImageRect; + + // Indicates whether or not frames may contain alpha information. + bool mAlphaPresent = false; + + Maybe<int32_t> mFrameRate; +}; + +class AudioInfo : public TrackInfo { + public: + AudioInfo() + : TrackInfo(kAudioTrack, u"1"_ns, u"main"_ns, u""_ns, u""_ns, true, 1), + mRate(0), + mChannels(0), + mChannelMap(AudioConfig::ChannelLayout::UNKNOWN_MAP), + mBitDepth(0), + mProfile(0), + mExtendedProfile(0) {} + + AudioInfo(const AudioInfo& aOther) = default; + + bool operator==(const AudioInfo& rhs) const; + + static const uint32_t MAX_RATE = 640000; + static const uint32_t MAX_CHANNEL_COUNT = 256; + + bool IsValid() const override { + return mChannels > 0 && mChannels <= MAX_CHANNEL_COUNT && mRate > 0 && + mRate <= MAX_RATE; + } + + AudioInfo* GetAsAudioInfo() override { return this; } + + const AudioInfo* GetAsAudioInfo() const override { return this; } + + nsCString ToString() const; + + UniquePtr<TrackInfo> Clone() const override { + return MakeUnique<AudioInfo>(*this); + } + + // Sample rate. + uint32_t mRate; + + // Number of audio channels. + uint32_t mChannels; + // The AudioConfig::ChannelLayout map. Channels are ordered as per SMPTE + // definition. A value of UNKNOWN_MAP indicates unknown layout. + // ChannelMap is an unsigned bitmap compatible with Windows' WAVE and FFmpeg + // channel map. + AudioConfig::ChannelLayout::ChannelMap mChannelMap; + + // Bits per sample. + uint32_t mBitDepth; + + // Codec profile. + uint8_t mProfile; + + // Extended codec profile. + uint8_t mExtendedProfile; + + AudioCodecSpecificVariant mCodecSpecificConfig{NoCodecSpecificData{}}; +}; + +class EncryptionInfo { + public: + EncryptionInfo() : mEncrypted(false) {} + + struct InitData { + template <typename AInitDatas> + InitData(const nsAString& aType, AInitDatas&& aInitData) + : mType(aType), mInitData(std::forward<AInitDatas>(aInitData)) {} + + // Encryption type to be passed to JS. Usually `cenc'. + nsString mType; + + // Encryption data. + CopyableTArray<uint8_t> mInitData; + }; + using InitDatas = CopyableTArray<InitData>; + + // True if the stream has encryption metadata + bool IsEncrypted() const { return mEncrypted; } + + void Reset() { + mEncrypted = false; + mInitDatas.Clear(); + } + + template <typename AInitDatas> + void AddInitData(const nsAString& aType, AInitDatas&& aInitData) { + mInitDatas.AppendElement( + InitData(aType, std::forward<AInitDatas>(aInitData))); + mEncrypted = true; + } + + void AddInitData(const EncryptionInfo& aInfo) { + mInitDatas.AppendElements(aInfo.mInitDatas); + mEncrypted = !!mInitDatas.Length(); + } + + // One 'InitData' per encrypted buffer. + InitDatas mInitDatas; + + private: + bool mEncrypted; +}; + +class MediaInfo { + public: + bool HasVideo() const { return mVideo.IsValid(); } + + void EnableVideo() { + if (HasVideo()) { + return; + } + // Set dummy values so that HasVideo() will return true; + // See VideoInfo::IsValid() + mVideo.mDisplay = gfx::IntSize(1, 1); + } + + bool HasAudio() const { return mAudio.IsValid(); } + + void EnableAudio() { + if (HasAudio()) { + return; + } + // Set dummy values so that HasAudio() will return true; + // See AudioInfo::IsValid() + mAudio.mChannels = 2; + mAudio.mRate = 44100; + } + + bool IsEncrypted() const { + return (HasAudio() && mAudio.mCrypto.IsEncrypted()) || + (HasVideo() && mVideo.mCrypto.IsEncrypted()); + } + + bool HasValidMedia() const { return HasVideo() || HasAudio(); } + + // TODO: Store VideoInfo and AudioIndo in arrays to support multi-tracks. + VideoInfo mVideo; + AudioInfo mAudio; + + // If the metadata includes a duration, we store it here. + media::NullableTimeUnit mMetadataDuration; + + // The Ogg reader tries to kinda-sorta compute the duration by seeking to the + // end and determining the timestamp of the last frame. This isn't useful as + // a duration until we know the start time, so we need to track it separately. + media::NullableTimeUnit mUnadjustedMetadataEndTime; + + // True if the media is seekable (i.e. supports random access). + bool mMediaSeekable = true; + + // True if the media is only seekable within its buffered ranges. + bool mMediaSeekableOnlyInBufferedRanges = false; + + EncryptionInfo mCrypto; + + // The minimum of start times of audio and video tracks. + // Use to map the zero time on the media timeline to the first frame. + media::TimeUnit mStartTime; +}; + +class TrackInfoSharedPtr { + NS_INLINE_DECL_THREADSAFE_REFCOUNTING(TrackInfoSharedPtr) + public: + TrackInfoSharedPtr(const TrackInfo& aOriginal, uint32_t aStreamID) + : mInfo(aOriginal.Clone()), + mStreamSourceID(aStreamID), + mMimeType(mInfo->mMimeType) {} + + uint32_t GetID() const { return mStreamSourceID; } + + operator const TrackInfo*() const { return mInfo.get(); } + + const TrackInfo* operator*() const { return mInfo.get(); } + + const TrackInfo* operator->() const { + MOZ_ASSERT(mInfo.get(), "dereferencing a UniquePtr containing nullptr"); + return mInfo.get(); + } + + const AudioInfo* GetAsAudioInfo() const { + return mInfo ? mInfo->GetAsAudioInfo() : nullptr; + } + + const VideoInfo* GetAsVideoInfo() const { + return mInfo ? mInfo->GetAsVideoInfo() : nullptr; + } + + const TextInfo* GetAsTextInfo() const { + return mInfo ? mInfo->GetAsTextInfo() : nullptr; + } + + private: + ~TrackInfoSharedPtr() = default; + UniquePtr<TrackInfo> mInfo; + // A unique ID, guaranteed to change when changing streams. + uint32_t mStreamSourceID; + + public: + const nsCString& mMimeType; +}; + +} // namespace mozilla + +#endif // MediaInfo_h |