summaryrefslogtreecommitdiffstats
path: root/dom/media/mediasource/ContainerParser.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'dom/media/mediasource/ContainerParser.cpp')
-rw-r--r--dom/media/mediasource/ContainerParser.cpp767
1 files changed, 767 insertions, 0 deletions
diff --git a/dom/media/mediasource/ContainerParser.cpp b/dom/media/mediasource/ContainerParser.cpp
new file mode 100644
index 0000000000..b31ee42d67
--- /dev/null
+++ b/dom/media/mediasource/ContainerParser.cpp
@@ -0,0 +1,767 @@
+/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
+/* vim: set ts=8 sts=2 et sw=2 tw=80: */
+/* This Source Code Form is subject to the terms of the Mozilla Public
+ * License, v. 2.0. If a copy of the MPL was not distributed with this
+ * file, You can obtain one at http://mozilla.org/MPL/2.0/. */
+
+#include "ContainerParser.h"
+
+#include "WebMBufferedParser.h"
+#include "mozilla/EndianUtils.h"
+#include "mozilla/IntegerPrintfMacros.h"
+#include "mozilla/ErrorResult.h"
+#include "MoofParser.h"
+#include "mozilla/Logging.h"
+#include "mozilla/Maybe.h"
+#include "mozilla/Result.h"
+#include "MediaData.h"
+#include "nsMimeTypes.h"
+#ifdef MOZ_FMP4
+# include "AtomType.h"
+# include "BufferReader.h"
+# include "ByteStream.h"
+# include "MP4Interval.h"
+# include "SampleIterator.h"
+#endif
+#include "SourceBufferResource.h"
+#include <algorithm>
+
+extern mozilla::LogModule* GetMediaSourceSamplesLog();
+
+#define MSE_DEBUG(arg, ...) \
+ DDMOZ_LOG(GetMediaSourceSamplesLog(), mozilla::LogLevel::Debug, \
+ "(%s)::%s: " arg, mType.OriginalString().Data(), __func__, \
+ ##__VA_ARGS__)
+#define MSE_DEBUGV(arg, ...) \
+ DDMOZ_LOG(GetMediaSourceSamplesLog(), mozilla::LogLevel::Verbose, \
+ "(%s)::%s: " arg, mType.OriginalString().Data(), __func__, \
+ ##__VA_ARGS__)
+#define MSE_DEBUGVEX(_this, arg, ...) \
+ DDMOZ_LOGEX(_this, GetMediaSourceSamplesLog(), mozilla::LogLevel::Verbose, \
+ "(%s)::%s: " arg, mType.OriginalString().Data(), __func__, \
+ ##__VA_ARGS__)
+
+namespace mozilla {
+
+ContainerParser::ContainerParser(const MediaContainerType& aType)
+ : mHasInitData(false), mTotalParsed(0), mGlobalOffset(0), mType(aType) {}
+
+ContainerParser::~ContainerParser() = default;
+
+MediaResult ContainerParser::IsInitSegmentPresent(const MediaSpan& aData) {
+ MSE_DEBUG(
+ "aLength=%zu [%x%x%x%x]", aData.Length(),
+ aData.Length() > 0 ? aData[0] : 0, aData.Length() > 1 ? aData[1] : 0,
+ aData.Length() > 2 ? aData[2] : 0, aData.Length() > 3 ? aData[3] : 0);
+ return NS_ERROR_NOT_AVAILABLE;
+}
+
+MediaResult ContainerParser::IsMediaSegmentPresent(const MediaSpan& aData) {
+ MSE_DEBUG(
+ "aLength=%zu [%x%x%x%x]", aData.Length(),
+ aData.Length() > 0 ? aData[0] : 0, aData.Length() > 1 ? aData[1] : 0,
+ aData.Length() > 2 ? aData[2] : 0, aData.Length() > 3 ? aData[3] : 0);
+ return NS_ERROR_NOT_AVAILABLE;
+}
+
+MediaResult ContainerParser::ParseStartAndEndTimestamps(const MediaSpan& aData,
+ media::TimeUnit& aStart,
+ media::TimeUnit& aEnd) {
+ return NS_ERROR_NOT_AVAILABLE;
+}
+
+bool ContainerParser::TimestampsFuzzyEqual(int64_t aLhs, int64_t aRhs) {
+ return llabs(aLhs - aRhs) <= GetRoundingError();
+}
+
+int64_t ContainerParser::GetRoundingError() {
+ NS_WARNING("Using default ContainerParser::GetRoundingError implementation");
+ return 0;
+}
+
+bool ContainerParser::HasCompleteInitData() {
+ return mHasInitData && !!mInitData->Length();
+}
+
+MediaByteBuffer* ContainerParser::InitData() { return mInitData; }
+
+MediaByteRange ContainerParser::InitSegmentRange() {
+ return mCompleteInitSegmentRange;
+}
+
+MediaByteRange ContainerParser::MediaHeaderRange() {
+ return mCompleteMediaHeaderRange;
+}
+
+MediaByteRange ContainerParser::MediaSegmentRange() {
+ return mCompleteMediaSegmentRange;
+}
+
+DDLoggedTypeDeclNameAndBase(WebMContainerParser, ContainerParser);
+
+class WebMContainerParser
+ : public ContainerParser,
+ public DecoderDoctorLifeLogger<WebMContainerParser> {
+ public:
+ explicit WebMContainerParser(const MediaContainerType& aType)
+ : ContainerParser(aType), mParser(0), mOffset(0) {}
+
+ static const unsigned NS_PER_USEC = 1000;
+
+ MediaResult IsInitSegmentPresent(const MediaSpan& aData) override {
+ ContainerParser::IsInitSegmentPresent(aData);
+ if (aData.Length() < 4) {
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+
+ WebMBufferedParser parser(0);
+ nsTArray<WebMTimeDataOffset> mapping;
+ if (auto result = parser.Append(aData.Elements(), aData.Length(), mapping);
+ NS_FAILED(result)) {
+ return result;
+ }
+ return parser.mInitEndOffset > 0 ? NS_OK : NS_ERROR_NOT_AVAILABLE;
+ }
+
+ MediaResult IsMediaSegmentPresent(const MediaSpan& aData) override {
+ ContainerParser::IsMediaSegmentPresent(aData);
+ if (aData.Length() < 4) {
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+
+ WebMBufferedParser parser(0);
+ nsTArray<WebMTimeDataOffset> mapping;
+ parser.AppendMediaSegmentOnly();
+ if (auto result = parser.Append(aData.Elements(), aData.Length(), mapping);
+ NS_FAILED(result)) {
+ return result;
+ }
+ return parser.GetClusterOffset() >= 0 ? NS_OK : NS_ERROR_NOT_AVAILABLE;
+ }
+
+ MediaResult ParseStartAndEndTimestamps(const MediaSpan& aData,
+ media::TimeUnit& aStart,
+ media::TimeUnit& aEnd) override {
+ bool initSegment = NS_SUCCEEDED(IsInitSegmentPresent(aData));
+
+ if (mLastMapping &&
+ (initSegment || NS_SUCCEEDED(IsMediaSegmentPresent(aData)))) {
+ // The last data contained a complete cluster but we can only detect it
+ // now that a new one is starting.
+ // We use mOffset as end position to ensure that any blocks not reported
+ // by WebMBufferParser are properly skipped.
+ mCompleteMediaSegmentRange =
+ MediaByteRange(mLastMapping.ref().mSyncOffset, mOffset) +
+ mGlobalOffset;
+ mLastMapping.reset();
+ MSE_DEBUG("New cluster found at start, ending previous one");
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+
+ if (initSegment) {
+ mOffset = 0;
+ mParser = WebMBufferedParser(0);
+ mOverlappedMapping.Clear();
+ mInitData = new MediaByteBuffer();
+ mResource = new SourceBufferResource();
+ DDLINKCHILD("resource", mResource.get());
+ mCompleteInitSegmentRange = MediaByteRange();
+ mCompleteMediaHeaderRange = MediaByteRange();
+ mCompleteMediaSegmentRange = MediaByteRange();
+ mGlobalOffset = mTotalParsed;
+ }
+
+ // XXX if it only adds new mappings, overlapped but not available
+ // (e.g. overlap < 0) frames are "lost" from the reported mappings here.
+ nsTArray<WebMTimeDataOffset> mapping;
+ mapping.AppendElements(mOverlappedMapping);
+ mOverlappedMapping.Clear();
+ if (auto result = mParser.Append(aData.Elements(), aData.Length(), mapping);
+ NS_FAILED(result)) {
+ return result;
+ }
+ if (mResource) {
+ mResource->AppendData(aData);
+ }
+
+ // XXX This is a bit of a hack. Assume if there are no timecodes
+ // present and it's an init segment that it's _just_ an init segment.
+ // We should be more precise.
+ if (initSegment || !HasCompleteInitData()) {
+ if (mParser.mInitEndOffset > 0) {
+ MOZ_DIAGNOSTIC_ASSERT(mInitData && mResource &&
+ mParser.mInitEndOffset <= mResource->GetLength());
+ if (!mInitData->SetLength(mParser.mInitEndOffset, fallible)) {
+ // Super unlikely OOM
+ return NS_ERROR_OUT_OF_MEMORY;
+ }
+ mCompleteInitSegmentRange =
+ MediaByteRange(0, mParser.mInitEndOffset) + mGlobalOffset;
+ char* buffer = reinterpret_cast<char*>(mInitData->Elements());
+ mResource->ReadFromCache(buffer, 0, mParser.mInitEndOffset);
+ MSE_DEBUG("Stashed init of %" PRId64 " bytes.", mParser.mInitEndOffset);
+ mResource = nullptr;
+ } else {
+ MSE_DEBUG("Incomplete init found.");
+ }
+ mHasInitData = true;
+ }
+ mOffset += aData.Length();
+ mTotalParsed += aData.Length();
+
+ if (mapping.IsEmpty()) {
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+
+ // Calculate media range for first media segment.
+
+ // Check if we have a cluster finishing in the current data.
+ uint32_t endIdx = mapping.Length() - 1;
+ bool foundNewCluster = false;
+ while (mapping[0].mSyncOffset != mapping[endIdx].mSyncOffset) {
+ endIdx -= 1;
+ foundNewCluster = true;
+ }
+
+ int32_t completeIdx = endIdx;
+ while (completeIdx >= 0 && mOffset < mapping[completeIdx].mEndOffset) {
+ MSE_DEBUG("block is incomplete, missing: %" PRId64,
+ mapping[completeIdx].mEndOffset - mOffset);
+ completeIdx -= 1;
+ }
+
+ // Save parsed blocks for which we do not have all data yet.
+ mOverlappedMapping.AppendElements(mapping.Elements() + completeIdx + 1,
+ mapping.Length() - completeIdx - 1);
+
+ if (completeIdx < 0) {
+ mLastMapping.reset();
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+
+ if (mCompleteMediaHeaderRange.IsEmpty()) {
+ mCompleteMediaHeaderRange =
+ MediaByteRange(mapping[0].mSyncOffset, mapping[0].mEndOffset) +
+ mGlobalOffset;
+ }
+
+ if (foundNewCluster && mOffset >= mapping[endIdx].mEndOffset) {
+ // We now have all information required to delimit a complete cluster.
+ int64_t endOffset = mapping[endIdx + 1].mSyncOffset;
+ if (mapping[endIdx + 1].mInitOffset > mapping[endIdx].mInitOffset) {
+ // We have a new init segment before this cluster.
+ endOffset = mapping[endIdx + 1].mInitOffset;
+ }
+ mCompleteMediaSegmentRange =
+ MediaByteRange(mapping[endIdx].mSyncOffset, endOffset) +
+ mGlobalOffset;
+ } else if (mapping[endIdx].mClusterEndOffset >= 0 &&
+ mOffset >= mapping[endIdx].mClusterEndOffset) {
+ mCompleteMediaSegmentRange =
+ MediaByteRange(
+ mapping[endIdx].mSyncOffset,
+ mParser.EndSegmentOffset(mapping[endIdx].mClusterEndOffset)) +
+ mGlobalOffset;
+ }
+
+ Maybe<WebMTimeDataOffset> previousMapping;
+ if (completeIdx) {
+ previousMapping = Some(mapping[completeIdx - 1]);
+ } else {
+ previousMapping = mLastMapping;
+ }
+
+ mLastMapping = Some(mapping[completeIdx]);
+
+ if (!previousMapping && completeIdx + 1u >= mapping.Length()) {
+ // We have no previous nor next block available,
+ // so we can't estimate this block's duration.
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+
+ uint64_t frameDuration =
+ (completeIdx + 1u < mapping.Length())
+ ? mapping[completeIdx + 1].mTimecode -
+ mapping[completeIdx].mTimecode
+ : mapping[completeIdx].mTimecode - previousMapping.ref().mTimecode;
+ aStart = media::TimeUnit::FromNanoseconds(
+ AssertedCast<int64_t>(mapping[0].mTimecode));
+ aEnd = media::TimeUnit::FromNanoseconds(
+ AssertedCast<int64_t>(mapping[completeIdx].mTimecode + frameDuration));
+
+ MSE_DEBUG("[%" PRId64 ", %" PRId64 "] [fso=%" PRId64 ", leo=%" PRId64
+ ", l=%zu processedIdx=%u fs=%" PRId64 "]",
+ aStart.ToMicroseconds(), aEnd.ToMicroseconds(),
+ mapping[0].mSyncOffset, mapping[completeIdx].mEndOffset,
+ mapping.Length(), completeIdx, mCompleteMediaSegmentRange.mEnd);
+
+ return NS_OK;
+ }
+
+ int64_t GetRoundingError() override {
+ int64_t error = mParser.GetTimecodeScale() / NS_PER_USEC;
+ return error * 2;
+ }
+
+ private:
+ WebMBufferedParser mParser;
+ nsTArray<WebMTimeDataOffset> mOverlappedMapping;
+ int64_t mOffset;
+ Maybe<WebMTimeDataOffset> mLastMapping;
+};
+
+#ifdef MOZ_FMP4
+
+DDLoggedTypeDeclNameAndBase(MP4Stream, ByteStream);
+
+class MP4Stream : public ByteStream, public DecoderDoctorLifeLogger<MP4Stream> {
+ public:
+ explicit MP4Stream(SourceBufferResource* aResource);
+ virtual ~MP4Stream();
+ bool ReadAt(int64_t aOffset, void* aBuffer, size_t aCount,
+ size_t* aBytesRead) override;
+ bool CachedReadAt(int64_t aOffset, void* aBuffer, size_t aCount,
+ size_t* aBytesRead) override;
+ bool Length(int64_t* aSize) override;
+ const uint8_t* GetContiguousAccess(int64_t aOffset, size_t aSize) override;
+
+ private:
+ RefPtr<SourceBufferResource> mResource;
+};
+
+MP4Stream::MP4Stream(SourceBufferResource* aResource) : mResource(aResource) {
+ MOZ_COUNT_CTOR(MP4Stream);
+ MOZ_ASSERT(aResource);
+ DDLINKCHILD("resource", aResource);
+}
+
+MP4Stream::~MP4Stream() { MOZ_COUNT_DTOR(MP4Stream); }
+
+bool MP4Stream::ReadAt(int64_t aOffset, void* aBuffer, size_t aCount,
+ size_t* aBytesRead) {
+ return CachedReadAt(aOffset, aBuffer, aCount, aBytesRead);
+}
+
+bool MP4Stream::CachedReadAt(int64_t aOffset, void* aBuffer, size_t aCount,
+ size_t* aBytesRead) {
+ nsresult rv = mResource->ReadFromCache(reinterpret_cast<char*>(aBuffer),
+ aOffset, aCount);
+ if (NS_FAILED(rv)) {
+ *aBytesRead = 0;
+ return false;
+ }
+ *aBytesRead = aCount;
+ return true;
+}
+
+const uint8_t* MP4Stream::GetContiguousAccess(int64_t aOffset, size_t aSize) {
+ return mResource->GetContiguousAccess(aOffset, aSize);
+}
+
+bool MP4Stream::Length(int64_t* aSize) {
+ if (mResource->GetLength() < 0) return false;
+ *aSize = mResource->GetLength();
+ return true;
+}
+
+DDLoggedTypeDeclNameAndBase(MP4ContainerParser, ContainerParser);
+
+class MP4ContainerParser : public ContainerParser,
+ public DecoderDoctorLifeLogger<MP4ContainerParser> {
+ public:
+ explicit MP4ContainerParser(const MediaContainerType& aType)
+ : ContainerParser(aType) {}
+
+ MediaResult IsInitSegmentPresent(const MediaSpan& aData) override {
+ ContainerParser::IsInitSegmentPresent(aData);
+ // Each MP4 atom has a chunk size and chunk type. The root chunk in an MP4
+ // file is the 'ftyp' atom followed by a file type. We just check for a
+ // vaguely valid 'ftyp' atom.
+ if (aData.Length() < 8) {
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+ AtomParser parser(*this, aData, AtomParser::StopAt::eInitSegment);
+ if (!parser.IsValid()) {
+ return MediaResult(
+ NS_ERROR_FAILURE,
+ RESULT_DETAIL("Invalid Top-Level Box:%s", parser.LastInvalidBox()));
+ }
+ return parser.StartWithInitSegment() ? NS_OK : NS_ERROR_NOT_AVAILABLE;
+ }
+
+ MediaResult IsMediaSegmentPresent(const MediaSpan& aData) override {
+ if (aData.Length() < 8) {
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+ AtomParser parser(*this, aData, AtomParser::StopAt::eMediaSegment);
+ if (!parser.IsValid()) {
+ return MediaResult(
+ NS_ERROR_FAILURE,
+ RESULT_DETAIL("Invalid Box:%s", parser.LastInvalidBox()));
+ }
+ return parser.StartWithMediaSegment() ? NS_OK : NS_ERROR_NOT_AVAILABLE;
+ }
+
+ private:
+ class AtomParser {
+ public:
+ enum class StopAt { eInitSegment, eMediaSegment, eEnd };
+
+ AtomParser(const MP4ContainerParser& aParser, const MediaSpan& aData,
+ StopAt aStop = StopAt::eEnd) {
+ mValid = Init(aParser, aData, aStop).isOk();
+ }
+
+ Result<Ok, nsresult> Init(const MP4ContainerParser& aParser,
+ const MediaSpan& aData, StopAt aStop) {
+ const MediaContainerType mType(
+ aParser.ContainerType()); // for logging macro.
+ BufferReader reader(aData);
+ AtomType initAtom("moov");
+ AtomType mediaAtom("moof");
+ AtomType dataAtom("mdat");
+
+ // Valid top-level boxes defined in ISO/IEC 14496-12 (Table 1)
+ static const AtomType validBoxes[] = {
+ "ftyp", "moov", // init segment
+ "pdin", "free", "sidx", // optional prior moov box
+ "styp", "moof", "mdat", // media segment
+ "mfra", "skip", "meta", "meco", "ssix", "prft", // others.
+ "pssh", // optional with encrypted EME, though ignored.
+ "emsg", // ISO23009-1:2014 Section 5.10.3.3
+ "bloc", "uuid" // boxes accepted by chrome.
+ };
+
+ while (reader.Remaining() >= 8) {
+ uint32_t tmp;
+ MOZ_TRY_VAR(tmp, reader.ReadU32());
+ uint64_t size = tmp;
+ const uint8_t* typec = reader.Peek(4);
+ MOZ_TRY_VAR(tmp, reader.ReadU32());
+ AtomType type(tmp);
+ MSE_DEBUGVEX(&aParser, "Checking atom:'%c%c%c%c' @ %u", typec[0],
+ typec[1], typec[2], typec[3],
+ (uint32_t)reader.Offset() - 8);
+ if (std::find(std::begin(validBoxes), std::end(validBoxes), type) ==
+ std::end(validBoxes)) {
+ // No valid box found, no point continuing.
+ mLastInvalidBox[0] = typec[0];
+ mLastInvalidBox[1] = typec[1];
+ mLastInvalidBox[2] = typec[2];
+ mLastInvalidBox[3] = typec[3];
+ mLastInvalidBox[4] = '\0';
+ return Err(NS_ERROR_FAILURE);
+ }
+ if (mInitOffset.isNothing() && AtomType(type) == initAtom) {
+ mInitOffset = Some(reader.Offset());
+ }
+ if (mMediaOffset.isNothing() && AtomType(type) == mediaAtom) {
+ mMediaOffset = Some(reader.Offset());
+ }
+ if (mDataOffset.isNothing() && AtomType(type) == dataAtom) {
+ mDataOffset = Some(reader.Offset());
+ }
+ if (size == 1) {
+ // 64 bits size.
+ MOZ_TRY_VAR(size, reader.ReadU64());
+ } else if (size == 0) {
+ // Atom extends to the end of the buffer, it can't have what we're
+ // looking for.
+ break;
+ }
+ if (reader.Remaining() < size - 8) {
+ // Incomplete atom.
+ break;
+ }
+ reader.Read(size - 8);
+
+ if (aStop == StopAt::eInitSegment && (mInitOffset || mMediaOffset)) {
+ // When we're looking for an init segment, if we encountered a media
+ // segment, it we will need to be processed first. So we can stop
+ // right away if we have found a media segment.
+ break;
+ }
+ if (aStop == StopAt::eMediaSegment &&
+ (mInitOffset || (mMediaOffset && mDataOffset))) {
+ // When we're looking for a media segment, if we encountered an init
+ // segment, it we will need to be processed first. So we can stop
+ // right away if we have found an init segment.
+ break;
+ }
+ }
+
+ return Ok();
+ }
+
+ bool StartWithInitSegment() const {
+ return mInitOffset.isSome() && (mMediaOffset.isNothing() ||
+ mInitOffset.ref() < mMediaOffset.ref());
+ }
+ bool StartWithMediaSegment() const {
+ return mMediaOffset.isSome() && (mInitOffset.isNothing() ||
+ mMediaOffset.ref() < mInitOffset.ref());
+ }
+ bool IsValid() const { return mValid; }
+ const char* LastInvalidBox() const { return mLastInvalidBox; }
+
+ private:
+ Maybe<size_t> mInitOffset;
+ Maybe<size_t> mMediaOffset;
+ Maybe<size_t> mDataOffset;
+ bool mValid;
+ char mLastInvalidBox[5];
+ };
+
+ public:
+ MediaResult ParseStartAndEndTimestamps(const MediaSpan& aData,
+ media::TimeUnit& aStart,
+ media::TimeUnit& aEnd) override {
+ bool initSegment = NS_SUCCEEDED(IsInitSegmentPresent(aData));
+ if (initSegment) {
+ mResource = new SourceBufferResource();
+ DDLINKCHILD("resource", mResource.get());
+ mStream = new MP4Stream(mResource);
+ // We use a timestampOffset of 0 for ContainerParser, and require
+ // consumers of ParseStartAndEndTimestamps to add their timestamp offset
+ // manually. This allows the ContainerParser to be shared across different
+ // timestampOffsets.
+ mParser = MakeUnique<MoofParser>(mStream, AsVariant(ParseAllTracks{}),
+ /* aIsAudio = */ false);
+ DDLINKCHILD("parser", mParser.get());
+ mInitData = new MediaByteBuffer();
+ mCompleteInitSegmentRange = MediaByteRange();
+ mCompleteMediaHeaderRange = MediaByteRange();
+ mCompleteMediaSegmentRange = MediaByteRange();
+ mGlobalOffset = mTotalParsed;
+ } else if (!mStream || !mParser) {
+ mTotalParsed += aData.Length();
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+
+ MOZ_DIAGNOSTIC_ASSERT(mResource && mParser && mInitData,
+ "Should have received an init segment first");
+
+ mResource->AppendData(aData);
+ MediaByteRangeSet byteRanges;
+ byteRanges +=
+ MediaByteRange(int64_t(mParser->mOffset), mResource->GetLength());
+ mParser->RebuildFragmentedIndex(byteRanges);
+
+ if (initSegment || !HasCompleteInitData()) {
+ MediaByteRange& range = mParser->mInitRange;
+ if (range.Length()) {
+ mCompleteInitSegmentRange = range + mGlobalOffset;
+ if (!mInitData->SetLength(range.Length(), fallible)) {
+ // Super unlikely OOM
+ return NS_ERROR_OUT_OF_MEMORY;
+ }
+ char* buffer = reinterpret_cast<char*>(mInitData->Elements());
+ mResource->ReadFromCache(buffer, range.mStart, range.Length());
+ MSE_DEBUG("Stashed init of %" PRIu64 " bytes.", range.Length());
+ } else {
+ MSE_DEBUG("Incomplete init found.");
+ }
+ mHasInitData = true;
+ }
+ mTotalParsed += aData.Length();
+
+ MP4Interval<media::TimeUnit> compositionRange =
+ mParser->GetCompositionRange(byteRanges);
+
+ mCompleteMediaHeaderRange =
+ mParser->FirstCompleteMediaHeader() + mGlobalOffset;
+ mCompleteMediaSegmentRange =
+ mParser->FirstCompleteMediaSegment() + mGlobalOffset;
+
+ if (HasCompleteInitData()) {
+ mResource->EvictData(mParser->mOffset, mParser->mOffset);
+ }
+
+ if (compositionRange.IsNull()) {
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+ aStart = compositionRange.start;
+ aEnd = compositionRange.end;
+ MSE_DEBUG("[%" PRId64 ", %" PRId64 "]", aStart.ToMicroseconds(),
+ aEnd.ToMicroseconds());
+ return NS_OK;
+ }
+
+ // Gaps of up to 35ms (marginally longer than a single frame at 30fps) are
+ // considered to be sequential frames.
+ int64_t GetRoundingError() override { return 35000; }
+
+ private:
+ RefPtr<MP4Stream> mStream;
+ UniquePtr<MoofParser> mParser;
+};
+#endif // MOZ_FMP4
+
+#ifdef MOZ_FMP4
+DDLoggedTypeDeclNameAndBase(ADTSContainerParser, ContainerParser);
+
+class ADTSContainerParser
+ : public ContainerParser,
+ public DecoderDoctorLifeLogger<ADTSContainerParser> {
+ public:
+ explicit ADTSContainerParser(const MediaContainerType& aType)
+ : ContainerParser(aType) {}
+
+ typedef struct {
+ size_t header_length; // Length of just the initialization data.
+ size_t frame_length; // Includes header_length;
+ uint8_t aac_frames; // Number of AAC frames in the ADTS frame.
+ bool have_crc;
+ } Header;
+
+ /// Helper to parse the ADTS header, returning data we care about.
+ /// Returns true if the header is parsed successfully.
+ /// Returns false if the header is invalid or incomplete,
+ /// without modifying the passed-in Header object.
+ bool Parse(const MediaSpan& aData, Header& header) {
+ // ADTS initialization segments are just the packet header.
+ if (aData.Length() < 7) {
+ MSE_DEBUG("buffer too short for header.");
+ return false;
+ }
+ // Check 0xfffx sync word plus layer 0.
+ if ((aData[0] != 0xff) || ((aData[1] & 0xf6) != 0xf0)) {
+ MSE_DEBUG("no syncword.");
+ return false;
+ }
+ bool have_crc = !(aData[1] & 0x01);
+ if (have_crc && aData.Length() < 9) {
+ MSE_DEBUG("buffer too short for header with crc.");
+ return false;
+ }
+ uint8_t frequency_index = (aData[2] & 0x3c) >> 2;
+ MOZ_ASSERT(frequency_index < 16);
+ if (frequency_index == 15) {
+ MSE_DEBUG("explicit frequency disallowed.");
+ return false;
+ }
+ size_t header_length = have_crc ? 9 : 7;
+ size_t data_length = ((aData[3] & 0x03) << 11) | ((aData[4] & 0xff) << 3) |
+ ((aData[5] & 0xe0) >> 5);
+ uint8_t frames = (aData[6] & 0x03) + 1;
+ MOZ_ASSERT(frames > 0);
+ MOZ_ASSERT(frames < 4);
+
+ // Return successfully parsed data.
+ header.header_length = header_length;
+ header.frame_length = header_length + data_length;
+ header.aac_frames = frames;
+ header.have_crc = have_crc;
+ return true;
+ }
+
+ MediaResult IsInitSegmentPresent(const MediaSpan& aData) override {
+ // Call superclass for logging.
+ ContainerParser::IsInitSegmentPresent(aData);
+
+ Header header;
+ if (!Parse(aData, header)) {
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+
+ MSE_DEBUGV("%llu byte frame %d aac frames%s",
+ (unsigned long long)header.frame_length, (int)header.aac_frames,
+ header.have_crc ? " crc" : "");
+
+ return NS_OK;
+ }
+
+ MediaResult IsMediaSegmentPresent(const MediaSpan& aData) override {
+ // Call superclass for logging.
+ ContainerParser::IsMediaSegmentPresent(aData);
+
+ // Make sure we have a header so we know how long the frame is.
+ // NB this assumes the media segment buffer starts with an
+ // initialization segment. Since every frame has an ADTS header
+ // this is a normal place to divide packets, but we can re-parse
+ // mInitData if we need to handle separate media segments.
+ Header header;
+ if (!Parse(aData, header)) {
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+ // We're supposed to return true as long as aData contains the
+ // start of a media segment, whether or not it's complete. So
+ // return true if we have any data beyond the header.
+ if (aData.Length() <= header.header_length) {
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+
+ // We should have at least a partial frame.
+ return NS_OK;
+ }
+
+ MediaResult ParseStartAndEndTimestamps(const MediaSpan& aData,
+ media::TimeUnit& aStart,
+ media::TimeUnit& aEnd) override {
+ // ADTS header.
+ Header header;
+ if (!Parse(aData, header)) {
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+ mHasInitData = true;
+ mCompleteInitSegmentRange =
+ MediaByteRange(0, int64_t(header.header_length));
+
+ // Cache raw header in case the caller wants a copy.
+ mInitData = new MediaByteBuffer(header.header_length);
+ mInitData->AppendElements(aData.Elements(), header.header_length);
+
+ // Check that we have enough data for the frame body.
+ if (aData.Length() < header.frame_length) {
+ MSE_DEBUGV(
+ "Not enough data for %llu byte frame"
+ " in %llu byte buffer.",
+ (unsigned long long)header.frame_length,
+ (unsigned long long)(aData.Length()));
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+ mCompleteMediaSegmentRange =
+ MediaByteRange(header.header_length, header.frame_length);
+ // The ADTS MediaSource Byte Stream Format document doesn't
+ // define media header. Just treat it the same as the whole
+ // media segment.
+ mCompleteMediaHeaderRange = mCompleteMediaSegmentRange;
+
+ MSE_DEBUG("[%" PRId64 ", %" PRId64 "]", aStart.ToMicroseconds(),
+ aEnd.ToMicroseconds());
+ // We don't update timestamps, regardless.
+ return NS_ERROR_NOT_AVAILABLE;
+ }
+
+ // Audio shouldn't have gaps.
+ // Especially when we generate the timestamps ourselves.
+ int64_t GetRoundingError() override { return 0; }
+};
+#endif // MOZ_FMP4
+
+/*static*/
+UniquePtr<ContainerParser> ContainerParser::CreateForMIMEType(
+ const MediaContainerType& aType) {
+ if (aType.Type() == MEDIAMIMETYPE(VIDEO_WEBM) ||
+ aType.Type() == MEDIAMIMETYPE(AUDIO_WEBM)) {
+ return MakeUnique<WebMContainerParser>(aType);
+ }
+
+#ifdef MOZ_FMP4
+ if (aType.Type() == MEDIAMIMETYPE(VIDEO_MP4) ||
+ aType.Type() == MEDIAMIMETYPE(AUDIO_MP4)) {
+ return MakeUnique<MP4ContainerParser>(aType);
+ }
+ if (aType.Type() == MEDIAMIMETYPE("audio/aac")) {
+ return MakeUnique<ADTSContainerParser>(aType);
+ }
+#endif
+
+ return MakeUnique<ContainerParser>(aType);
+}
+
+#undef MSE_DEBUG
+#undef MSE_DEBUGV
+#undef MSE_DEBUGVEX
+
+} // namespace mozilla