1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
|
/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
/* vim: set ts=8 sts=2 et sw=2 tw=80: */
/* This Source Code Form is subject to the terms of the Mozilla Public
* License, v. 2.0. If a copy of the MPL was not distributed with this
* file, You can obtain one at http://mozilla.org/MPL/2.0/. */
#if !defined(MediaInfo_h)
# define MediaInfo_h
# include "mozilla/UniquePtr.h"
# include "mozilla/RefPtr.h"
# include "mozilla/Variant.h"
# include "nsTHashMap.h"
# include "nsString.h"
# include "nsTArray.h"
# include "AudioConfig.h"
# include "ImageTypes.h"
# include "MediaData.h"
# include "TimeUnits.h"
# include "mozilla/gfx/Point.h" // for gfx::IntSize
# include "mozilla/gfx/Rect.h" // for gfx::IntRect
# include "mozilla/gfx/Types.h" // for gfx::ColorDepth
namespace mozilla {
class AudioInfo;
class VideoInfo;
class TextInfo;
class MetadataTag {
public:
MetadataTag(const nsACString& aKey, const nsACString& aValue)
: mKey(aKey), mValue(aValue) {}
nsCString mKey;
nsCString mValue;
bool operator==(const MetadataTag& rhs) const {
return mKey == rhs.mKey && mValue == rhs.mValue;
}
};
using MetadataTags = nsTHashMap<nsCStringHashKey, nsCString>;
// Start codec specific data structs. If modifying these remember to also
// modify the MediaIPCUtils so that any new members are sent across IPC.
// Generic types, we should prefer a specific type when we can.
// Generic empty type. Prefer to use a specific type but not populate members
// if possible, as that helps with type checking.
struct NoCodecSpecificData {
bool operator==(const NoCodecSpecificData& rhs) const { return true; }
};
// Generic binary blob type. Prefer not to use this structure. It's here to ease
// the transition to codec specific structures in the code.
struct AudioCodecSpecificBinaryBlob {
bool operator==(const AudioCodecSpecificBinaryBlob& rhs) const {
return *mBinaryBlob == *rhs.mBinaryBlob;
}
RefPtr<MediaByteBuffer> mBinaryBlob{new MediaByteBuffer};
};
// End generic types.
// Audio codec specific data types.
struct AacCodecSpecificData {
bool operator==(const AacCodecSpecificData& rhs) const {
return *mEsDescriptorBinaryBlob == *rhs.mEsDescriptorBinaryBlob &&
*mDecoderConfigDescriptorBinaryBlob ==
*rhs.mDecoderConfigDescriptorBinaryBlob;
}
// An explanation for the necessity of handling the encoder delay and the
// padding is available here:
// https://developer.apple.com/library/archive/documentation/QuickTime/QTFF/QTFFAppenG/QTFFAppenG.html
// The number of frames that should be skipped from the beginning of the
// decoded stream.
uint32_t mEncoderDelayFrames{0};
// The total number of frames of the media, that is, excluding the encoder
// delay and the padding of the last packet, that must be discarded.
uint64_t mMediaFrameCount{0};
// The bytes of the ES_Descriptor field parsed out of esds box. We store
// this as a blob as some decoders want this.
RefPtr<MediaByteBuffer> mEsDescriptorBinaryBlob{new MediaByteBuffer};
// The bytes of the DecoderConfigDescriptor field within the parsed
// ES_Descriptor. This is a subset of the ES_Descriptor, so it is technically
// redundant to store both. However, some decoders expect this binary blob
// instead of the whole ES_Descriptor, so both are stored for convenience
// and clarity (rather than reparsing the ES_Descriptor).
// TODO(bug 1768562): use a Span to track this rather than duplicating data.
RefPtr<MediaByteBuffer> mDecoderConfigDescriptorBinaryBlob{
new MediaByteBuffer};
};
struct FlacCodecSpecificData {
bool operator==(const FlacCodecSpecificData& rhs) const {
return *mStreamInfoBinaryBlob == *rhs.mStreamInfoBinaryBlob;
}
// A binary blob of the data from the METADATA_BLOCK_STREAMINFO block
// in the flac header.
// See https://xiph.org/flac/format.html#metadata_block_streaminfo
// Consumers of this data (ffmpeg) take a blob, so we don't parse the data,
// just store the blob. For headerless flac files this will be left empty.
RefPtr<MediaByteBuffer> mStreamInfoBinaryBlob{new MediaByteBuffer};
};
struct Mp3CodecSpecificData final {
bool operator==(const Mp3CodecSpecificData& rhs) const {
return mEncoderDelayFrames == rhs.mEncoderDelayFrames &&
mEncoderPaddingFrames == rhs.mEncoderPaddingFrames;
}
auto MutTiedFields() {
return std::tie(mEncoderDelayFrames, mEncoderPaddingFrames);
}
// The number of frames that should be skipped from the beginning of the
// decoded stream.
// See https://bugzilla.mozilla.org/show_bug.cgi?id=1566389 for more info.
uint32_t mEncoderDelayFrames{0};
// The number of frames that should be skipped from the end of the decoded
// stream.
// See https://bugzilla.mozilla.org/show_bug.cgi?id=1566389 for more info.
uint32_t mEncoderPaddingFrames{0};
};
struct OpusCodecSpecificData {
bool operator==(const OpusCodecSpecificData& rhs) const {
return mContainerCodecDelayFrames == rhs.mContainerCodecDelayFrames &&
*mHeadersBinaryBlob == *rhs.mHeadersBinaryBlob;
}
// The codec delay (aka pre-skip) in audio frames.
// See https://tools.ietf.org/html/rfc7845#section-4.2 for more info.
// This member should store the codec delay parsed from the container file.
// In some cases (such as the ogg container), this information is derived
// from the same headers stored in the header blob, making storing this
// separately redundant. However, other containers store the delay in
// addition to the header blob, in which case we can check this container
// delay against the header delay to ensure they're consistent.
int64_t mContainerCodecDelayFrames{-1};
// A binary blob of opus header data, specifically the Identification Header.
// See https://datatracker.ietf.org/doc/html/rfc7845.html#section-5.1
RefPtr<MediaByteBuffer> mHeadersBinaryBlob{new MediaByteBuffer};
};
struct VorbisCodecSpecificData {
bool operator==(const VorbisCodecSpecificData& rhs) const {
return *mHeadersBinaryBlob == *rhs.mHeadersBinaryBlob;
}
// A binary blob of headers in the 'extradata' format (the format ffmpeg
// expects for packing the extradata field). This is also the format some
// containers use for storing the data. Specifically, this format consists of
// the page_segments field, followed by the segment_table field, followed by
// the three Vorbis header packets, respectively the identification header,
// the comments header, and the setup header, in that order.
// See also https://xiph.org/vorbis/doc/framing.html and
// https://xiph.org/vorbis/doc/Vorbis_I_spec.html#x1-610004.2
RefPtr<MediaByteBuffer> mHeadersBinaryBlob{new MediaByteBuffer};
};
struct WaveCodecSpecificData {
bool operator==(const WaveCodecSpecificData& rhs) const { return true; }
// Intentionally empty. We don't store any wave specific data, but this
// variant is useful for type checking.
};
using AudioCodecSpecificVariant =
mozilla::Variant<NoCodecSpecificData, AudioCodecSpecificBinaryBlob,
AacCodecSpecificData, FlacCodecSpecificData,
Mp3CodecSpecificData, OpusCodecSpecificData,
VorbisCodecSpecificData, WaveCodecSpecificData>;
// Returns a binary blob representation of the AudioCodecSpecificVariant. This
// does not guarantee that a binary representation exists. Will return an empty
// buffer if no representation exists. Prefer `GetAudioCodecSpecificBlob` which
// asserts if getting a blob is unexpected for a given codec config.
inline already_AddRefed<MediaByteBuffer> ForceGetAudioCodecSpecificBlob(
const AudioCodecSpecificVariant& v) {
return v.match(
[](const NoCodecSpecificData&) {
return RefPtr<MediaByteBuffer>(new MediaByteBuffer).forget();
},
[](const AudioCodecSpecificBinaryBlob& binaryBlob) {
return RefPtr<MediaByteBuffer>(binaryBlob.mBinaryBlob).forget();
},
[](const AacCodecSpecificData& aacData) {
// We return the mDecoderConfigDescriptor blob here, as it is more
// commonly used by decoders at time of writing than the
// ES_Descriptor data. However, consumers of this data should
// prefer getting one or the other specifically, rather than
// calling this.
return RefPtr<MediaByteBuffer>(
aacData.mDecoderConfigDescriptorBinaryBlob)
.forget();
},
[](const FlacCodecSpecificData& flacData) {
return RefPtr<MediaByteBuffer>(flacData.mStreamInfoBinaryBlob).forget();
},
[](const Mp3CodecSpecificData&) {
return RefPtr<MediaByteBuffer>(new MediaByteBuffer).forget();
},
[](const OpusCodecSpecificData& opusData) {
return RefPtr<MediaByteBuffer>(opusData.mHeadersBinaryBlob).forget();
},
[](const VorbisCodecSpecificData& vorbisData) {
return RefPtr<MediaByteBuffer>(vorbisData.mHeadersBinaryBlob).forget();
},
[](const WaveCodecSpecificData&) {
return RefPtr<MediaByteBuffer>(new MediaByteBuffer).forget();
});
}
// Same as `ForceGetAudioCodecSpecificBlob` but with extra asserts to ensure
// we're not trying to get a binary blob from codecs where we don't store the
// information as a blob or where a blob is ambiguous.
inline already_AddRefed<MediaByteBuffer> GetAudioCodecSpecificBlob(
const AudioCodecSpecificVariant& v) {
return ForceGetAudioCodecSpecificBlob(v);
}
// End audio codec specific data types.
// End codec specific data structs.
class TrackInfo {
public:
enum TrackType { kUndefinedTrack, kAudioTrack, kVideoTrack, kTextTrack };
TrackInfo(TrackType aType, const nsAString& aId, const nsAString& aKind,
const nsAString& aLabel, const nsAString& aLanguage, bool aEnabled,
uint32_t aTrackId)
: mId(aId),
mKind(aKind),
mLabel(aLabel),
mLanguage(aLanguage),
mEnabled(aEnabled),
mTrackId(aTrackId),
mIsRenderedExternally(false),
mType(aType) {
MOZ_COUNT_CTOR(TrackInfo);
}
// Only used for backward compatibility. Do not use in new code.
void Init(const nsAString& aId, const nsAString& aKind,
const nsAString& aLabel, const nsAString& aLanguage,
bool aEnabled) {
mId = aId;
mKind = aKind;
mLabel = aLabel;
mLanguage = aLanguage;
mEnabled = aEnabled;
}
// Fields common with MediaTrack object.
nsString mId;
nsString mKind;
nsString mLabel;
nsString mLanguage;
bool mEnabled;
uint32_t mTrackId;
nsCString mMimeType;
media::TimeUnit mDuration;
media::TimeUnit mMediaTime;
uint32_t mTimeScale = 0;
CryptoTrack mCrypto;
CopyableTArray<MetadataTag> mTags;
// True if the track is gonna be (decrypted)/decoded and
// rendered directly by non-gecko components.
bool mIsRenderedExternally;
virtual AudioInfo* GetAsAudioInfo() { return nullptr; }
virtual VideoInfo* GetAsVideoInfo() { return nullptr; }
virtual TextInfo* GetAsTextInfo() { return nullptr; }
virtual const AudioInfo* GetAsAudioInfo() const { return nullptr; }
virtual const VideoInfo* GetAsVideoInfo() const { return nullptr; }
virtual const TextInfo* GetAsTextInfo() const { return nullptr; }
bool IsAudio() const { return !!GetAsAudioInfo(); }
bool IsVideo() const { return !!GetAsVideoInfo(); }
bool IsText() const { return !!GetAsTextInfo(); }
TrackType GetType() const { return mType; }
nsCString ToString() const;
bool virtual IsValid() const = 0;
virtual UniquePtr<TrackInfo> Clone() const = 0;
MOZ_COUNTED_DTOR_VIRTUAL(TrackInfo)
protected:
TrackInfo(const TrackInfo& aOther) {
mId = aOther.mId;
mKind = aOther.mKind;
mLabel = aOther.mLabel;
mLanguage = aOther.mLanguage;
mEnabled = aOther.mEnabled;
mTrackId = aOther.mTrackId;
mMimeType = aOther.mMimeType;
mDuration = aOther.mDuration;
mMediaTime = aOther.mMediaTime;
mCrypto = aOther.mCrypto;
mIsRenderedExternally = aOther.mIsRenderedExternally;
mType = aOther.mType;
mTags = aOther.mTags.Clone();
MOZ_COUNT_CTOR(TrackInfo);
}
bool IsEqualTo(const TrackInfo& rhs) const;
private:
TrackType mType;
};
// String version of track type.
const char* TrackTypeToStr(TrackInfo::TrackType aTrack);
enum class VideoRotation {
kDegree_0 = 0,
kDegree_90 = 90,
kDegree_180 = 180,
kDegree_270 = 270,
};
// Stores info relevant to presenting media frames.
class VideoInfo : public TrackInfo {
public:
VideoInfo() : VideoInfo(-1, -1) {}
VideoInfo(int32_t aWidth, int32_t aHeight)
: VideoInfo(gfx::IntSize(aWidth, aHeight)) {}
explicit VideoInfo(const gfx::IntSize& aSize)
: TrackInfo(kVideoTrack, u"2"_ns, u"main"_ns, u""_ns, u""_ns, true, 2),
mDisplay(aSize),
mStereoMode(StereoMode::MONO),
mImage(aSize),
mCodecSpecificConfig(new MediaByteBuffer),
mExtraData(new MediaByteBuffer),
mRotation(VideoRotation::kDegree_0) {}
VideoInfo(const VideoInfo& aOther) : TrackInfo(aOther) {
if (aOther.mCodecSpecificConfig) {
mCodecSpecificConfig = new MediaByteBuffer();
mCodecSpecificConfig->AppendElements(
reinterpret_cast<uint8_t*>(aOther.mCodecSpecificConfig->Elements()),
aOther.mCodecSpecificConfig->Length());
}
if (aOther.mExtraData) {
mExtraData = new MediaByteBuffer();
mExtraData->AppendElements(
reinterpret_cast<uint8_t*>(aOther.mExtraData->Elements()),
aOther.mExtraData->Length());
}
mDisplay = aOther.mDisplay;
mStereoMode = aOther.mStereoMode;
mImage = aOther.mImage;
mRotation = aOther.mRotation;
mColorDepth = aOther.mColorDepth;
mColorSpace = aOther.mColorSpace;
mColorPrimaries = aOther.mColorPrimaries;
mTransferFunction = aOther.mTransferFunction;
mColorRange = aOther.mColorRange;
mImageRect = aOther.mImageRect;
mAlphaPresent = aOther.mAlphaPresent;
mFrameRate = aOther.mFrameRate;
};
bool operator==(const VideoInfo& rhs) const;
bool IsValid() const override {
return mDisplay.width > 0 && mDisplay.height > 0;
}
VideoInfo* GetAsVideoInfo() override { return this; }
const VideoInfo* GetAsVideoInfo() const override { return this; }
UniquePtr<TrackInfo> Clone() const override {
return MakeUnique<VideoInfo>(*this);
}
void SetAlpha(bool aAlphaPresent) { mAlphaPresent = aAlphaPresent; }
bool HasAlpha() const { return mAlphaPresent; }
gfx::IntRect ImageRect() const {
if (!mImageRect) {
return gfx::IntRect(0, 0, mImage.width, mImage.height);
}
return *mImageRect;
}
void SetImageRect(const gfx::IntRect& aRect) { mImageRect = Some(aRect); }
void ResetImageRect() { mImageRect.reset(); }
// Returned the crop rectangle scaled to aWidth/aHeight size relative to
// mImage size.
// If aWidth and aHeight are identical to the original
// mImage.width/mImage.height then the scaling ratio will be 1. This is used
// for when the frame size is different from what the container reports. This
// is legal in WebM, and we will preserve the ratio of the crop rectangle as
// it was reported relative to the picture size reported by the container.
gfx::IntRect ScaledImageRect(int64_t aWidth, int64_t aHeight) const {
if ((aWidth == mImage.width && aHeight == mImage.height) || !mImage.width ||
!mImage.height) {
return ImageRect();
}
gfx::IntRect imageRect = ImageRect();
int64_t w = (aWidth * imageRect.Width()) / mImage.width;
int64_t h = (aHeight * imageRect.Height()) / mImage.height;
if (!w || !h) {
return imageRect;
}
imageRect.x = AssertedCast<int>((imageRect.x * aWidth) / mImage.width);
imageRect.y = AssertedCast<int>((imageRect.y * aHeight) / mImage.height);
imageRect.SetWidth(AssertedCast<int>(w));
imageRect.SetHeight(AssertedCast<int>(h));
return imageRect;
}
VideoRotation ToSupportedRotation(int32_t aDegree) const {
switch (aDegree) {
case 90:
return VideoRotation::kDegree_90;
case 180:
return VideoRotation::kDegree_180;
case 270:
return VideoRotation::kDegree_270;
default:
NS_WARNING_ASSERTION(aDegree == 0, "Invalid rotation degree, ignored");
return VideoRotation::kDegree_0;
}
}
nsString ToString() const {
std::array YUVColorSpaceStrings = {"BT601", "BT709", "BT2020", "Identity",
"Default"};
std::array ColorDepthStrings = {
"COLOR_8",
"COLOR_10",
"COLOR_12",
"COLOR_16",
};
std::array TransferFunctionStrings = {
"BT709",
"SRGB",
"PQ",
"HLG",
};
std::array ColorRangeStrings = {
"LIMITED",
"FULL",
};
std::array ColorPrimariesStrings = {"Display",
"UNKNOWN"
"SRGB",
"DISPLAY_P3",
"BT601_525",
"BT709",
"BT601_625"
"BT709",
"BT2020"};
nsString rv;
rv.AppendLiteral(u"VideoInfo: ");
rv.AppendPrintf("display size: %dx%d ", mDisplay.Width(),
mDisplay.Height());
rv.AppendPrintf("stereo mode: %d", static_cast<int>(mStereoMode));
rv.AppendPrintf("image size: %dx%d ", mImage.Width(), mImage.Height());
if (mCodecSpecificConfig) {
rv.AppendPrintf("codec specific config: %zu bytes",
mCodecSpecificConfig->Length());
}
if (mExtraData) {
rv.AppendPrintf("extra data: %zu bytes", mExtraData->Length());
}
rv.AppendPrintf("rotation: %d", static_cast<int>(mRotation));
rv.AppendPrintf("colors: %s",
ColorDepthStrings[static_cast<int>(mColorDepth)]);
if (mColorSpace) {
rv.AppendPrintf(
"YUV colorspace: %s ",
YUVColorSpaceStrings[static_cast<int>(mColorSpace.value())]);
}
if (mColorPrimaries) {
rv.AppendPrintf(
"color primaries: %s ",
ColorPrimariesStrings[static_cast<int>(mColorPrimaries.value())]);
}
if (mTransferFunction) {
rv.AppendPrintf(
"transfer function %s ",
TransferFunctionStrings[static_cast<int>(mTransferFunction.value())]);
}
rv.AppendPrintf("color range: %s",
ColorRangeStrings[static_cast<int>(mColorRange)]);
if (mImageRect) {
rv.AppendPrintf("image rect: %dx%d", mImageRect->Width(),
mImageRect->Height());
}
rv.AppendPrintf("alpha present: %s", mAlphaPresent ? "true" : "false");
if (mFrameRate) {
rv.AppendPrintf("frame rate: %dHz", mFrameRate.value());
}
return rv;
}
// Size in pixels at which the video is rendered. This is after it has
// been scaled by its aspect ratio.
gfx::IntSize mDisplay;
// Indicates the frame layout for single track stereo videos.
StereoMode mStereoMode;
// Size of the decoded video's image.
gfx::IntSize mImage;
RefPtr<MediaByteBuffer> mCodecSpecificConfig;
RefPtr<MediaByteBuffer> mExtraData;
// Describing how many degrees video frames should be rotated in clock-wise to
// get correct view.
VideoRotation mRotation;
// Should be 8, 10 or 12. Default value is 8.
gfx::ColorDepth mColorDepth = gfx::ColorDepth::COLOR_8;
// Matrix coefficients (if specified by the video) imply a colorspace.
Maybe<gfx::YUVColorSpace> mColorSpace;
// Color primaries are independent from the coefficients.
Maybe<gfx::ColorSpace2> mColorPrimaries;
// Transfer functions get their own member, which may not be strongly
// correlated to the colorspace.
Maybe<gfx::TransferFunction> mTransferFunction;
// True indicates no restriction on Y, U, V values (otherwise 16-235 for 8
// bits etc)
gfx::ColorRange mColorRange = gfx::ColorRange::LIMITED;
Maybe<int32_t> GetFrameRate() const { return mFrameRate; }
void SetFrameRate(int32_t aRate) { mFrameRate = Some(aRate); }
private:
friend struct IPC::ParamTraits<VideoInfo>;
// mImage may be cropped; currently only used with the WebM container.
// If unset, no cropping is to occur.
Maybe<gfx::IntRect> mImageRect;
// Indicates whether or not frames may contain alpha information.
bool mAlphaPresent = false;
Maybe<int32_t> mFrameRate;
};
class AudioInfo : public TrackInfo {
public:
AudioInfo()
: TrackInfo(kAudioTrack, u"1"_ns, u"main"_ns, u""_ns, u""_ns, true, 1),
mRate(0),
mChannels(0),
mChannelMap(AudioConfig::ChannelLayout::UNKNOWN_MAP),
mBitDepth(0),
mProfile(0),
mExtendedProfile(0) {}
AudioInfo(const AudioInfo& aOther) = default;
bool operator==(const AudioInfo& rhs) const;
static const uint32_t MAX_RATE = 768000;
static const uint32_t MAX_CHANNEL_COUNT = 256;
bool IsValid() const override {
return mChannels > 0 && mChannels <= MAX_CHANNEL_COUNT && mRate > 0 &&
mRate <= MAX_RATE;
}
AudioInfo* GetAsAudioInfo() override { return this; }
const AudioInfo* GetAsAudioInfo() const override { return this; }
nsCString ToString() const;
UniquePtr<TrackInfo> Clone() const override {
return MakeUnique<AudioInfo>(*this);
}
// Sample rate.
uint32_t mRate;
// Number of audio channels.
uint32_t mChannels;
// The AudioConfig::ChannelLayout map. Channels are ordered as per SMPTE
// definition. A value of UNKNOWN_MAP indicates unknown layout.
// ChannelMap is an unsigned bitmap compatible with Windows' WAVE and FFmpeg
// channel map.
AudioConfig::ChannelLayout::ChannelMap mChannelMap;
// Bits per sample.
uint32_t mBitDepth;
// Codec profile.
uint8_t mProfile;
// Extended codec profile.
uint8_t mExtendedProfile;
AudioCodecSpecificVariant mCodecSpecificConfig{NoCodecSpecificData{}};
};
class EncryptionInfo {
public:
EncryptionInfo() : mEncrypted(false) {}
struct InitData {
template <typename AInitDatas>
InitData(const nsAString& aType, AInitDatas&& aInitData)
: mType(aType), mInitData(std::forward<AInitDatas>(aInitData)) {}
// Encryption type to be passed to JS. Usually `cenc'.
nsString mType;
// Encryption data.
CopyableTArray<uint8_t> mInitData;
};
using InitDatas = CopyableTArray<InitData>;
// True if the stream has encryption metadata
bool IsEncrypted() const { return mEncrypted; }
void Reset() {
mEncrypted = false;
mInitDatas.Clear();
}
template <typename AInitDatas>
void AddInitData(const nsAString& aType, AInitDatas&& aInitData) {
mInitDatas.AppendElement(
InitData(aType, std::forward<AInitDatas>(aInitData)));
mEncrypted = true;
}
void AddInitData(const EncryptionInfo& aInfo) {
mInitDatas.AppendElements(aInfo.mInitDatas);
mEncrypted = !!mInitDatas.Length();
}
// One 'InitData' per encrypted buffer.
InitDatas mInitDatas;
private:
bool mEncrypted;
};
class MediaInfo {
public:
bool HasVideo() const { return mVideo.IsValid(); }
void EnableVideo() {
if (HasVideo()) {
return;
}
// Set dummy values so that HasVideo() will return true;
// See VideoInfo::IsValid()
mVideo.mDisplay = gfx::IntSize(1, 1);
}
bool HasAudio() const { return mAudio.IsValid(); }
void EnableAudio() {
if (HasAudio()) {
return;
}
// Set dummy values so that HasAudio() will return true;
// See AudioInfo::IsValid()
mAudio.mChannels = 2;
mAudio.mRate = 44100;
}
bool IsEncrypted() const {
return (HasAudio() && mAudio.mCrypto.IsEncrypted()) ||
(HasVideo() && mVideo.mCrypto.IsEncrypted());
}
bool HasValidMedia() const { return HasVideo() || HasAudio(); }
// TODO: Store VideoInfo and AudioIndo in arrays to support multi-tracks.
VideoInfo mVideo;
AudioInfo mAudio;
// If the metadata includes a duration, we store it here.
media::NullableTimeUnit mMetadataDuration;
// The Ogg reader tries to kinda-sorta compute the duration by seeking to the
// end and determining the timestamp of the last frame. This isn't useful as
// a duration until we know the start time, so we need to track it separately.
media::NullableTimeUnit mUnadjustedMetadataEndTime;
// True if the media is seekable (i.e. supports random access).
bool mMediaSeekable = true;
// True if the media is only seekable within its buffered ranges.
bool mMediaSeekableOnlyInBufferedRanges = false;
EncryptionInfo mCrypto;
// The minimum of start times of audio and video tracks.
// Use to map the zero time on the media timeline to the first frame.
media::TimeUnit mStartTime;
};
class TrackInfoSharedPtr {
NS_INLINE_DECL_THREADSAFE_REFCOUNTING(TrackInfoSharedPtr)
public:
TrackInfoSharedPtr(const TrackInfo& aOriginal, uint32_t aStreamID)
: mInfo(aOriginal.Clone()),
mStreamSourceID(aStreamID),
mMimeType(mInfo->mMimeType) {}
uint32_t GetID() const { return mStreamSourceID; }
operator const TrackInfo*() const { return mInfo.get(); }
const TrackInfo* operator*() const { return mInfo.get(); }
const TrackInfo* operator->() const {
MOZ_ASSERT(mInfo.get(), "dereferencing a UniquePtr containing nullptr");
return mInfo.get();
}
const AudioInfo* GetAsAudioInfo() const {
return mInfo ? mInfo->GetAsAudioInfo() : nullptr;
}
const VideoInfo* GetAsVideoInfo() const {
return mInfo ? mInfo->GetAsVideoInfo() : nullptr;
}
const TextInfo* GetAsTextInfo() const {
return mInfo ? mInfo->GetAsTextInfo() : nullptr;
}
private:
~TrackInfoSharedPtr() = default;
UniquePtr<TrackInfo> mInfo;
// A unique ID, guaranteed to change when changing streams.
uint32_t mStreamSourceID;
public:
const nsCString& mMimeType;
};
} // namespace mozilla
#endif // MediaInfo_h
|