1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
|
/* This Source Code Form is subject to the terms of the Mozilla Public
* License, v. 2.0. If a copy of the MPL was not distributed with this
* file, You can obtain one at http://mozilla.org/MPL/2.0/. */
#include "MFMediaEngineVideoStream.h"
#include "mozilla/layers/DcompSurfaceImage.h"
#include "MFMediaEngineUtils.h"
#include "mozilla/StaticPrefs_media.h"
namespace mozilla {
#define LOG(msg, ...) \
MOZ_LOG(gMFMediaEngineLog, LogLevel::Debug, \
("MFMediaStream=%p (%s), " msg, this, \
this->GetDescriptionName().get(), ##__VA_ARGS__))
#define LOGV(msg, ...) \
MOZ_LOG(gMFMediaEngineLog, LogLevel::Verbose, \
("MFMediaStream=%p (%s), " msg, this, \
this->GetDescriptionName().get(), ##__VA_ARGS__))
using Microsoft::WRL::ComPtr;
using Microsoft::WRL::MakeAndInitialize;
/* static */
MFMediaEngineVideoStream* MFMediaEngineVideoStream::Create(
uint64_t aStreamId, const TrackInfo& aInfo, MFMediaSource* aParentSource) {
MFMediaEngineVideoStream* stream;
MOZ_ASSERT(aInfo.IsVideo());
if (FAILED(MakeAndInitialize<MFMediaEngineVideoStream>(
&stream, aStreamId, aInfo, aParentSource))) {
return nullptr;
}
stream->mStreamType =
GetStreamTypeFromMimeType(aInfo.GetAsVideoInfo()->mMimeType);
MOZ_ASSERT(StreamTypeIsVideo(stream->mStreamType));
stream->mHasReceivedInitialCreateDecoderConfig = false;
stream->SetDCompSurfaceHandle(INVALID_HANDLE_VALUE, gfx::IntSize{});
return stream;
}
void MFMediaEngineVideoStream::SetKnowsCompositor(
layers::KnowsCompositor* aKnowsCompositor) {
ComPtr<MFMediaEngineVideoStream> self = this;
Unused << mTaskQueue->Dispatch(NS_NewRunnableFunction(
"MFMediaEngineStream::SetKnowsCompositor",
[self, knowCompositor = RefPtr<layers::KnowsCompositor>{aKnowsCompositor},
this]() {
mKnowsCompositor = knowCompositor;
LOG("Set SetKnowsCompositor=%p", mKnowsCompositor.get());
ResolvePendingDrainPromiseIfNeeded();
}));
}
void MFMediaEngineVideoStream::SetDCompSurfaceHandle(HANDLE aDCompSurfaceHandle,
gfx::IntSize aDisplay) {
ComPtr<MFMediaEngineVideoStream> self = this;
Unused << mTaskQueue->Dispatch(NS_NewRunnableFunction(
"MFMediaEngineStream::SetDCompSurfaceHandle",
[self, aDCompSurfaceHandle, aDisplay, this]() {
if (mDCompSurfaceHandle == aDCompSurfaceHandle) {
return;
}
mDCompSurfaceHandle = aDCompSurfaceHandle;
mNeedRecreateImage = true;
{
MutexAutoLock lock(mMutex);
if (aDCompSurfaceHandle != INVALID_HANDLE_VALUE &&
aDisplay != mDisplay) {
LOG("Update display [%dx%d] -> [%dx%d]", mDisplay.Width(),
mDisplay.Height(), aDisplay.Width(), aDisplay.Height());
mDisplay = aDisplay;
}
}
LOG("Set DCompSurfaceHandle, handle=%p", mDCompSurfaceHandle);
ResolvePendingDrainPromiseIfNeeded();
}));
}
HRESULT MFMediaEngineVideoStream::CreateMediaType(const TrackInfo& aInfo,
IMFMediaType** aMediaType) {
auto& videoInfo = *aInfo.GetAsVideoInfo();
mIsEncrypted = videoInfo.mCrypto.IsEncrypted();
GUID subType = VideoMimeTypeToMediaFoundationSubtype(videoInfo.mMimeType);
NS_ENSURE_TRUE(subType != GUID_NULL, MF_E_TOPO_CODEC_NOT_FOUND);
// https://docs.microsoft.com/en-us/windows/win32/medfound/media-type-attributes
ComPtr<IMFMediaType> mediaType;
RETURN_IF_FAILED(wmf::MFCreateMediaType(&mediaType));
RETURN_IF_FAILED(mediaType->SetGUID(MF_MT_MAJOR_TYPE, MFMediaType_Video));
RETURN_IF_FAILED(mediaType->SetGUID(MF_MT_SUBTYPE, subType));
const auto& image = videoInfo.mImage;
UINT32 imageWidth = image.Width();
UINT32 imageHeight = image.Height();
RETURN_IF_FAILED(MFSetAttributeSize(mediaType.Get(), MF_MT_FRAME_SIZE,
imageWidth, imageHeight));
UINT32 displayWidth = videoInfo.mDisplay.Width();
UINT32 displayHeight = videoInfo.mDisplay.Height();
{
MutexAutoLock lock(mMutex);
mDisplay = videoInfo.mDisplay;
}
// PAR = DAR / SAR = (DW / DH) / (SW / SH) = (DW * SH) / (DH * SW)
RETURN_IF_FAILED(MFSetAttributeRatio(
mediaType.Get(), MF_MT_PIXEL_ASPECT_RATIO, displayWidth * imageHeight,
displayHeight * imageWidth));
// https://docs.microsoft.com/en-us/windows/win32/api/mfobjects/ns-mfobjects-mfoffset
// The value of the MFOffset number is value + (fract / 65536.0f).
static const auto ToMFOffset = [](float aValue) {
MFOffset offset;
offset.value = static_cast<short>(aValue);
offset.fract = static_cast<WORD>(65536 * (aValue - offset.value));
return offset;
};
MFVideoArea area;
area.OffsetX = ToMFOffset(videoInfo.ImageRect().x);
area.OffsetY = ToMFOffset(videoInfo.ImageRect().y);
area.Area = {(LONG)imageWidth, (LONG)imageHeight};
RETURN_IF_FAILED(mediaType->SetBlob(MF_MT_GEOMETRIC_APERTURE, (UINT8*)&area,
sizeof(area)));
// https://docs.microsoft.com/en-us/windows/win32/api/mfapi/ne-mfapi-mfvideorotationformat
static const auto ToMFVideoRotationFormat =
[](VideoInfo::Rotation aRotation) {
using Rotation = VideoInfo::Rotation;
switch (aRotation) {
case Rotation::kDegree_0:
return MFVideoRotationFormat_0;
case Rotation::kDegree_90:
return MFVideoRotationFormat_90;
case Rotation::kDegree_180:
return MFVideoRotationFormat_180;
default:
MOZ_ASSERT(aRotation == Rotation::kDegree_270);
return MFVideoRotationFormat_270;
}
};
const auto rotation = ToMFVideoRotationFormat(videoInfo.mRotation);
RETURN_IF_FAILED(mediaType->SetUINT32(MF_MT_VIDEO_ROTATION, rotation));
static const auto ToMFVideoTransFunc =
[](const Maybe<gfx::YUVColorSpace>& aColorSpace) {
using YUVColorSpace = gfx::YUVColorSpace;
if (!aColorSpace) {
return MFVideoTransFunc_Unknown;
}
// https://docs.microsoft.com/en-us/windows/win32/api/mfobjects/ne-mfobjects-mfvideotransferfunction
switch (*aColorSpace) {
case YUVColorSpace::BT601:
case YUVColorSpace::BT709:
return MFVideoTransFunc_709;
case YUVColorSpace::BT2020:
return MFVideoTransFunc_2020;
case YUVColorSpace::Identity:
return MFVideoTransFunc_sRGB;
default:
return MFVideoTransFunc_Unknown;
}
};
const auto transFunc = ToMFVideoTransFunc(videoInfo.mColorSpace);
RETURN_IF_FAILED(mediaType->SetUINT32(MF_MT_TRANSFER_FUNCTION, transFunc));
static const auto ToMFVideoPrimaries =
[](const Maybe<gfx::YUVColorSpace>& aColorSpace) {
using YUVColorSpace = gfx::YUVColorSpace;
if (!aColorSpace) {
return MFVideoPrimaries_Unknown;
}
// https://docs.microsoft.com/en-us/windows/win32/api/mfobjects/ne-mfobjects-mfvideoprimaries
switch (*aColorSpace) {
case YUVColorSpace::BT601:
return MFVideoPrimaries_Unknown;
case YUVColorSpace::BT709:
return MFVideoPrimaries_BT709;
case YUVColorSpace::BT2020:
return MFVideoPrimaries_BT2020;
case YUVColorSpace::Identity:
return MFVideoPrimaries_BT709;
default:
return MFVideoPrimaries_Unknown;
}
};
const auto videoPrimaries = ToMFVideoPrimaries(videoInfo.mColorSpace);
RETURN_IF_FAILED(mediaType->SetUINT32(MF_MT_VIDEO_PRIMARIES, videoPrimaries));
LOG("Created video type, subtype=%s, image=[%ux%u], display=[%ux%u], "
"rotation=%s, tranFuns=%s, primaries=%s, encrypted=%d",
GUIDToStr(subType), imageWidth, imageHeight, displayWidth, displayHeight,
MFVideoRotationFormatToStr(rotation),
MFVideoTransferFunctionToStr(transFunc),
MFVideoPrimariesToStr(videoPrimaries), mIsEncrypted);
if (IsEncrypted()) {
ComPtr<IMFMediaType> protectedMediaType;
RETURN_IF_FAILED(wmf::MFWrapMediaType(mediaType.Get(),
MFMediaType_Protected, subType,
protectedMediaType.GetAddressOf()));
LOG("Wrap MFMediaType_Video into MFMediaType_Protected");
*aMediaType = protectedMediaType.Detach();
} else {
*aMediaType = mediaType.Detach();
}
return S_OK;
}
bool MFMediaEngineVideoStream::HasEnoughRawData() const {
// If more than this much raw video is queued, we'll hold off request more
// video.
return mRawDataQueueForFeedingEngine.Duration() >=
StaticPrefs::media_wmf_media_engine_raw_data_threshold_video();
}
bool MFMediaEngineVideoStream::IsDCompImageReady() {
AssertOnTaskQueue();
if (!mDCompSurfaceHandle || mDCompSurfaceHandle == INVALID_HANDLE_VALUE) {
LOGV("Can't create image without a valid dcomp surface handle");
return false;
}
if (!mKnowsCompositor) {
LOGV("Can't create image without the knows compositor");
return false;
}
if (!mDcompSurfaceImage || mNeedRecreateImage) {
MutexAutoLock lock(mMutex);
// DirectComposition only supports RGBA. We use DXGI_FORMAT_B8G8R8A8_UNORM
// as a default because we can't know what format the dcomp surface is.
// https://docs.microsoft.com/en-us/windows/win32/api/dcomp/nf-dcomp-idcompositionsurfacefactory-createsurface
mDcompSurfaceImage = new layers::DcompSurfaceImage(
mDCompSurfaceHandle, mDisplay, gfx::SurfaceFormat::B8G8R8A8,
mKnowsCompositor);
mNeedRecreateImage = false;
LOG("Created dcomp surface image, handle=%p, size=[%u,%u]",
mDCompSurfaceHandle, mDisplay.Width(), mDisplay.Height());
}
return true;
}
already_AddRefed<MediaData> MFMediaEngineVideoStream::OutputDataInternal() {
AssertOnTaskQueue();
if (mRawDataQueueForGeneratingOutput.GetSize() == 0 || !IsDCompImageReady()) {
return nullptr;
}
RefPtr<MediaRawData> sample = mRawDataQueueForGeneratingOutput.PopFront();
RefPtr<VideoData> output;
{
MutexAutoLock lock(mMutex);
output = VideoData::CreateFromImage(
mDisplay, sample->mOffset, sample->mTime, sample->mDuration,
mDcompSurfaceImage, sample->mKeyframe, sample->mTimecode);
}
return output.forget();
}
RefPtr<MediaDataDecoder::DecodePromise> MFMediaEngineVideoStream::Drain() {
AssertOnTaskQueue();
MediaDataDecoder::DecodedData outputs;
if (!IsDCompImageReady()) {
LOGV("Waiting for dcomp image for draining");
return mPendingDrainPromise.Ensure(__func__);
}
return MFMediaEngineStream::Drain();
}
void MFMediaEngineVideoStream::ResolvePendingDrainPromiseIfNeeded() {
AssertOnTaskQueue();
if (mPendingDrainPromise.IsEmpty()) {
return;
}
if (!IsDCompImageReady()) {
return;
}
MediaDataDecoder::DecodedData outputs;
while (RefPtr<MediaData> outputData = OutputDataInternal()) {
outputs.AppendElement(outputData);
LOGV("Output data [%" PRId64 ",%" PRId64 "]",
outputData->mTime.ToMicroseconds(),
outputData->GetEndTime().ToMicroseconds());
}
mPendingDrainPromise.Resolve(std::move(outputs), __func__);
LOG("Resolved pending drain promise");
}
MediaDataDecoder::ConversionRequired MFMediaEngineVideoStream::NeedsConversion()
const {
return mStreamType == WMFStreamType::H264
? MediaDataDecoder::ConversionRequired::kNeedAnnexB
: MediaDataDecoder::ConversionRequired::kNeedNone;
}
void MFMediaEngineVideoStream::SetConfig(const TrackInfo& aConfig) {
MOZ_ASSERT(aConfig.IsVideo());
ComPtr<MFMediaEngineStream> self = this;
Unused << mTaskQueue->Dispatch(
NS_NewRunnableFunction("MFMediaEngineStream::SetConfig",
[self, info = *aConfig.GetAsVideoInfo(), this]() {
if (mHasReceivedInitialCreateDecoderConfig) {
// Here indicating a new config for video,
// which is triggered by the media change
// monitor, so we need to update the config.
UpdateConfig(info);
}
mHasReceivedInitialCreateDecoderConfig = true;
}));
}
void MFMediaEngineVideoStream::UpdateConfig(const VideoInfo& aInfo) {
AssertOnTaskQueue();
// Disable explicit format change event for H264 to allow switching to the
// new stream without a full re-create, which will be much faster. This is
// also due to the fact that the MFT decoder can handle some format changes
// without a format change event. For format changes that the MFT decoder
// cannot support (e.g. codec change), the playback will fail later with
// MF_E_INVALIDMEDIATYPE (0xC00D36B4).
if (mStreamType == WMFStreamType::H264) {
return;
}
LOG("Video config changed, will update stream descriptor");
PROFILER_MARKER_TEXT("VideoConfigChange", MEDIA_PLAYBACK, {},
nsPrintfCString("stream=%s, id=%" PRIu64,
GetDescriptionName().get(), mStreamId));
ComPtr<IMFMediaType> mediaType;
RETURN_VOID_IF_FAILED(CreateMediaType(aInfo, mediaType.GetAddressOf()));
RETURN_VOID_IF_FAILED(GenerateStreamDescriptor(mediaType));
RETURN_VOID_IF_FAILED(mMediaEventQueue->QueueEventParamUnk(
MEStreamFormatChanged, GUID_NULL, S_OK, mediaType.Get()));
}
void MFMediaEngineVideoStream::ShutdownCleanUpOnTaskQueue() {
AssertOnTaskQueue();
mPendingDrainPromise.RejectIfExists(NS_ERROR_DOM_MEDIA_CANCELED, __func__);
}
bool MFMediaEngineVideoStream::IsEnded() const {
AssertOnTaskQueue();
// If a video only contains one frame, the media engine won't return a decoded
// frame before we tell it the track is already ended. However, due to the
// constraint of our media pipeline, the format reader won't notify EOS until
// the draining finishes, which causes a deadlock. Therefore, we would
// consider having pending drain promise as a sign of EOS as well, in order to
// get the decoded frame and revolve the drain promise.
return (mReceivedEOS || !mPendingDrainPromise.IsEmpty()) &&
mRawDataQueueForFeedingEngine.GetSize() == 0;
}
bool MFMediaEngineVideoStream::IsEncrypted() const { return mIsEncrypted; }
nsCString MFMediaEngineVideoStream::GetCodecName() const {
switch (mStreamType) {
case WMFStreamType::H264:
return "h264"_ns;
case WMFStreamType::VP8:
return "vp8"_ns;
case WMFStreamType::VP9:
return "vp9"_ns;
case WMFStreamType::AV1:
return "av1"_ns;
default:
return "unknown"_ns;
};
}
#undef LOG
#undef LOGV
} // namespace mozilla
|