diff options
Diffstat (limited to 'dom/media/gtest/TestMediaDataEncoder.cpp')
-rw-r--r-- | dom/media/gtest/TestMediaDataEncoder.cpp | 510 |
1 files changed, 510 insertions, 0 deletions
diff --git a/dom/media/gtest/TestMediaDataEncoder.cpp b/dom/media/gtest/TestMediaDataEncoder.cpp new file mode 100644 index 0000000000..7bc975816f --- /dev/null +++ b/dom/media/gtest/TestMediaDataEncoder.cpp @@ -0,0 +1,510 @@ +/* -*- Mode: C++; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*- */ +/* This Source Code Form is subject to the terms of the Mozilla Public + * License, v. 2.0. If a copy of the MPL was not distributed with this + * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ + +#include "gtest/gtest.h" + +#include "AnnexB.h" +#include "ImageContainer.h" +#include "mozilla/AbstractThread.h" +#include "mozilla/SpinEventLoopUntil.h" +#include "mozilla/media/MediaUtils.h" // For media::Await +#include "nsMimeTypes.h" +#include "PEMFactory.h" +#include "TimeUnits.h" +#include "VideoUtils.h" +#include "VPXDecoder.h" +#include <algorithm> + +#include <fstream> + +#ifdef XP_WIN +#include "mozilla/WindowsVersion.h" +#endif + +#define RUN_IF_SUPPORTED(mimeType, test) \ + do { \ + if (!isWin7()) { \ + RefPtr<PEMFactory> f(new PEMFactory()); \ + if (f->SupportsMimeType(nsLiteralCString(mimeType))) { \ + test(); \ + } \ + } \ + } while (0) + +#define BLOCK_SIZE 64 +#define WIDTH 640 +#define HEIGHT 480 +#define NUM_FRAMES 150UL +#define FRAME_RATE 30 +#define FRAME_DURATION (1000000 / FRAME_RATE) +#define BIT_RATE (1000 * 1000) // 1Mbps +#define KEYFRAME_INTERVAL FRAME_RATE // 1 keyframe per second +#define VIDEO_VP8 "video/vp8" +#define VIDEO_VP9 "video/vp9" + +using namespace mozilla; + +static gfx::IntSize kImageSize(WIDTH, HEIGHT); + +class MediaDataEncoderTest : public testing::Test { + protected: + void SetUp() override { mData.Init(kImageSize); } + + void TearDown() override { mData.Deinit(); } + + public: + struct FrameSource final { + layers::PlanarYCbCrData mYUV; + UniquePtr<uint8_t[]> mBuffer; + RefPtr<layers::BufferRecycleBin> mRecycleBin; + int16_t mColorStep = 4; + + void Init(const gfx::IntSize& aSize) { + mYUV.mPictureRect = gfx::IntRect(0, 0, aSize.width, aSize.height); + mYUV.mYStride = aSize.width; + mYUV.mCbCrStride = (aSize.width + 1) / 2; + mYUV.mChromaSubsampling = gfx::ChromaSubsampling::HALF_WIDTH_AND_HEIGHT; + auto ySize = mYUV.YDataSize(); + auto cbcrSize = mYUV.CbCrDataSize(); + size_t bufferSize = + mYUV.mYStride * ySize.height + 2 * mYUV.mCbCrStride * cbcrSize.height; + mBuffer = MakeUnique<uint8_t[]>(bufferSize); + std::fill_n(mBuffer.get(), bufferSize, 0x7F); + mYUV.mYChannel = mBuffer.get(); + mYUV.mCbChannel = mYUV.mYChannel + mYUV.mYStride * ySize.height; + mYUV.mCrChannel = mYUV.mCbChannel + mYUV.mCbCrStride * cbcrSize.height; + mYUV.mChromaSubsampling = gfx::ChromaSubsampling::HALF_WIDTH_AND_HEIGHT; + mRecycleBin = new layers::BufferRecycleBin(); + } + + void Deinit() { + mBuffer.reset(); + mRecycleBin = nullptr; + } + + already_AddRefed<MediaData> GetFrame(const size_t aIndex) { + Draw(aIndex); + RefPtr<layers::PlanarYCbCrImage> img = + new layers::RecyclingPlanarYCbCrImage(mRecycleBin); + img->CopyData(mYUV); + RefPtr<MediaData> frame = VideoData::CreateFromImage( + kImageSize, 0, + media::TimeUnit::FromMicroseconds(aIndex * FRAME_DURATION), + media::TimeUnit::FromMicroseconds(FRAME_DURATION), img, + (aIndex & 0xF) == 0, + media::TimeUnit::FromMicroseconds(aIndex * FRAME_DURATION)); + return frame.forget(); + } + + void DrawChessboard(uint8_t* aAddr, const size_t aWidth, + const size_t aHeight, const size_t aOffset) { + uint8_t pixels[2][BLOCK_SIZE]; + size_t x = aOffset % BLOCK_SIZE; + if ((aOffset / BLOCK_SIZE) & 1) { + x = BLOCK_SIZE - x; + } + for (size_t i = 0; i < x; i++) { + pixels[0][i] = 0x00; + pixels[1][i] = 0xFF; + } + for (size_t i = x; i < BLOCK_SIZE; i++) { + pixels[0][i] = 0xFF; + pixels[1][i] = 0x00; + } + + uint8_t* p = aAddr; + for (size_t row = 0; row < aHeight; row++) { + for (size_t col = 0; col < aWidth; col += BLOCK_SIZE) { + memcpy(p, pixels[((row / BLOCK_SIZE) + (col / BLOCK_SIZE)) % 2], + BLOCK_SIZE); + p += BLOCK_SIZE; + } + } + } + + void Draw(const size_t aIndex) { + auto ySize = mYUV.YDataSize(); + DrawChessboard(mYUV.mYChannel, ySize.width, ySize.height, aIndex << 1); + int16_t color = mYUV.mCbChannel[0] + mColorStep; + if (color > 255 || color < 0) { + mColorStep = -mColorStep; + color = mYUV.mCbChannel[0] + mColorStep; + } + + size_t size = (mYUV.mCrChannel - mYUV.mCbChannel); + + std::fill_n(mYUV.mCbChannel, size, static_cast<uint8_t>(color)); + std::fill_n(mYUV.mCrChannel, size, 0xFF - static_cast<uint8_t>(color)); + } + }; + + public: + FrameSource mData; +}; + +template <typename T> +already_AddRefed<MediaDataEncoder> CreateVideoEncoder( + const char* aMimeType, MediaDataEncoder::Usage aUsage, + MediaDataEncoder::PixelFormat aPixelFormat, int32_t aWidth, int32_t aHeight, + const Maybe<T>& aSpecific) { + RefPtr<PEMFactory> f(new PEMFactory()); + + if (!f->SupportsMimeType(nsCString(aMimeType))) { + return nullptr; + } + + VideoInfo videoInfo(aWidth, aHeight); + videoInfo.mMimeType = nsCString(aMimeType); + const RefPtr<TaskQueue> taskQueue( + TaskQueue::Create(GetMediaThreadPool(MediaThreadType::PLATFORM_ENCODER), + "TestMediaDataEncoder")); + + RefPtr<MediaDataEncoder> e; +#ifdef MOZ_WIDGET_ANDROID + const bool hardwareNotAllowed = false; +#else + const bool hardwareNotAllowed = true; +#endif + if (aSpecific) { + e = f->CreateEncoder( + CreateEncoderParams(videoInfo /* track info */, aUsage, taskQueue, + aPixelFormat, FRAME_RATE /* FPS */, + KEYFRAME_INTERVAL /* keyframe interval */, + BIT_RATE /* bitrate */, aSpecific.value()), + hardwareNotAllowed); + } else { + e = f->CreateEncoder( + CreateEncoderParams(videoInfo /* track info */, aUsage, taskQueue, + aPixelFormat, FRAME_RATE /* FPS */, + KEYFRAME_INTERVAL /* keyframe interval */, + BIT_RATE /* bitrate */), + hardwareNotAllowed); + } + + return e.forget(); +} + +static already_AddRefed<MediaDataEncoder> CreateH264Encoder( + MediaDataEncoder::Usage aUsage = MediaDataEncoder::Usage::Realtime, + MediaDataEncoder::PixelFormat aPixelFormat = + MediaDataEncoder::PixelFormat::YUV420P, + int32_t aWidth = WIDTH, int32_t aHeight = HEIGHT, + const Maybe<MediaDataEncoder::H264Specific>& aSpecific = + Some(MediaDataEncoder::H264Specific( + MediaDataEncoder::H264Specific::ProfileLevel::BaselineAutoLevel))) { + return CreateVideoEncoder(VIDEO_MP4, aUsage, aPixelFormat, aWidth, aHeight, + aSpecific); +} + +void WaitForShutdown(RefPtr<MediaDataEncoder> aEncoder) { + MOZ_ASSERT(aEncoder); + + Maybe<bool> result; + // media::Await() supports exclusive promises only, but ShutdownPromise is + // not. + aEncoder->Shutdown()->Then( + AbstractThread::MainThread(), __func__, + [&result](bool rv) { + EXPECT_TRUE(rv); + result = Some(true); + }, + []() { FAIL() << "Shutdown should never be rejected"; }); + SpinEventLoopUntil("TestMediaDataEncoder.cpp:WaitForShutdown"_ns, + [&result]() { return result; }); +} + +bool isWin7() { + #ifdef XP_WIN + if (!IsWin8OrLater()) { + return true; + } + #endif + return false; +} + +TEST_F(MediaDataEncoderTest, H264Create) { + RUN_IF_SUPPORTED(VIDEO_MP4, []() { + RefPtr<MediaDataEncoder> e = CreateH264Encoder(); + EXPECT_TRUE(e); + WaitForShutdown(e); + }); +} + +static bool EnsureInit(RefPtr<MediaDataEncoder> aEncoder) { + if (!aEncoder) { + return false; + } + + bool succeeded; + media::Await( + GetMediaThreadPool(MediaThreadType::SUPERVISOR), aEncoder->Init(), + [&succeeded](TrackInfo::TrackType t) { + EXPECT_EQ(TrackInfo::TrackType::kVideoTrack, t); + succeeded = true; + }, + [&succeeded](MediaResult r) { succeeded = false; }); + return succeeded; +} + +TEST_F(MediaDataEncoderTest, H264Inits) { + RUN_IF_SUPPORTED(VIDEO_MP4, []() { + // w/o codec specific. + RefPtr<MediaDataEncoder> e = CreateH264Encoder( + MediaDataEncoder::Usage::Realtime, + MediaDataEncoder::PixelFormat::YUV420P, WIDTH, HEIGHT, Nothing()); + EXPECT_TRUE(EnsureInit(e)); + WaitForShutdown(e); + + // w/ codec specific + e = CreateH264Encoder(); + EXPECT_TRUE(EnsureInit(e)); + WaitForShutdown(e); + }); +} + +static MediaDataEncoder::EncodedData Encode( + const RefPtr<MediaDataEncoder> aEncoder, const size_t aNumFrames, + MediaDataEncoderTest::FrameSource& aSource) { + MediaDataEncoder::EncodedData output; + bool succeeded; + for (size_t i = 0; i < aNumFrames; i++) { + RefPtr<MediaData> frame = aSource.GetFrame(i); + media::Await( + GetMediaThreadPool(MediaThreadType::SUPERVISOR), + aEncoder->Encode(frame), + [&output, &succeeded](MediaDataEncoder::EncodedData encoded) { + output.AppendElements(std::move(encoded)); + succeeded = true; + }, + [&succeeded](MediaResult r) { succeeded = false; }); + EXPECT_TRUE(succeeded); + if (!succeeded) { + return output; + } + } + + size_t pending = 0; + do { + media::Await( + GetMediaThreadPool(MediaThreadType::SUPERVISOR), aEncoder->Drain(), + [&pending, &output, &succeeded](MediaDataEncoder::EncodedData encoded) { + pending = encoded.Length(); + output.AppendElements(std::move(encoded)); + succeeded = true; + }, + [&succeeded](MediaResult r) { succeeded = false; }); + EXPECT_TRUE(succeeded); + if (!succeeded) { + return output; + } + } while (pending > 0); + + return output; +} + +TEST_F(MediaDataEncoderTest, H264Encodes) { + RUN_IF_SUPPORTED(VIDEO_MP4, [this]() { + // Encode one frame and output in AnnexB format. + RefPtr<MediaDataEncoder> e = CreateH264Encoder(); + EnsureInit(e); + MediaDataEncoder::EncodedData output = Encode(e, 1UL, mData); + EXPECT_EQ(output.Length(), 1UL); + EXPECT_TRUE(AnnexB::IsAnnexB(output[0])); + WaitForShutdown(e); + + // Encode multiple frames and output in AnnexB format. + e = CreateH264Encoder(); + EnsureInit(e); + output = Encode(e, NUM_FRAMES, mData); + EXPECT_EQ(output.Length(), NUM_FRAMES); + for (auto frame : output) { + EXPECT_TRUE(AnnexB::IsAnnexB(frame)); + } + WaitForShutdown(e); + + // Encode one frame and output in avcC format. + e = CreateH264Encoder(MediaDataEncoder::Usage::Record); + EnsureInit(e); + output = Encode(e, NUM_FRAMES, mData); + EXPECT_EQ(output.Length(), NUM_FRAMES); + AnnexB::IsAVCC(output[0]); // Only 1st frame has extra data. + for (auto frame : output) { + EXPECT_FALSE(AnnexB::IsAnnexB(frame)); + } + WaitForShutdown(e); + }); +} + +#ifndef DEBUG // Zero width or height will assert/crash in debug builds. +TEST_F(MediaDataEncoderTest, InvalidSize) { + RUN_IF_SUPPORTED(VIDEO_MP4, []() { + RefPtr<MediaDataEncoder> e0x0 = + CreateH264Encoder(MediaDataEncoder::Usage::Realtime, + MediaDataEncoder::PixelFormat::YUV420P, 0, 0); + EXPECT_NE(e0x0, nullptr); + EXPECT_FALSE(EnsureInit(e0x0)); + + RefPtr<MediaDataEncoder> e0x1 = + CreateH264Encoder(MediaDataEncoder::Usage::Realtime, + MediaDataEncoder::PixelFormat::YUV420P, 0, 1); + EXPECT_NE(e0x1, nullptr); + EXPECT_FALSE(EnsureInit(e0x1)); + + RefPtr<MediaDataEncoder> e1x0 = + CreateH264Encoder(MediaDataEncoder::Usage::Realtime, + MediaDataEncoder::PixelFormat::YUV420P, 1, 0); + EXPECT_NE(e1x0, nullptr); + EXPECT_FALSE(EnsureInit(e1x0)); + }); +} +#endif + +#ifdef MOZ_WIDGET_ANDROID +TEST_F(MediaDataEncoderTest, AndroidNotSupportedSize) { + RUN_IF_SUPPORTED(VIDEO_MP4, []() { + RefPtr<MediaDataEncoder> e = + CreateH264Encoder(MediaDataEncoder::Usage::Realtime, + MediaDataEncoder::PixelFormat::YUV420P, 1, 1); + EXPECT_NE(e, nullptr); + EXPECT_FALSE(EnsureInit(e)); + }); +} +#endif + +static already_AddRefed<MediaDataEncoder> CreateVP8Encoder( + MediaDataEncoder::Usage aUsage = MediaDataEncoder::Usage::Realtime, + MediaDataEncoder::PixelFormat aPixelFormat = + MediaDataEncoder::PixelFormat::YUV420P, + int32_t aWidth = WIDTH, int32_t aHeight = HEIGHT, + const Maybe<MediaDataEncoder::VPXSpecific::VP8>& aSpecific = + Some(MediaDataEncoder::VPXSpecific::VP8())) { + return CreateVideoEncoder(VIDEO_VP8, aUsage, aPixelFormat, aWidth, aHeight, + aSpecific); +} + +static already_AddRefed<MediaDataEncoder> CreateVP9Encoder( + MediaDataEncoder::Usage aUsage = MediaDataEncoder::Usage::Realtime, + MediaDataEncoder::PixelFormat aPixelFormat = + MediaDataEncoder::PixelFormat::YUV420P, + int32_t aWidth = WIDTH, int32_t aHeight = HEIGHT, + const Maybe<MediaDataEncoder::VPXSpecific::VP9>& aSpecific = + Some(MediaDataEncoder::VPXSpecific::VP9())) { + return CreateVideoEncoder(VIDEO_VP9, aUsage, aPixelFormat, aWidth, aHeight, + aSpecific); +} + +TEST_F(MediaDataEncoderTest, VP8Create) { + RUN_IF_SUPPORTED(VIDEO_VP8, []() { + RefPtr<MediaDataEncoder> e = CreateVP8Encoder(); + EXPECT_TRUE(e); + WaitForShutdown(e); + }); +} + +TEST_F(MediaDataEncoderTest, VP8Inits) { + RUN_IF_SUPPORTED(VIDEO_VP8, []() { + // w/o codec specific. + RefPtr<MediaDataEncoder> e = CreateVP8Encoder( + MediaDataEncoder::Usage::Realtime, + MediaDataEncoder::PixelFormat::YUV420P, WIDTH, HEIGHT, Nothing()); + EXPECT_TRUE(EnsureInit(e)); + WaitForShutdown(e); + + // w/ codec specific + e = CreateVP8Encoder(); + EXPECT_TRUE(EnsureInit(e)); + WaitForShutdown(e); + }); +} + +TEST_F(MediaDataEncoderTest, VP8Encodes) { + RUN_IF_SUPPORTED(VIDEO_VP8, [this]() { + // Encode one VPX frame. + RefPtr<MediaDataEncoder> e = CreateVP8Encoder(); + EnsureInit(e); + MediaDataEncoder::EncodedData output = Encode(e, 1UL, mData); + EXPECT_EQ(output.Length(), 1UL); + VPXDecoder::VPXStreamInfo info; + EXPECT_TRUE( + VPXDecoder::GetStreamInfo(*output[0], info, VPXDecoder::Codec::VP8)); + EXPECT_EQ(info.mKeyFrame, output[0]->mKeyframe); + if (info.mKeyFrame) { + EXPECT_EQ(info.mImage, kImageSize); + } + WaitForShutdown(e); + + // Encode multiple VPX frames. + e = CreateVP8Encoder(); + EnsureInit(e); + output = Encode(e, NUM_FRAMES, mData); + EXPECT_EQ(output.Length(), NUM_FRAMES); + for (auto frame : output) { + VPXDecoder::VPXStreamInfo info; + EXPECT_TRUE( + VPXDecoder::GetStreamInfo(*frame, info, VPXDecoder::Codec::VP8)); + EXPECT_EQ(info.mKeyFrame, frame->mKeyframe); + if (info.mKeyFrame) { + EXPECT_EQ(info.mImage, kImageSize); + } + } + WaitForShutdown(e); + }); +} + +TEST_F(MediaDataEncoderTest, VP9Create) { + RUN_IF_SUPPORTED(VIDEO_VP9, []() { + RefPtr<MediaDataEncoder> e = CreateVP9Encoder(); + EXPECT_TRUE(e); + WaitForShutdown(e); + }); +} + +TEST_F(MediaDataEncoderTest, VP9Inits) { + RUN_IF_SUPPORTED(VIDEO_VP9, []() { + // w/o codec specific. + RefPtr<MediaDataEncoder> e = CreateVP9Encoder( + MediaDataEncoder::Usage::Realtime, + MediaDataEncoder::PixelFormat::YUV420P, WIDTH, HEIGHT, Nothing()); + EXPECT_TRUE(EnsureInit(e)); + WaitForShutdown(e); + + // w/ codec specific + e = CreateVP9Encoder(); + EXPECT_TRUE(EnsureInit(e)); + WaitForShutdown(e); + }); +} + +TEST_F(MediaDataEncoderTest, VP9Encodes) { + RUN_IF_SUPPORTED(VIDEO_VP9, [this]() { + RefPtr<MediaDataEncoder> e = CreateVP9Encoder(); + EnsureInit(e); + MediaDataEncoder::EncodedData output = Encode(e, 1UL, mData); + EXPECT_EQ(output.Length(), 1UL); + VPXDecoder::VPXStreamInfo info; + EXPECT_TRUE( + VPXDecoder::GetStreamInfo(*output[0], info, VPXDecoder::Codec::VP9)); + EXPECT_EQ(info.mKeyFrame, output[0]->mKeyframe); + if (info.mKeyFrame) { + EXPECT_EQ(info.mImage, kImageSize); + } + WaitForShutdown(e); + + e = CreateVP9Encoder(); + EnsureInit(e); + output = Encode(e, NUM_FRAMES, mData); + EXPECT_EQ(output.Length(), NUM_FRAMES); + for (auto frame : output) { + VPXDecoder::VPXStreamInfo info; + EXPECT_TRUE( + VPXDecoder::GetStreamInfo(*frame, info, VPXDecoder::Codec::VP9)); + EXPECT_EQ(info.mKeyFrame, frame->mKeyframe); + if (info.mKeyFrame) { + EXPECT_EQ(info.mImage, kImageSize); + } + } + WaitForShutdown(e); + }); +} |