diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-07 19:33:14 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-07 19:33:14 +0000 |
commit | 36d22d82aa202bb199967e9512281e9a53db42c9 (patch) | |
tree | 105e8c98ddea1c1e4784a60a5a6410fa416be2de /third_party/libwebrtc/video/video_quality_test.cc | |
parent | Initial commit. (diff) | |
download | firefox-esr-36d22d82aa202bb199967e9512281e9a53db42c9.tar.xz firefox-esr-36d22d82aa202bb199967e9512281e9a53db42c9.zip |
Adding upstream version 115.7.0esr.upstream/115.7.0esrupstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'third_party/libwebrtc/video/video_quality_test.cc')
-rw-r--r-- | third_party/libwebrtc/video/video_quality_test.cc | 1577 |
1 files changed, 1577 insertions, 0 deletions
diff --git a/third_party/libwebrtc/video/video_quality_test.cc b/third_party/libwebrtc/video/video_quality_test.cc new file mode 100644 index 0000000000..5d179178f2 --- /dev/null +++ b/third_party/libwebrtc/video/video_quality_test.cc @@ -0,0 +1,1577 @@ +/* + * Copyright (c) 2015 The WebRTC project authors. All Rights Reserved. + * + * Use of this source code is governed by a BSD-style license + * that can be found in the LICENSE file in the root of the source + * tree. An additional intellectual property rights grant can be found + * in the file PATENTS. All contributing project authors may + * be found in the AUTHORS file in the root of the source tree. + */ +#include "video/video_quality_test.h" + +#include <stdio.h> + +#if defined(WEBRTC_WIN) +#include <conio.h> +#endif + +#include <algorithm> +#include <deque> +#include <map> +#include <memory> +#include <string> +#include <vector> + +#include "api/fec_controller_override.h" +#include "api/rtc_event_log_output_file.h" +#include "api/task_queue/default_task_queue_factory.h" +#include "api/task_queue/task_queue_base.h" +#include "api/test/create_frame_generator.h" +#include "api/video/builtin_video_bitrate_allocator_factory.h" +#include "api/video_codecs/video_encoder.h" +#include "call/fake_network_pipe.h" +#include "call/simulated_network.h" +#include "media/base/media_constants.h" +#include "media/engine/adm_helpers.h" +#include "media/engine/encoder_simulcast_proxy.h" +#include "media/engine/fake_video_codec_factory.h" +#include "media/engine/internal_encoder_factory.h" +#include "media/engine/webrtc_video_engine.h" +#include "modules/audio_device/include/audio_device.h" +#include "modules/audio_mixer/audio_mixer_impl.h" +#include "modules/video_coding/codecs/h264/include/h264.h" +#include "modules/video_coding/codecs/multiplex/include/multiplex_decoder_adapter.h" +#include "modules/video_coding/codecs/multiplex/include/multiplex_encoder_adapter.h" +#include "modules/video_coding/codecs/vp8/include/vp8.h" +#include "modules/video_coding/codecs/vp9/include/vp9.h" +#include "modules/video_coding/utility/ivf_file_writer.h" +#include "rtc_base/strings/string_builder.h" +#include "rtc_base/task_queue_for_test.h" +#include "test/platform_video_capturer.h" +#include "test/testsupport/file_utils.h" +#include "test/video_renderer.h" +#include "video/frame_dumping_decoder.h" +#ifdef WEBRTC_WIN +#include "modules/audio_device/include/audio_device_factory.h" +#endif +#include "video/config/encoder_stream_factory.h" + +namespace webrtc { + +namespace { +enum : int { // The first valid value is 1. + kAbsSendTimeExtensionId = 1, + kGenericFrameDescriptorExtensionId00, + kGenericFrameDescriptorExtensionId01, + kTransportSequenceNumberExtensionId, + kVideoContentTypeExtensionId, + kVideoTimingExtensionId, +}; + +constexpr char kSyncGroup[] = "av_sync"; +constexpr int kOpusMinBitrateBps = 6000; +constexpr int kOpusBitrateFbBps = 32000; +constexpr int kFramesSentInQuickTest = 1; +constexpr uint32_t kThumbnailSendSsrcStart = 0xE0000; +constexpr uint32_t kThumbnailRtxSsrcStart = 0xF0000; + +constexpr int kDefaultMaxQp = cricket::WebRtcVideoChannel::kDefaultQpMax; + +const VideoEncoder::Capabilities kCapabilities(false); + +std::pair<uint32_t, uint32_t> GetMinMaxBitratesBps(const VideoCodec& codec, + size_t spatial_idx) { + uint32_t min_bitrate = codec.minBitrate; + uint32_t max_bitrate = codec.maxBitrate; + if (spatial_idx < codec.numberOfSimulcastStreams) { + min_bitrate = + std::max(min_bitrate, codec.simulcastStream[spatial_idx].minBitrate); + max_bitrate = + std::min(max_bitrate, codec.simulcastStream[spatial_idx].maxBitrate); + } + if (codec.codecType == VideoCodecType::kVideoCodecVP9 && + spatial_idx < codec.VP9().numberOfSpatialLayers) { + min_bitrate = + std::max(min_bitrate, codec.spatialLayers[spatial_idx].minBitrate); + max_bitrate = + std::min(max_bitrate, codec.spatialLayers[spatial_idx].maxBitrate); + } + max_bitrate = std::max(max_bitrate, min_bitrate); + return {min_bitrate * 1000, max_bitrate * 1000}; +} + +class VideoStreamFactory + : public VideoEncoderConfig::VideoStreamFactoryInterface { + public: + explicit VideoStreamFactory(const std::vector<VideoStream>& streams) + : streams_(streams) {} + + private: + std::vector<VideoStream> CreateEncoderStreams( + int frame_width, + int frame_height, + const VideoEncoderConfig& encoder_config) override { + // The highest layer must match the incoming resolution. + std::vector<VideoStream> streams = streams_; + streams[streams_.size() - 1].height = frame_height; + streams[streams_.size() - 1].width = frame_width; + + streams[0].bitrate_priority = encoder_config.bitrate_priority; + return streams; + } + + std::vector<VideoStream> streams_; +}; + +// This wrapper provides two features needed by the video quality tests: +// 1. Invoke VideoAnalyzer callbacks before and after encoding each frame. +// 2. Write the encoded frames to file, one file per simulcast layer. +class QualityTestVideoEncoder : public VideoEncoder, + private EncodedImageCallback { + public: + QualityTestVideoEncoder(std::unique_ptr<VideoEncoder> encoder, + VideoAnalyzer* analyzer, + std::vector<FileWrapper> files, + double overshoot_factor) + : encoder_(std::move(encoder)), + overshoot_factor_(overshoot_factor), + analyzer_(analyzer) { + for (FileWrapper& file : files) { + writers_.push_back( + IvfFileWriter::Wrap(std::move(file), /* byte_limit= */ 100000000)); + } + } + + // Implement VideoEncoder + void SetFecControllerOverride( + FecControllerOverride* fec_controller_override) { + // Ignored. + } + + int32_t InitEncode(const VideoCodec* codec_settings, + const Settings& settings) override { + codec_settings_ = *codec_settings; + return encoder_->InitEncode(codec_settings, settings); + } + + int32_t RegisterEncodeCompleteCallback( + EncodedImageCallback* callback) override { + callback_ = callback; + return encoder_->RegisterEncodeCompleteCallback(this); + } + + int32_t Release() override { return encoder_->Release(); } + + int32_t Encode(const VideoFrame& frame, + const std::vector<VideoFrameType>* frame_types) { + if (analyzer_) { + analyzer_->PreEncodeOnFrame(frame); + } + return encoder_->Encode(frame, frame_types); + } + + void SetRates(const RateControlParameters& parameters) override { + RTC_DCHECK_GT(overshoot_factor_, 0.0); + if (overshoot_factor_ == 1.0) { + encoder_->SetRates(parameters); + return; + } + + // Simulating encoder overshooting target bitrate, by configuring actual + // encoder too high. Take care not to adjust past limits of config, + // otherwise encoders may crash on DCHECK. + VideoBitrateAllocation overshot_allocation; + for (size_t si = 0; si < kMaxSpatialLayers; ++si) { + const uint32_t spatial_layer_bitrate_bps = + parameters.bitrate.GetSpatialLayerSum(si); + if (spatial_layer_bitrate_bps == 0) { + continue; + } + + uint32_t min_bitrate_bps; + uint32_t max_bitrate_bps; + std::tie(min_bitrate_bps, max_bitrate_bps) = + GetMinMaxBitratesBps(codec_settings_, si); + double overshoot_factor = overshoot_factor_; + const uint32_t corrected_bitrate = rtc::checked_cast<uint32_t>( + overshoot_factor * spatial_layer_bitrate_bps); + if (corrected_bitrate < min_bitrate_bps) { + overshoot_factor = min_bitrate_bps / spatial_layer_bitrate_bps; + } else if (corrected_bitrate > max_bitrate_bps) { + overshoot_factor = max_bitrate_bps / spatial_layer_bitrate_bps; + } + + for (size_t ti = 0; ti < kMaxTemporalStreams; ++ti) { + if (parameters.bitrate.HasBitrate(si, ti)) { + overshot_allocation.SetBitrate( + si, ti, + rtc::checked_cast<uint32_t>( + overshoot_factor * parameters.bitrate.GetBitrate(si, ti))); + } + } + } + + return encoder_->SetRates( + RateControlParameters(overshot_allocation, parameters.framerate_fps, + parameters.bandwidth_allocation)); + } + + void OnPacketLossRateUpdate(float packet_loss_rate) override { + encoder_->OnPacketLossRateUpdate(packet_loss_rate); + } + + void OnRttUpdate(int64_t rtt_ms) override { encoder_->OnRttUpdate(rtt_ms); } + + void OnLossNotification(const LossNotification& loss_notification) override { + encoder_->OnLossNotification(loss_notification); + } + + EncoderInfo GetEncoderInfo() const override { + EncoderInfo info = encoder_->GetEncoderInfo(); + if (overshoot_factor_ != 1.0) { + // We're simulating bad encoder, don't forward trusted setting + // from eg libvpx. + info.has_trusted_rate_controller = false; + } + return info; + } + + private: + // Implement EncodedImageCallback + Result OnEncodedImage(const EncodedImage& encoded_image, + const CodecSpecificInfo* codec_specific_info) override { + if (codec_specific_info) { + int simulcast_index; + if (codec_specific_info->codecType == kVideoCodecVP9) { + simulcast_index = 0; + } else { + simulcast_index = encoded_image.SpatialIndex().value_or(0); + } + RTC_DCHECK_GE(simulcast_index, 0); + if (analyzer_) { + analyzer_->PostEncodeOnFrame(simulcast_index, + encoded_image.Timestamp()); + } + if (static_cast<size_t>(simulcast_index) < writers_.size()) { + writers_[simulcast_index]->WriteFrame(encoded_image, + codec_specific_info->codecType); + } + } + + return callback_->OnEncodedImage(encoded_image, codec_specific_info); + } + + void OnDroppedFrame(DropReason reason) override { + callback_->OnDroppedFrame(reason); + } + + const std::unique_ptr<VideoEncoder> encoder_; + const double overshoot_factor_; + VideoAnalyzer* const analyzer_; + std::vector<std::unique_ptr<IvfFileWriter>> writers_; + EncodedImageCallback* callback_ = nullptr; + VideoCodec codec_settings_; +}; + +#if defined(WEBRTC_WIN) && !defined(WINUWP) +void PressEnterToContinue(TaskQueueBase* task_queue) { + puts(">> Press ENTER to continue..."); + + while (!_kbhit() || _getch() != '\r') { + // Drive the message loop for the thread running the task_queue + SendTask(task_queue, [&]() { + MSG msg; + if (PeekMessage(&msg, NULL, 0, 0, PM_REMOVE)) { + TranslateMessage(&msg); + DispatchMessage(&msg); + } + }); + } +} +#else +void PressEnterToContinue(TaskQueueBase* /*task_queue*/) { + puts(">> Press ENTER to continue..."); + while (getc(stdin) != '\n' && !feof(stdin)) + ; // NOLINT +} +#endif + +} // namespace + +std::unique_ptr<VideoDecoder> VideoQualityTest::CreateVideoDecoder( + const SdpVideoFormat& format) { + std::unique_ptr<VideoDecoder> decoder; + if (format.name == "multiplex") { + decoder = std::make_unique<MultiplexDecoderAdapter>( + decoder_factory_.get(), SdpVideoFormat(cricket::kVp9CodecName)); + } else if (format.name == "FakeCodec") { + decoder = webrtc::FakeVideoDecoderFactory::CreateVideoDecoder(); + } else { + decoder = decoder_factory_->CreateVideoDecoder(format); + } + if (!params_.logging.encoded_frame_base_path.empty()) { + rtc::StringBuilder str; + str << receive_logs_++; + std::string path = + params_.logging.encoded_frame_base_path + "." + str.str() + ".recv.ivf"; + decoder = CreateFrameDumpingDecoderWrapper( + std::move(decoder), FileWrapper::OpenWriteOnly(path)); + } + return decoder; +} + +std::unique_ptr<VideoEncoder> VideoQualityTest::CreateVideoEncoder( + const SdpVideoFormat& format, + VideoAnalyzer* analyzer) { + std::unique_ptr<VideoEncoder> encoder; + if (format.name == "VP8") { + encoder = + std::make_unique<EncoderSimulcastProxy>(encoder_factory_.get(), format); + } else if (format.name == "multiplex") { + encoder = std::make_unique<MultiplexEncoderAdapter>( + encoder_factory_.get(), SdpVideoFormat(cricket::kVp9CodecName)); + } else if (format.name == "FakeCodec") { + encoder = webrtc::FakeVideoEncoderFactory::CreateVideoEncoder(); + } else { + encoder = encoder_factory_->CreateVideoEncoder(format); + } + + std::vector<FileWrapper> encoded_frame_dump_files; + if (!params_.logging.encoded_frame_base_path.empty()) { + char ss_buf[100]; + rtc::SimpleStringBuilder sb(ss_buf); + sb << send_logs_++; + std::string prefix = + params_.logging.encoded_frame_base_path + "." + sb.str() + ".send."; + encoded_frame_dump_files.push_back( + FileWrapper::OpenWriteOnly(prefix + "1.ivf")); + encoded_frame_dump_files.push_back( + FileWrapper::OpenWriteOnly(prefix + "2.ivf")); + encoded_frame_dump_files.push_back( + FileWrapper::OpenWriteOnly(prefix + "3.ivf")); + } + + double overshoot_factor = 1.0; + // Match format to either of the streams in dual-stream mode in order to get + // the overshoot factor. This is not very robust but we can't know for sure + // which stream this encoder is meant for, from within the factory. + if (format == + SdpVideoFormat(params_.video[0].codec, params_.video[0].sdp_params)) { + overshoot_factor = params_.video[0].encoder_overshoot_factor; + } else if (format == SdpVideoFormat(params_.video[1].codec, + params_.video[1].sdp_params)) { + overshoot_factor = params_.video[1].encoder_overshoot_factor; + } + if (overshoot_factor == 0.0) { + // If params were zero-initialized, set to 1.0 instead. + overshoot_factor = 1.0; + } + + if (analyzer || !encoded_frame_dump_files.empty() || overshoot_factor > 1.0) { + encoder = std::make_unique<QualityTestVideoEncoder>( + std::move(encoder), analyzer, std::move(encoded_frame_dump_files), + overshoot_factor); + } + + return encoder; +} + +VideoQualityTest::VideoQualityTest( + std::unique_ptr<InjectionComponents> injection_components) + : clock_(Clock::GetRealTimeClock()), + task_queue_factory_(CreateDefaultTaskQueueFactory()), + rtc_event_log_factory_(task_queue_factory_.get()), + video_decoder_factory_([this](const SdpVideoFormat& format) { + return this->CreateVideoDecoder(format); + }), + video_encoder_factory_([this](const SdpVideoFormat& format) { + return this->CreateVideoEncoder(format, nullptr); + }), + video_encoder_factory_with_analyzer_( + [this](const SdpVideoFormat& format) { + return this->CreateVideoEncoder(format, analyzer_.get()); + }), + video_bitrate_allocator_factory_( + CreateBuiltinVideoBitrateAllocatorFactory()), + receive_logs_(0), + send_logs_(0), + injection_components_(std::move(injection_components)), + num_video_streams_(0) { + if (injection_components_ == nullptr) { + injection_components_ = std::make_unique<InjectionComponents>(); + } + if (injection_components_->video_decoder_factory != nullptr) { + decoder_factory_ = std::move(injection_components_->video_decoder_factory); + } else { + decoder_factory_ = std::make_unique<InternalDecoderFactory>(); + } + if (injection_components_->video_encoder_factory != nullptr) { + encoder_factory_ = std::move(injection_components_->video_encoder_factory); + } else { + encoder_factory_ = std::make_unique<InternalEncoderFactory>(); + } + + fec_controller_factory_ = + std::move(injection_components_->fec_controller_factory); + network_state_predictor_factory_ = + std::move(injection_components_->network_state_predictor_factory); + network_controller_factory_ = + std::move(injection_components_->network_controller_factory); + + // Register header extensions that are used by transport to identify + // extensions when parsing incomig packets. + RegisterRtpExtension(RtpExtension(RtpExtension::kTransportSequenceNumberUri, + kTransportSequenceNumberExtensionId)); + RegisterRtpExtension( + RtpExtension(RtpExtension::kAbsSendTimeUri, kAbsSendTimeExtensionId)); + RegisterRtpExtension(RtpExtension(RtpExtension::kGenericFrameDescriptorUri00, + kGenericFrameDescriptorExtensionId00)); + RegisterRtpExtension(RtpExtension(RtpExtension::kDependencyDescriptorUri, + kRtpExtensionDependencyDescriptor)); + RegisterRtpExtension(RtpExtension(RtpExtension::kVideoContentTypeUri, + kVideoContentTypeExtensionId)); + RegisterRtpExtension( + RtpExtension(RtpExtension::kVideoTimingUri, kVideoTimingExtensionId)); +} + +VideoQualityTest::InjectionComponents::InjectionComponents() = default; + +VideoQualityTest::InjectionComponents::~InjectionComponents() = default; + +void VideoQualityTest::TestBody() {} + +std::string VideoQualityTest::GenerateGraphTitle() const { + rtc::StringBuilder ss; + ss << params_.video[0].codec; + ss << " (" << params_.video[0].target_bitrate_bps / 1000 << "kbps"; + ss << ", " << params_.video[0].fps << " FPS"; + if (params_.screenshare[0].scroll_duration) + ss << ", " << params_.screenshare[0].scroll_duration << "s scroll"; + if (params_.ss[0].streams.size() > 1) + ss << ", Stream #" << params_.ss[0].selected_stream; + if (params_.ss[0].num_spatial_layers > 1) + ss << ", Layer #" << params_.ss[0].selected_sl; + ss << ")"; + return ss.Release(); +} + +void VideoQualityTest::CheckParamsAndInjectionComponents() { + if (injection_components_ == nullptr) { + injection_components_ = std::make_unique<InjectionComponents>(); + } + if (!params_.config && injection_components_->sender_network == nullptr && + injection_components_->receiver_network == nullptr) { + params_.config = BuiltInNetworkBehaviorConfig(); + } + RTC_CHECK( + (params_.config && injection_components_->sender_network == nullptr && + injection_components_->receiver_network == nullptr) || + (!params_.config && injection_components_->sender_network != nullptr && + injection_components_->receiver_network != nullptr)); + for (size_t video_idx = 0; video_idx < num_video_streams_; ++video_idx) { + // Iterate over primary and secondary video streams. + if (!params_.video[video_idx].enabled) + return; + // Add a default stream in none specified. + if (params_.ss[video_idx].streams.empty()) + params_.ss[video_idx].streams.push_back( + VideoQualityTest::DefaultVideoStream(params_, video_idx)); + if (params_.ss[video_idx].num_spatial_layers == 0) + params_.ss[video_idx].num_spatial_layers = 1; + + if (params_.config) { + if (params_.config->loss_percent != 0 || + params_.config->queue_length_packets != 0) { + // Since LayerFilteringTransport changes the sequence numbers, we can't + // use that feature with pack loss, since the NACK request would end up + // retransmitting the wrong packets. + RTC_CHECK(params_.ss[video_idx].selected_sl == -1 || + params_.ss[video_idx].selected_sl == + params_.ss[video_idx].num_spatial_layers - 1); + RTC_CHECK(params_.video[video_idx].selected_tl == -1 || + params_.video[video_idx].selected_tl == + params_.video[video_idx].num_temporal_layers - 1); + } + } + + // TODO(ivica): Should max_bitrate_bps == -1 represent inf max bitrate, as + // it does in some parts of the code? + RTC_CHECK_GE(params_.video[video_idx].max_bitrate_bps, + params_.video[video_idx].target_bitrate_bps); + RTC_CHECK_GE(params_.video[video_idx].target_bitrate_bps, + params_.video[video_idx].min_bitrate_bps); + int selected_stream = params_.ss[video_idx].selected_stream; + if (params_.video[video_idx].selected_tl > -1) { + RTC_CHECK_LT(selected_stream, params_.ss[video_idx].streams.size()) + << "Can not use --selected_tl when --selected_stream is all streams"; + int stream_tl = params_.ss[video_idx] + .streams[selected_stream] + .num_temporal_layers.value_or(1); + RTC_CHECK_LT(params_.video[video_idx].selected_tl, stream_tl); + } + RTC_CHECK_LE(params_.ss[video_idx].selected_stream, + params_.ss[video_idx].streams.size()); + for (const VideoStream& stream : params_.ss[video_idx].streams) { + RTC_CHECK_GE(stream.min_bitrate_bps, 0); + RTC_CHECK_GE(stream.target_bitrate_bps, stream.min_bitrate_bps); + RTC_CHECK_GE(stream.max_bitrate_bps, stream.target_bitrate_bps); + } + // TODO(ivica): Should we check if the sum of all streams/layers is equal to + // the total bitrate? We anyway have to update them in the case bitrate + // estimator changes the total bitrates. + RTC_CHECK_GE(params_.ss[video_idx].num_spatial_layers, 1); + RTC_CHECK_LE(params_.ss[video_idx].selected_sl, + params_.ss[video_idx].num_spatial_layers); + RTC_CHECK( + params_.ss[video_idx].spatial_layers.empty() || + params_.ss[video_idx].spatial_layers.size() == + static_cast<size_t>(params_.ss[video_idx].num_spatial_layers)); + if (params_.video[video_idx].codec == "VP8") { + RTC_CHECK_EQ(params_.ss[video_idx].num_spatial_layers, 1); + } else if (params_.video[video_idx].codec == "VP9") { + RTC_CHECK_EQ(params_.ss[video_idx].streams.size(), 1); + } + RTC_CHECK_GE(params_.call.num_thumbnails, 0); + if (params_.call.num_thumbnails > 0) { + RTC_CHECK_EQ(params_.ss[video_idx].num_spatial_layers, 1); + RTC_CHECK_EQ(params_.ss[video_idx].streams.size(), 3); + RTC_CHECK_EQ(params_.video[video_idx].num_temporal_layers, 3); + RTC_CHECK_EQ(params_.video[video_idx].codec, "VP8"); + } + // Dual streams with FEC not supported in tests yet. + RTC_CHECK(!params_.video[video_idx].flexfec || num_video_streams_ == 1); + RTC_CHECK(!params_.video[video_idx].ulpfec || num_video_streams_ == 1); + } +} + +// Static. +std::vector<int> VideoQualityTest::ParseCSV(const std::string& str) { + // Parse comma separated nonnegative integers, where some elements may be + // empty. The empty values are replaced with -1. + // E.g. "10,-20,,30,40" --> {10, 20, -1, 30,40} + // E.g. ",,10,,20," --> {-1, -1, 10, -1, 20, -1} + std::vector<int> result; + if (str.empty()) + return result; + + const char* p = str.c_str(); + int value = -1; + int pos; + while (*p) { + if (*p == ',') { + result.push_back(value); + value = -1; + ++p; + continue; + } + RTC_CHECK_EQ(sscanf(p, "%d%n", &value, &pos), 1) + << "Unexpected non-number value."; + p += pos; + } + result.push_back(value); + return result; +} + +// Static. +VideoStream VideoQualityTest::DefaultVideoStream(const Params& params, + size_t video_idx) { + VideoStream stream; + stream.width = params.video[video_idx].width; + stream.height = params.video[video_idx].height; + stream.max_framerate = params.video[video_idx].fps; + stream.min_bitrate_bps = params.video[video_idx].min_bitrate_bps; + stream.target_bitrate_bps = params.video[video_idx].target_bitrate_bps; + stream.max_bitrate_bps = params.video[video_idx].max_bitrate_bps; + stream.max_qp = kDefaultMaxQp; + stream.num_temporal_layers = params.video[video_idx].num_temporal_layers; + stream.active = true; + return stream; +} + +// Static. +VideoStream VideoQualityTest::DefaultThumbnailStream() { + VideoStream stream; + stream.width = 320; + stream.height = 180; + stream.max_framerate = 7; + stream.min_bitrate_bps = 7500; + stream.target_bitrate_bps = 37500; + stream.max_bitrate_bps = 50000; + stream.max_qp = kDefaultMaxQp; + return stream; +} + +// Static. +void VideoQualityTest::FillScalabilitySettings( + Params* params, + size_t video_idx, + const std::vector<std::string>& stream_descriptors, + int num_streams, + size_t selected_stream, + int num_spatial_layers, + int selected_sl, + InterLayerPredMode inter_layer_pred, + const std::vector<std::string>& sl_descriptors) { + if (params->ss[video_idx].streams.empty() && + params->ss[video_idx].infer_streams) { + webrtc::VideoEncoder::EncoderInfo encoder_info; + webrtc::VideoEncoderConfig encoder_config; + encoder_config.codec_type = + PayloadStringToCodecType(params->video[video_idx].codec); + encoder_config.content_type = + params->screenshare[video_idx].enabled + ? webrtc::VideoEncoderConfig::ContentType::kScreen + : webrtc::VideoEncoderConfig::ContentType::kRealtimeVideo; + encoder_config.max_bitrate_bps = params->video[video_idx].max_bitrate_bps; + encoder_config.min_transmit_bitrate_bps = + params->video[video_idx].min_transmit_bps; + encoder_config.number_of_streams = num_streams; + encoder_config.spatial_layers = params->ss[video_idx].spatial_layers; + encoder_config.simulcast_layers = std::vector<VideoStream>(num_streams); + encoder_config.video_stream_factory = + rtc::make_ref_counted<cricket::EncoderStreamFactory>( + params->video[video_idx].codec, kDefaultMaxQp, + params->screenshare[video_idx].enabled, true, encoder_info); + params->ss[video_idx].streams = + encoder_config.video_stream_factory->CreateEncoderStreams( + params->video[video_idx].width, params->video[video_idx].height, + encoder_config); + } else { + // Read VideoStream and SpatialLayer elements from a list of comma separated + // lists. To use a default value for an element, use -1 or leave empty. + // Validity checks performed in CheckParamsAndInjectionComponents. + RTC_CHECK(params->ss[video_idx].streams.empty()); + for (const auto& descriptor : stream_descriptors) { + if (descriptor.empty()) + continue; + VideoStream stream = + VideoQualityTest::DefaultVideoStream(*params, video_idx); + std::vector<int> v = VideoQualityTest::ParseCSV(descriptor); + if (v[0] != -1) + stream.width = static_cast<size_t>(v[0]); + if (v[1] != -1) + stream.height = static_cast<size_t>(v[1]); + if (v[2] != -1) + stream.max_framerate = v[2]; + if (v[3] != -1) + stream.min_bitrate_bps = v[3]; + if (v[4] != -1) + stream.target_bitrate_bps = v[4]; + if (v[5] != -1) + stream.max_bitrate_bps = v[5]; + if (v.size() > 6 && v[6] != -1) + stream.max_qp = v[6]; + if (v.size() > 7 && v[7] != -1) { + stream.num_temporal_layers = v[7]; + } else { + // Automatic TL thresholds for more than two layers not supported. + RTC_CHECK_LE(params->video[video_idx].num_temporal_layers, 2); + } + params->ss[video_idx].streams.push_back(stream); + } + } + + params->ss[video_idx].num_spatial_layers = std::max(1, num_spatial_layers); + params->ss[video_idx].selected_stream = selected_stream; + + params->ss[video_idx].selected_sl = selected_sl; + params->ss[video_idx].inter_layer_pred = inter_layer_pred; + RTC_CHECK(params->ss[video_idx].spatial_layers.empty()); + for (const auto& descriptor : sl_descriptors) { + if (descriptor.empty()) + continue; + std::vector<int> v = VideoQualityTest::ParseCSV(descriptor); + RTC_CHECK_EQ(v.size(), 8); + + SpatialLayer layer = {0}; + layer.width = v[0]; + layer.height = v[1]; + layer.maxFramerate = v[2]; + layer.numberOfTemporalLayers = v[3]; + layer.maxBitrate = v[4]; + layer.minBitrate = v[5]; + layer.targetBitrate = v[6]; + layer.qpMax = v[7]; + layer.active = true; + + params->ss[video_idx].spatial_layers.push_back(layer); + } +} + +void VideoQualityTest::SetupVideo(Transport* send_transport, + Transport* recv_transport) { + size_t total_streams_used = 0; + video_receive_configs_.clear(); + video_send_configs_.clear(); + video_encoder_configs_.clear(); + bool decode_all_receive_streams = true; + size_t num_video_substreams = params_.ss[0].streams.size(); + RTC_CHECK(num_video_streams_ > 0); + video_encoder_configs_.resize(num_video_streams_); + std::string generic_codec_name; + webrtc::VideoEncoder::EncoderInfo encoder_info; + for (size_t video_idx = 0; video_idx < num_video_streams_; ++video_idx) { + VideoSendStream::Config config(send_transport); + config.rtp.extmap_allow_mixed = true; + video_send_configs_.push_back(std::move(config)); + video_encoder_configs_.push_back(VideoEncoderConfig()); + num_video_substreams = params_.ss[video_idx].streams.size(); + RTC_CHECK_GT(num_video_substreams, 0); + for (size_t i = 0; i < num_video_substreams; ++i) + video_send_configs_[video_idx].rtp.ssrcs.push_back( + kVideoSendSsrcs[total_streams_used + i]); + + int payload_type; + if (params_.video[video_idx].codec == "H264") { + payload_type = kPayloadTypeH264; + } else if (params_.video[video_idx].codec == "VP8") { + payload_type = kPayloadTypeVP8; + } else if (params_.video[video_idx].codec == "VP9") { + payload_type = kPayloadTypeVP9; + } else if (params_.video[video_idx].codec == "multiplex") { + payload_type = kPayloadTypeVP9; + } else if (params_.video[video_idx].codec == "FakeCodec") { + payload_type = kFakeVideoSendPayloadType; + } else { + RTC_CHECK(generic_codec_name.empty() || + generic_codec_name == params_.video[video_idx].codec) + << "Supplying multiple generic codecs is unsupported."; + RTC_LOG(LS_INFO) << "Treating codec " << params_.video[video_idx].codec + << " as generic."; + payload_type = kPayloadTypeGeneric; + generic_codec_name = params_.video[video_idx].codec; + } + video_send_configs_[video_idx].encoder_settings.encoder_factory = + (video_idx == 0) ? &video_encoder_factory_with_analyzer_ + : &video_encoder_factory_; + video_send_configs_[video_idx].encoder_settings.bitrate_allocator_factory = + video_bitrate_allocator_factory_.get(); + + video_send_configs_[video_idx].rtp.payload_name = + params_.video[video_idx].codec; + video_send_configs_[video_idx].rtp.payload_type = payload_type; + video_send_configs_[video_idx].rtp.nack.rtp_history_ms = kNackRtpHistoryMs; + video_send_configs_[video_idx].rtp.rtx.payload_type = kSendRtxPayloadType; + for (size_t i = 0; i < num_video_substreams; ++i) { + video_send_configs_[video_idx].rtp.rtx.ssrcs.push_back( + kSendRtxSsrcs[i + total_streams_used]); + } + video_send_configs_[video_idx].rtp.extensions.clear(); + if (params_.call.send_side_bwe) { + video_send_configs_[video_idx].rtp.extensions.emplace_back( + RtpExtension::kTransportSequenceNumberUri, + kTransportSequenceNumberExtensionId); + } else { + video_send_configs_[video_idx].rtp.extensions.emplace_back( + RtpExtension::kAbsSendTimeUri, kAbsSendTimeExtensionId); + } + + if (params_.call.generic_descriptor) { + video_send_configs_[video_idx].rtp.extensions.emplace_back( + RtpExtension::kGenericFrameDescriptorUri00, + kGenericFrameDescriptorExtensionId00); + } + + if (params_.call.dependency_descriptor) { + video_send_configs_[video_idx].rtp.extensions.emplace_back( + RtpExtension::kDependencyDescriptorUri, + kRtpExtensionDependencyDescriptor); + } + + video_send_configs_[video_idx].rtp.extensions.emplace_back( + RtpExtension::kVideoContentTypeUri, kVideoContentTypeExtensionId); + video_send_configs_[video_idx].rtp.extensions.emplace_back( + RtpExtension::kVideoTimingUri, kVideoTimingExtensionId); + + video_encoder_configs_[video_idx].video_format.name = + params_.video[video_idx].codec; + + video_encoder_configs_[video_idx].video_format.parameters = + params_.video[video_idx].sdp_params; + + video_encoder_configs_[video_idx].codec_type = + PayloadStringToCodecType(params_.video[video_idx].codec); + + video_encoder_configs_[video_idx].min_transmit_bitrate_bps = + params_.video[video_idx].min_transmit_bps; + + video_send_configs_[video_idx].suspend_below_min_bitrate = + params_.video[video_idx].suspend_below_min_bitrate; + + video_encoder_configs_[video_idx].number_of_streams = + params_.ss[video_idx].streams.size(); + video_encoder_configs_[video_idx].max_bitrate_bps = 0; + for (size_t i = 0; i < params_.ss[video_idx].streams.size(); ++i) { + video_encoder_configs_[video_idx].max_bitrate_bps += + params_.ss[video_idx].streams[i].max_bitrate_bps; + } + video_encoder_configs_[video_idx].simulcast_layers = + std::vector<VideoStream>(params_.ss[video_idx].streams.size()); + if (!params_.ss[video_idx].infer_streams) { + video_encoder_configs_[video_idx].simulcast_layers = + params_.ss[video_idx].streams; + } + video_encoder_configs_[video_idx].video_stream_factory = + rtc::make_ref_counted<cricket::EncoderStreamFactory>( + params_.video[video_idx].codec, + params_.ss[video_idx].streams[0].max_qp, + params_.screenshare[video_idx].enabled, true, encoder_info); + + video_encoder_configs_[video_idx].spatial_layers = + params_.ss[video_idx].spatial_layers; + + video_encoder_configs_[video_idx].frame_drop_enabled = true; + + decode_all_receive_streams = params_.ss[video_idx].selected_stream == + params_.ss[video_idx].streams.size(); + absl::optional<int> decode_sub_stream; + if (!decode_all_receive_streams) + decode_sub_stream = params_.ss[video_idx].selected_stream; + CreateMatchingVideoReceiveConfigs( + video_send_configs_[video_idx], recv_transport, &video_decoder_factory_, + decode_sub_stream, true, kNackRtpHistoryMs); + + if (params_.screenshare[video_idx].enabled) { + // Fill out codec settings. + video_encoder_configs_[video_idx].content_type = + VideoEncoderConfig::ContentType::kScreen; + degradation_preference_ = DegradationPreference::MAINTAIN_RESOLUTION; + if (params_.video[video_idx].codec == "VP8") { + VideoCodecVP8 vp8_settings = VideoEncoder::GetDefaultVp8Settings(); + vp8_settings.denoisingOn = false; + vp8_settings.numberOfTemporalLayers = static_cast<unsigned char>( + params_.video[video_idx].num_temporal_layers); + video_encoder_configs_[video_idx].encoder_specific_settings = + rtc::make_ref_counted< + VideoEncoderConfig::Vp8EncoderSpecificSettings>(vp8_settings); + } else if (params_.video[video_idx].codec == "VP9") { + VideoCodecVP9 vp9_settings = VideoEncoder::GetDefaultVp9Settings(); + vp9_settings.denoisingOn = false; + vp9_settings.automaticResizeOn = false; + vp9_settings.numberOfTemporalLayers = static_cast<unsigned char>( + params_.video[video_idx].num_temporal_layers); + vp9_settings.numberOfSpatialLayers = static_cast<unsigned char>( + params_.ss[video_idx].num_spatial_layers); + vp9_settings.interLayerPred = params_.ss[video_idx].inter_layer_pred; + // High FPS vp9 screenshare requires flexible mode. + if (params_.ss[video_idx].num_spatial_layers > 1) { + vp9_settings.flexibleMode = true; + } + video_encoder_configs_[video_idx].encoder_specific_settings = + rtc::make_ref_counted< + VideoEncoderConfig::Vp9EncoderSpecificSettings>(vp9_settings); + } + } else if (params_.ss[video_idx].num_spatial_layers > 1) { + // If SVC mode without screenshare, still need to set codec specifics. + RTC_CHECK(params_.video[video_idx].codec == "VP9"); + VideoCodecVP9 vp9_settings = VideoEncoder::GetDefaultVp9Settings(); + vp9_settings.numberOfTemporalLayers = static_cast<unsigned char>( + params_.video[video_idx].num_temporal_layers); + vp9_settings.numberOfSpatialLayers = + static_cast<unsigned char>(params_.ss[video_idx].num_spatial_layers); + vp9_settings.interLayerPred = params_.ss[video_idx].inter_layer_pred; + vp9_settings.automaticResizeOn = false; + video_encoder_configs_[video_idx].encoder_specific_settings = + rtc::make_ref_counted<VideoEncoderConfig::Vp9EncoderSpecificSettings>( + vp9_settings); + RTC_DCHECK_EQ(video_encoder_configs_[video_idx].simulcast_layers.size(), + 1); + // Min bitrate will be enforced by spatial layer config instead. + video_encoder_configs_[video_idx].simulcast_layers[0].min_bitrate_bps = 0; + } else if (params_.video[video_idx].automatic_scaling) { + if (params_.video[video_idx].codec == "VP8") { + VideoCodecVP8 vp8_settings = VideoEncoder::GetDefaultVp8Settings(); + vp8_settings.automaticResizeOn = true; + video_encoder_configs_[video_idx].encoder_specific_settings = + rtc::make_ref_counted< + VideoEncoderConfig::Vp8EncoderSpecificSettings>(vp8_settings); + } else if (params_.video[video_idx].codec == "VP9") { + VideoCodecVP9 vp9_settings = VideoEncoder::GetDefaultVp9Settings(); + // Only enable quality scaler for single spatial layer. + vp9_settings.automaticResizeOn = + params_.ss[video_idx].num_spatial_layers == 1; + video_encoder_configs_[video_idx].encoder_specific_settings = + rtc::make_ref_counted< + VideoEncoderConfig::Vp9EncoderSpecificSettings>(vp9_settings); + } else if (params_.video[video_idx].codec == "H264") { + // Quality scaling is always on for H.264. + } else if (params_.video[video_idx].codec == cricket::kAv1CodecName) { + // TODO(bugs.webrtc.org/11404): Propagate the flag to + // aom_codec_enc_cfg_t::rc_resize_mode in Av1 encoder wrapper. + // Until then do nothing, specially do not crash. + } else { + RTC_DCHECK_NOTREACHED() + << "Automatic scaling not supported for codec " + << params_.video[video_idx].codec << ", stream " << video_idx; + } + } else { + // Default mode. Single SL, no automatic_scaling, + if (params_.video[video_idx].codec == "VP8") { + VideoCodecVP8 vp8_settings = VideoEncoder::GetDefaultVp8Settings(); + vp8_settings.automaticResizeOn = false; + video_encoder_configs_[video_idx].encoder_specific_settings = + rtc::make_ref_counted< + VideoEncoderConfig::Vp8EncoderSpecificSettings>(vp8_settings); + } else if (params_.video[video_idx].codec == "VP9") { + VideoCodecVP9 vp9_settings = VideoEncoder::GetDefaultVp9Settings(); + vp9_settings.automaticResizeOn = false; + video_encoder_configs_[video_idx].encoder_specific_settings = + rtc::make_ref_counted< + VideoEncoderConfig::Vp9EncoderSpecificSettings>(vp9_settings); + } else if (params_.video[video_idx].codec == "H264") { + video_encoder_configs_[video_idx].encoder_specific_settings = nullptr; + } + } + total_streams_used += num_video_substreams; + } + + // FEC supported only for single video stream mode yet. + if (params_.video[0].flexfec) { + if (decode_all_receive_streams) { + SetSendFecConfig(GetVideoSendConfig()->rtp.ssrcs); + } else { + SetSendFecConfig({kVideoSendSsrcs[params_.ss[0].selected_stream]}); + } + + CreateMatchingFecConfig(recv_transport, *GetVideoSendConfig()); + if (params_.call.send_side_bwe) { + GetFlexFecConfig()->rtp.extensions.push_back( + RtpExtension(RtpExtension::kTransportSequenceNumberUri, + kTransportSequenceNumberExtensionId)); + } else { + GetFlexFecConfig()->rtp.extensions.push_back( + RtpExtension(RtpExtension::kAbsSendTimeUri, kAbsSendTimeExtensionId)); + } + } + + if (params_.video[0].ulpfec) { + SetSendUlpFecConfig(GetVideoSendConfig()); + if (decode_all_receive_streams) { + for (auto& receive_config : video_receive_configs_) { + SetReceiveUlpFecConfig(&receive_config); + } + } else { + SetReceiveUlpFecConfig( + &video_receive_configs_[params_.ss[0].selected_stream]); + } + } +} + +void VideoQualityTest::SetupThumbnails(Transport* send_transport, + Transport* recv_transport) { + for (int i = 0; i < params_.call.num_thumbnails; ++i) { + // Thumbnails will be send in the other way: from receiver_call to + // sender_call. + VideoSendStream::Config thumbnail_send_config(recv_transport); + thumbnail_send_config.rtp.ssrcs.push_back(kThumbnailSendSsrcStart + i); + thumbnail_send_config.encoder_settings.encoder_factory = + &video_encoder_factory_; + thumbnail_send_config.encoder_settings.bitrate_allocator_factory = + video_bitrate_allocator_factory_.get(); + thumbnail_send_config.rtp.payload_name = params_.video[0].codec; + thumbnail_send_config.rtp.payload_type = kPayloadTypeVP8; + thumbnail_send_config.rtp.nack.rtp_history_ms = kNackRtpHistoryMs; + thumbnail_send_config.rtp.rtx.payload_type = kSendRtxPayloadType; + thumbnail_send_config.rtp.rtx.ssrcs.push_back(kThumbnailRtxSsrcStart + i); + thumbnail_send_config.rtp.extensions.clear(); + if (params_.call.send_side_bwe) { + thumbnail_send_config.rtp.extensions.push_back( + RtpExtension(RtpExtension::kTransportSequenceNumberUri, + kTransportSequenceNumberExtensionId)); + } else { + thumbnail_send_config.rtp.extensions.push_back( + RtpExtension(RtpExtension::kAbsSendTimeUri, kAbsSendTimeExtensionId)); + } + + VideoEncoderConfig thumbnail_encoder_config; + thumbnail_encoder_config.codec_type = kVideoCodecVP8; + thumbnail_encoder_config.video_format.name = "VP8"; + thumbnail_encoder_config.min_transmit_bitrate_bps = 7500; + thumbnail_send_config.suspend_below_min_bitrate = + params_.video[0].suspend_below_min_bitrate; + thumbnail_encoder_config.number_of_streams = 1; + thumbnail_encoder_config.max_bitrate_bps = 50000; + std::vector<VideoStream> streams{params_.ss[0].streams[0]}; + thumbnail_encoder_config.video_stream_factory = + rtc::make_ref_counted<VideoStreamFactory>(streams); + thumbnail_encoder_config.spatial_layers = params_.ss[0].spatial_layers; + + thumbnail_encoder_configs_.push_back(thumbnail_encoder_config.Copy()); + thumbnail_send_configs_.push_back(thumbnail_send_config.Copy()); + + AddMatchingVideoReceiveConfigs( + &thumbnail_receive_configs_, thumbnail_send_config, send_transport, + &video_decoder_factory_, absl::nullopt, false, kNackRtpHistoryMs); + } + for (size_t i = 0; i < thumbnail_send_configs_.size(); ++i) { + thumbnail_send_streams_.push_back(receiver_call_->CreateVideoSendStream( + thumbnail_send_configs_[i].Copy(), + thumbnail_encoder_configs_[i].Copy())); + } + for (size_t i = 0; i < thumbnail_receive_configs_.size(); ++i) { + thumbnail_receive_streams_.push_back(sender_call_->CreateVideoReceiveStream( + thumbnail_receive_configs_[i].Copy())); + } +} + +void VideoQualityTest::DestroyThumbnailStreams() { + for (VideoSendStream* thumbnail_send_stream : thumbnail_send_streams_) { + receiver_call_->DestroyVideoSendStream(thumbnail_send_stream); + } + thumbnail_send_streams_.clear(); + for (VideoReceiveStreamInterface* thumbnail_receive_stream : + thumbnail_receive_streams_) { + sender_call_->DestroyVideoReceiveStream(thumbnail_receive_stream); + } + thumbnail_send_streams_.clear(); + thumbnail_receive_streams_.clear(); + for (std::unique_ptr<rtc::VideoSourceInterface<VideoFrame>>& video_capturer : + thumbnail_capturers_) { + video_capturer.reset(); + } +} + +void VideoQualityTest::SetupThumbnailCapturers(size_t num_thumbnail_streams) { + VideoStream thumbnail = DefaultThumbnailStream(); + for (size_t i = 0; i < num_thumbnail_streams; ++i) { + auto frame_generator_capturer = + std::make_unique<test::FrameGeneratorCapturer>( + clock_, + test::CreateSquareFrameGenerator(static_cast<int>(thumbnail.width), + static_cast<int>(thumbnail.height), + absl::nullopt, absl::nullopt), + thumbnail.max_framerate, *task_queue_factory_); + EXPECT_TRUE(frame_generator_capturer->Init()); + thumbnail_capturers_.push_back(std::move(frame_generator_capturer)); + } +} + +std::unique_ptr<test::FrameGeneratorInterface> +VideoQualityTest::CreateFrameGenerator(size_t video_idx) { + // Setup frame generator. + const size_t kWidth = 1850; + const size_t kHeight = 1110; + std::unique_ptr<test::FrameGeneratorInterface> frame_generator; + if (params_.screenshare[video_idx].generate_slides) { + frame_generator = test::CreateSlideFrameGenerator( + kWidth, kHeight, + params_.screenshare[video_idx].slide_change_interval * + params_.video[video_idx].fps); + } else { + std::vector<std::string> slides = params_.screenshare[video_idx].slides; + if (slides.empty()) { + slides.push_back(test::ResourcePath("web_screenshot_1850_1110", "yuv")); + slides.push_back(test::ResourcePath("presentation_1850_1110", "yuv")); + slides.push_back(test::ResourcePath("photo_1850_1110", "yuv")); + slides.push_back(test::ResourcePath("difficult_photo_1850_1110", "yuv")); + } + if (params_.screenshare[video_idx].scroll_duration == 0) { + // Cycle image every slide_change_interval seconds. + frame_generator = test::CreateFromYuvFileFrameGenerator( + slides, kWidth, kHeight, + params_.screenshare[video_idx].slide_change_interval * + params_.video[video_idx].fps); + } else { + RTC_CHECK_LE(params_.video[video_idx].width, kWidth); + RTC_CHECK_LE(params_.video[video_idx].height, kHeight); + RTC_CHECK_GT(params_.screenshare[video_idx].slide_change_interval, 0); + const int kPauseDurationMs = + (params_.screenshare[video_idx].slide_change_interval - + params_.screenshare[video_idx].scroll_duration) * + 1000; + RTC_CHECK_LE(params_.screenshare[video_idx].scroll_duration, + params_.screenshare[video_idx].slide_change_interval); + + frame_generator = test::CreateScrollingInputFromYuvFilesFrameGenerator( + clock_, slides, kWidth, kHeight, params_.video[video_idx].width, + params_.video[video_idx].height, + params_.screenshare[video_idx].scroll_duration * 1000, + kPauseDurationMs); + } + } + return frame_generator; +} + +void VideoQualityTest::CreateCapturers() { + RTC_DCHECK(video_sources_.empty()); + video_sources_.resize(num_video_streams_); + for (size_t video_idx = 0; video_idx < num_video_streams_; ++video_idx) { + std::unique_ptr<test::FrameGeneratorInterface> frame_generator; + if (params_.screenshare[video_idx].enabled) { + frame_generator = CreateFrameGenerator(video_idx); + } else if (params_.video[video_idx].clip_path == "Generator") { + frame_generator = test::CreateSquareFrameGenerator( + static_cast<int>(params_.video[video_idx].width), + static_cast<int>(params_.video[video_idx].height), absl::nullopt, + absl::nullopt); + } else if (params_.video[video_idx].clip_path == "GeneratorI420A") { + frame_generator = test::CreateSquareFrameGenerator( + static_cast<int>(params_.video[video_idx].width), + static_cast<int>(params_.video[video_idx].height), + test::FrameGeneratorInterface::OutputType::kI420A, absl::nullopt); + } else if (params_.video[video_idx].clip_path == "GeneratorI010") { + frame_generator = test::CreateSquareFrameGenerator( + static_cast<int>(params_.video[video_idx].width), + static_cast<int>(params_.video[video_idx].height), + test::FrameGeneratorInterface::OutputType::kI010, absl::nullopt); + } else if (params_.video[video_idx].clip_path == "GeneratorNV12") { + frame_generator = test::CreateSquareFrameGenerator( + static_cast<int>(params_.video[video_idx].width), + static_cast<int>(params_.video[video_idx].height), + test::FrameGeneratorInterface::OutputType::kNV12, absl::nullopt); + } else if (params_.video[video_idx].clip_path.empty()) { + video_sources_[video_idx] = test::CreateVideoCapturer( + params_.video[video_idx].width, params_.video[video_idx].height, + params_.video[video_idx].fps, + params_.video[video_idx].capture_device_index); + if (video_sources_[video_idx]) { + continue; + } else { + // Failed to get actual camera, use chroma generator as backup. + frame_generator = test::CreateSquareFrameGenerator( + static_cast<int>(params_.video[video_idx].width), + static_cast<int>(params_.video[video_idx].height), absl::nullopt, + absl::nullopt); + } + } else { + frame_generator = test::CreateFromYuvFileFrameGenerator( + {params_.video[video_idx].clip_path}, params_.video[video_idx].width, + params_.video[video_idx].height, 1); + ASSERT_TRUE(frame_generator) << "Could not create capturer for " + << params_.video[video_idx].clip_path + << ".yuv. Is this file present?"; + } + ASSERT_TRUE(frame_generator); + auto frame_generator_capturer = + std::make_unique<test::FrameGeneratorCapturer>( + clock_, std::move(frame_generator), params_.video[video_idx].fps, + *task_queue_factory_); + EXPECT_TRUE(frame_generator_capturer->Init()); + video_sources_[video_idx] = std::move(frame_generator_capturer); + } +} + +void VideoQualityTest::StartAudioStreams() { + audio_send_stream_->Start(); + for (AudioReceiveStreamInterface* audio_recv_stream : audio_receive_streams_) + audio_recv_stream->Start(); +} + +void VideoQualityTest::StartThumbnails() { + for (VideoSendStream* send_stream : thumbnail_send_streams_) + send_stream->Start(); + for (VideoReceiveStreamInterface* receive_stream : thumbnail_receive_streams_) + receive_stream->Start(); +} + +void VideoQualityTest::StopThumbnails() { + for (VideoReceiveStreamInterface* receive_stream : thumbnail_receive_streams_) + receive_stream->Stop(); + for (VideoSendStream* send_stream : thumbnail_send_streams_) + send_stream->Stop(); +} + +std::unique_ptr<test::LayerFilteringTransport> +VideoQualityTest::CreateSendTransport() { + std::unique_ptr<NetworkBehaviorInterface> network_behavior = nullptr; + if (injection_components_->sender_network == nullptr) { + network_behavior = std::make_unique<SimulatedNetwork>(*params_.config); + } else { + network_behavior = std::move(injection_components_->sender_network); + } + return std::make_unique<test::LayerFilteringTransport>( + task_queue(), + std::make_unique<FakeNetworkPipe>(clock_, std::move(network_behavior)), + sender_call_.get(), kPayloadTypeVP8, kPayloadTypeVP9, + params_.video[0].selected_tl, params_.ss[0].selected_sl, + payload_type_map_, kVideoSendSsrcs[0], + static_cast<uint32_t>(kVideoSendSsrcs[0] + params_.ss[0].streams.size() - + 1), + GetRegisteredExtensions(), GetRegisteredExtensions()); +} + +std::unique_ptr<test::DirectTransport> +VideoQualityTest::CreateReceiveTransport() { + std::unique_ptr<NetworkBehaviorInterface> network_behavior = nullptr; + if (injection_components_->receiver_network == nullptr) { + network_behavior = std::make_unique<SimulatedNetwork>(*params_.config); + } else { + network_behavior = std::move(injection_components_->receiver_network); + } + return std::make_unique<test::DirectTransport>( + task_queue(), + std::make_unique<FakeNetworkPipe>(clock_, std::move(network_behavior)), + receiver_call_.get(), payload_type_map_, GetRegisteredExtensions(), + GetRegisteredExtensions()); +} + +void VideoQualityTest::RunWithAnalyzer(const Params& params) { + num_video_streams_ = params.call.dual_video ? 2 : 1; + std::unique_ptr<test::LayerFilteringTransport> send_transport; + std::unique_ptr<test::DirectTransport> recv_transport; + FILE* graph_data_output_file = nullptr; + + params_ = params; + // TODO(ivica): Merge with RunWithRenderer and use a flag / argument to + // differentiate between the analyzer and the renderer case. + CheckParamsAndInjectionComponents(); + + if (!params_.analyzer.graph_data_output_filename.empty()) { + graph_data_output_file = + fopen(params_.analyzer.graph_data_output_filename.c_str(), "w"); + RTC_CHECK(graph_data_output_file) + << "Can't open the file " << params_.analyzer.graph_data_output_filename + << "!"; + } + + if (!params.logging.rtc_event_log_name.empty()) { + send_event_log_ = rtc_event_log_factory_.CreateRtcEventLog( + RtcEventLog::EncodingType::NewFormat); + recv_event_log_ = rtc_event_log_factory_.CreateRtcEventLog( + RtcEventLog::EncodingType::NewFormat); + std::unique_ptr<RtcEventLogOutputFile> send_output( + std::make_unique<RtcEventLogOutputFile>( + params.logging.rtc_event_log_name + "_send", + RtcEventLog::kUnlimitedOutput)); + std::unique_ptr<RtcEventLogOutputFile> recv_output( + std::make_unique<RtcEventLogOutputFile>( + params.logging.rtc_event_log_name + "_recv", + RtcEventLog::kUnlimitedOutput)); + bool event_log_started = + send_event_log_->StartLogging(std::move(send_output), + RtcEventLog::kImmediateOutput) && + recv_event_log_->StartLogging(std::move(recv_output), + RtcEventLog::kImmediateOutput); + RTC_DCHECK(event_log_started); + } else { + send_event_log_ = std::make_unique<RtcEventLogNull>(); + recv_event_log_ = std::make_unique<RtcEventLogNull>(); + } + + SendTask(task_queue(), [this, ¶ms, &send_transport, &recv_transport]() { + Call::Config send_call_config(send_event_log_.get()); + Call::Config recv_call_config(recv_event_log_.get()); + send_call_config.bitrate_config = params.call.call_bitrate_config; + recv_call_config.bitrate_config = params.call.call_bitrate_config; + if (params_.audio.enabled) + InitializeAudioDevice(&send_call_config, &recv_call_config, + params_.audio.use_real_adm); + + CreateCalls(send_call_config, recv_call_config); + send_transport = CreateSendTransport(); + recv_transport = CreateReceiveTransport(); + }); + + std::string graph_title = params_.analyzer.graph_title; + if (graph_title.empty()) + graph_title = VideoQualityTest::GenerateGraphTitle(); + bool is_quick_test_enabled = field_trial::IsEnabled("WebRTC-QuickPerfTest"); + analyzer_ = std::make_unique<VideoAnalyzer>( + send_transport.get(), params_.analyzer.test_label, + params_.analyzer.avg_psnr_threshold, params_.analyzer.avg_ssim_threshold, + is_quick_test_enabled + ? kFramesSentInQuickTest + : params_.analyzer.test_durations_secs * params_.video[0].fps, + is_quick_test_enabled + ? TimeDelta::Millis(1) + : TimeDelta::Seconds(params_.analyzer.test_durations_secs), + graph_data_output_file, graph_title, + kVideoSendSsrcs[params_.ss[0].selected_stream], + kSendRtxSsrcs[params_.ss[0].selected_stream], + static_cast<size_t>(params_.ss[0].selected_stream), + params.ss[0].selected_sl, params_.video[0].selected_tl, + is_quick_test_enabled, clock_, params_.logging.rtp_dump_name, + task_queue()); + + SendTask(task_queue(), [&]() { + analyzer_->SetCall(sender_call_.get()); + analyzer_->SetReceiver(receiver_call_->Receiver()); + send_transport->SetReceiver(analyzer_.get()); + recv_transport->SetReceiver(sender_call_->Receiver()); + + SetupVideo(analyzer_.get(), recv_transport.get()); + SetupThumbnails(analyzer_.get(), recv_transport.get()); + video_receive_configs_[params_.ss[0].selected_stream].renderer = + analyzer_.get(); + + CreateFlexfecStreams(); + CreateVideoStreams(); + analyzer_->SetSendStream(video_send_streams_[0]); + analyzer_->SetReceiveStream( + video_receive_streams_[params_.ss[0].selected_stream]); + + GetVideoSendStream()->SetSource(analyzer_->OutputInterface(), + degradation_preference_); + SetupThumbnailCapturers(params_.call.num_thumbnails); + for (size_t i = 0; i < thumbnail_send_streams_.size(); ++i) { + thumbnail_send_streams_[i]->SetSource(thumbnail_capturers_[i].get(), + degradation_preference_); + } + + CreateCapturers(); + + analyzer_->SetSource(video_sources_[0].get(), true); + + for (size_t video_idx = 1; video_idx < num_video_streams_; ++video_idx) { + video_send_streams_[video_idx]->SetSource(video_sources_[video_idx].get(), + degradation_preference_); + } + + if (params_.audio.enabled) { + SetupAudio(send_transport.get()); + StartAudioStreams(); + analyzer_->SetAudioReceiveStream(audio_receive_streams_[0]); + } + StartVideoStreams(); + StartThumbnails(); + analyzer_->StartMeasuringCpuProcessTime(); + }); + + analyzer_->Wait(); + + SendTask(task_queue(), [&]() { + StopThumbnails(); + Stop(); + + DestroyStreams(); + DestroyThumbnailStreams(); + + if (graph_data_output_file) + fclose(graph_data_output_file); + + send_transport.reset(); + recv_transport.reset(); + + DestroyCalls(); + }); + analyzer_ = nullptr; +} + +rtc::scoped_refptr<AudioDeviceModule> VideoQualityTest::CreateAudioDevice() { +#ifdef WEBRTC_WIN + RTC_LOG(LS_INFO) << "Using latest version of ADM on Windows"; + // We must initialize the COM library on a thread before we calling any of + // the library functions. All COM functions in the ADM will return + // CO_E_NOTINITIALIZED otherwise. The legacy ADM for Windows used internal + // COM initialization but the new ADM requires COM to be initialized + // externally. + com_initializer_ = + std::make_unique<ScopedCOMInitializer>(ScopedCOMInitializer::kMTA); + RTC_CHECK(com_initializer_->Succeeded()); + RTC_CHECK(webrtc_win::core_audio_utility::IsSupported()); + RTC_CHECK(webrtc_win::core_audio_utility::IsMMCSSSupported()); + return CreateWindowsCoreAudioAudioDeviceModule(task_queue_factory_.get()); +#else + // Use legacy factory method on all platforms except Windows. + return AudioDeviceModule::Create(AudioDeviceModule::kPlatformDefaultAudio, + task_queue_factory_.get()); +#endif +} + +void VideoQualityTest::InitializeAudioDevice(Call::Config* send_call_config, + Call::Config* recv_call_config, + bool use_real_adm) { + rtc::scoped_refptr<AudioDeviceModule> audio_device; + if (use_real_adm) { + // Run test with real ADM (using default audio devices) if user has + // explicitly set the --audio and --use_real_adm command-line flags. + audio_device = CreateAudioDevice(); + } else { + // By default, create a test ADM which fakes audio. + audio_device = TestAudioDeviceModule::Create( + task_queue_factory_.get(), + TestAudioDeviceModule::CreatePulsedNoiseCapturer(32000, 48000), + TestAudioDeviceModule::CreateDiscardRenderer(48000), 1.f); + } + RTC_CHECK(audio_device); + + AudioState::Config audio_state_config; + audio_state_config.audio_mixer = AudioMixerImpl::Create(); + audio_state_config.audio_processing = AudioProcessingBuilder().Create(); + audio_state_config.audio_device_module = audio_device; + send_call_config->audio_state = AudioState::Create(audio_state_config); + recv_call_config->audio_state = AudioState::Create(audio_state_config); + if (use_real_adm) { + // The real ADM requires extra initialization: setting default devices, + // setting up number of channels etc. Helper class also calls + // AudioDeviceModule::Init(). + webrtc::adm_helpers::Init(audio_device.get()); + } else { + audio_device->Init(); + } + // Always initialize the ADM before injecting a valid audio transport. + RTC_CHECK(audio_device->RegisterAudioCallback( + send_call_config->audio_state->audio_transport()) == 0); +} + +void VideoQualityTest::SetupAudio(Transport* transport) { + AudioSendStream::Config audio_send_config(transport); + audio_send_config.rtp.ssrc = kAudioSendSsrc; + + // Add extension to enable audio send side BWE, and allow audio bit rate + // adaptation. + audio_send_config.rtp.extensions.clear(); + audio_send_config.send_codec_spec = AudioSendStream::Config::SendCodecSpec( + kAudioSendPayloadType, + {"OPUS", + 48000, + 2, + {{"usedtx", (params_.audio.dtx ? "1" : "0")}, {"stereo", "1"}}}); + + if (params_.call.send_side_bwe) { + audio_send_config.rtp.extensions.push_back( + webrtc::RtpExtension(webrtc::RtpExtension::kTransportSequenceNumberUri, + kTransportSequenceNumberExtensionId)); + audio_send_config.min_bitrate_bps = kOpusMinBitrateBps; + audio_send_config.max_bitrate_bps = kOpusBitrateFbBps; + audio_send_config.send_codec_spec->transport_cc_enabled = true; + // Only allow ANA when send-side BWE is enabled. + audio_send_config.audio_network_adaptor_config = params_.audio.ana_config; + } + audio_send_config.encoder_factory = audio_encoder_factory_; + SetAudioConfig(audio_send_config); + + std::string sync_group; + if (params_.video[0].enabled && params_.audio.sync_video) + sync_group = kSyncGroup; + + CreateMatchingAudioConfigs(transport, sync_group); + CreateAudioStreams(); +} + +void VideoQualityTest::RunWithRenderers(const Params& params) { + RTC_LOG(LS_INFO) << __FUNCTION__; + num_video_streams_ = params.call.dual_video ? 2 : 1; + std::unique_ptr<test::LayerFilteringTransport> send_transport; + std::unique_ptr<test::DirectTransport> recv_transport; + std::unique_ptr<test::VideoRenderer> local_preview; + std::vector<std::unique_ptr<test::VideoRenderer>> loopback_renderers; + + if (!params.logging.rtc_event_log_name.empty()) { + send_event_log_ = rtc_event_log_factory_.CreateRtcEventLog( + RtcEventLog::EncodingType::NewFormat); + recv_event_log_ = rtc_event_log_factory_.CreateRtcEventLog( + RtcEventLog::EncodingType::NewFormat); + std::unique_ptr<RtcEventLogOutputFile> send_output( + std::make_unique<RtcEventLogOutputFile>( + params.logging.rtc_event_log_name + "_send", + RtcEventLog::kUnlimitedOutput)); + std::unique_ptr<RtcEventLogOutputFile> recv_output( + std::make_unique<RtcEventLogOutputFile>( + params.logging.rtc_event_log_name + "_recv", + RtcEventLog::kUnlimitedOutput)); + bool event_log_started = + send_event_log_->StartLogging(std::move(send_output), + /*output_period_ms=*/5000) && + recv_event_log_->StartLogging(std::move(recv_output), + /*output_period_ms=*/5000); + RTC_DCHECK(event_log_started); + } else { + send_event_log_ = std::make_unique<RtcEventLogNull>(); + recv_event_log_ = std::make_unique<RtcEventLogNull>(); + } + + SendTask(task_queue(), [&]() { + params_ = params; + CheckParamsAndInjectionComponents(); + + // TODO(ivica): Remove bitrate_config and use the default Call::Config(), to + // match the full stack tests. + Call::Config send_call_config(send_event_log_.get()); + send_call_config.bitrate_config = params_.call.call_bitrate_config; + Call::Config recv_call_config(recv_event_log_.get()); + + if (params_.audio.enabled) + InitializeAudioDevice(&send_call_config, &recv_call_config, + params_.audio.use_real_adm); + + CreateCalls(send_call_config, recv_call_config); + + // TODO(minyue): consider if this is a good transport even for audio only + // calls. + send_transport = CreateSendTransport(); + + recv_transport = CreateReceiveTransport(); + + // TODO(ivica): Use two calls to be able to merge with RunWithAnalyzer or at + // least share as much code as possible. That way this test would also match + // the full stack tests better. + send_transport->SetReceiver(receiver_call_->Receiver()); + recv_transport->SetReceiver(sender_call_->Receiver()); + + if (params_.video[0].enabled) { + // Create video renderers. + SetupVideo(send_transport.get(), recv_transport.get()); + size_t num_streams_processed = 0; + for (size_t video_idx = 0; video_idx < num_video_streams_; ++video_idx) { + const size_t selected_stream_id = params_.ss[video_idx].selected_stream; + const size_t num_streams = params_.ss[video_idx].streams.size(); + if (selected_stream_id == num_streams) { + for (size_t stream_id = 0; stream_id < num_streams; ++stream_id) { + rtc::StringBuilder oss; + oss << "Loopback Video #" << video_idx << " - Stream #" + << static_cast<int>(stream_id); + loopback_renderers.emplace_back(test::VideoRenderer::Create( + oss.str().c_str(), + params_.ss[video_idx].streams[stream_id].width, + params_.ss[video_idx].streams[stream_id].height)); + video_receive_configs_[stream_id + num_streams_processed].renderer = + loopback_renderers.back().get(); + if (params_.audio.enabled && params_.audio.sync_video) + video_receive_configs_[stream_id + num_streams_processed] + .sync_group = kSyncGroup; + } + } else { + rtc::StringBuilder oss; + oss << "Loopback Video #" << video_idx; + loopback_renderers.emplace_back(test::VideoRenderer::Create( + oss.str().c_str(), + params_.ss[video_idx].streams[selected_stream_id].width, + params_.ss[video_idx].streams[selected_stream_id].height)); + video_receive_configs_[selected_stream_id + num_streams_processed] + .renderer = loopback_renderers.back().get(); + if (params_.audio.enabled && params_.audio.sync_video) + video_receive_configs_[num_streams_processed + selected_stream_id] + .sync_group = kSyncGroup; + } + num_streams_processed += num_streams; + } + CreateFlexfecStreams(); + CreateVideoStreams(); + + CreateCapturers(); + if (params_.video[0].enabled) { + // Create local preview + local_preview.reset(test::VideoRenderer::Create( + "Local Preview", params_.video[0].width, params_.video[0].height)); + + video_sources_[0]->AddOrUpdateSink(local_preview.get(), + rtc::VideoSinkWants()); + } + ConnectVideoSourcesToStreams(); + } + + if (params_.audio.enabled) { + SetupAudio(send_transport.get()); + } + + Start(); + }); + + PressEnterToContinue(task_queue()); + + SendTask(task_queue(), [&]() { + Stop(); + DestroyStreams(); + + send_transport.reset(); + recv_transport.reset(); + + local_preview.reset(); + loopback_renderers.clear(); + + DestroyCalls(); + }); +} + +} // namespace webrtc |