summaryrefslogtreecommitdiffstats
path: root/third_party/libwebrtc/webrtc/video/video_stream_encoder.cc
diff options
context:
space:
mode:
Diffstat (limited to 'third_party/libwebrtc/webrtc/video/video_stream_encoder.cc')
-rw-r--r--third_party/libwebrtc/webrtc/video/video_stream_encoder.cc1273
1 files changed, 1273 insertions, 0 deletions
diff --git a/third_party/libwebrtc/webrtc/video/video_stream_encoder.cc b/third_party/libwebrtc/webrtc/video/video_stream_encoder.cc
new file mode 100644
index 0000000000..61e7f00ab0
--- /dev/null
+++ b/third_party/libwebrtc/webrtc/video/video_stream_encoder.cc
@@ -0,0 +1,1273 @@
+/*
+ * Copyright (c) 2012 The WebRTC project authors. All Rights Reserved.
+ *
+ * Use of this source code is governed by a BSD-style license
+ * that can be found in the LICENSE file in the root of the source
+ * tree. An additional intellectual property rights grant can be found
+ * in the file PATENTS. All contributing project authors may
+ * be found in the AUTHORS file in the root of the source tree.
+ */
+
+#include "video/video_stream_encoder.h"
+
+#include <algorithm>
+#include <limits>
+#include <numeric>
+#include <utility>
+
+#include "api/video/i420_buffer.h"
+#include "common_video/include/video_bitrate_allocator.h"
+#include "common_video/include/video_frame.h"
+#include "modules/pacing/paced_sender.h"
+#include "modules/video_coding/codecs/vp8/temporal_layers.h"
+#include "modules/video_coding/include/video_codec_initializer.h"
+#include "modules/video_coding/include/video_coding.h"
+#include "modules/video_coding/include/video_coding_defines.h"
+#include "rtc_base/arraysize.h"
+#include "rtc_base/checks.h"
+#include "rtc_base/location.h"
+#include "rtc_base/logging.h"
+#include "rtc_base/timeutils.h"
+#include "rtc_base/trace_event.h"
+#include "video/overuse_frame_detector.h"
+#include "video/send_statistics_proxy.h"
+
+namespace webrtc {
+
+namespace {
+
+// Time interval for logging frame counts.
+const int64_t kFrameLogIntervalMs = 60000;
+const int kMinFramerateFps = 2;
+const int kMaxFramerateFps = 120;
+
+// The maximum number of frames to drop at beginning of stream
+// to try and achieve desired bitrate.
+const int kMaxInitialFramedrop = 4;
+
+uint32_t MaximumFrameSizeForBitrate(uint32_t kbps) {
+ if (kbps > 0) {
+ if (kbps < 300 /* qvga */) {
+ return 320 * 240;
+ } else if (kbps < 500 /* vga */) {
+ return 640 * 480;
+ }
+ }
+ return std::numeric_limits<uint32_t>::max();
+}
+
+// Initial limits for kBalanced degradation preference.
+int MinFps(int pixels) {
+ if (pixels <= 320 * 240) {
+ return 7;
+ } else if (pixels <= 480 * 270) {
+ return 10;
+ } else if (pixels <= 640 * 480) {
+ return 15;
+ } else {
+ return std::numeric_limits<int>::max();
+ }
+}
+
+int MaxFps(int pixels) {
+ if (pixels <= 320 * 240) {
+ return 10;
+ } else if (pixels <= 480 * 270) {
+ return 15;
+ } else {
+ return std::numeric_limits<int>::max();
+ }
+}
+
+bool IsResolutionScalingEnabled(
+ VideoSendStream::DegradationPreference degradation_preference) {
+ return degradation_preference ==
+ VideoSendStream::DegradationPreference::kMaintainFramerate ||
+ degradation_preference ==
+ VideoSendStream::DegradationPreference::kBalanced;
+}
+
+bool IsFramerateScalingEnabled(
+ VideoSendStream::DegradationPreference degradation_preference) {
+ return degradation_preference ==
+ VideoSendStream::DegradationPreference::kMaintainResolution ||
+ degradation_preference ==
+ VideoSendStream::DegradationPreference::kBalanced;
+}
+
+} // namespace
+
+class VideoStreamEncoder::ConfigureEncoderTask : public rtc::QueuedTask {
+ public:
+ ConfigureEncoderTask(VideoStreamEncoder* video_stream_encoder,
+ VideoEncoderConfig config,
+ size_t max_data_payload_length,
+ bool nack_enabled)
+ : video_stream_encoder_(video_stream_encoder),
+ config_(std::move(config)),
+ max_data_payload_length_(max_data_payload_length),
+ nack_enabled_(nack_enabled) {}
+
+ private:
+ bool Run() override {
+ video_stream_encoder_->ConfigureEncoderOnTaskQueue(
+ std::move(config_), max_data_payload_length_, nack_enabled_);
+ return true;
+ }
+
+ VideoStreamEncoder* const video_stream_encoder_;
+ VideoEncoderConfig config_;
+ size_t max_data_payload_length_;
+ bool nack_enabled_;
+};
+
+class VideoStreamEncoder::EncodeTask : public rtc::QueuedTask {
+ public:
+ EncodeTask(const VideoFrame& frame,
+ VideoStreamEncoder* video_stream_encoder,
+ int64_t time_when_posted_us,
+ bool log_stats)
+ : frame_(frame),
+ video_stream_encoder_(video_stream_encoder),
+ time_when_posted_us_(time_when_posted_us),
+ log_stats_(log_stats) {
+ ++video_stream_encoder_->posted_frames_waiting_for_encode_;
+ }
+
+ private:
+ bool Run() override {
+ RTC_DCHECK_RUN_ON(&video_stream_encoder_->encoder_queue_);
+ video_stream_encoder_->stats_proxy_->OnIncomingFrame(frame_.width(),
+ frame_.height());
+ ++video_stream_encoder_->captured_frame_count_;
+ const int posted_frames_waiting_for_encode =
+ video_stream_encoder_->posted_frames_waiting_for_encode_.fetch_sub(1);
+ RTC_DCHECK_GT(posted_frames_waiting_for_encode, 0);
+ if (posted_frames_waiting_for_encode == 1) {
+ video_stream_encoder_->EncodeVideoFrame(frame_, time_when_posted_us_);
+ } else {
+ // There is a newer frame in flight. Do not encode this frame.
+ RTC_LOG(LS_VERBOSE)
+ << "Incoming frame dropped due to that the encoder is blocked.";
+ ++video_stream_encoder_->dropped_frame_count_;
+ video_stream_encoder_->stats_proxy_->OnFrameDroppedInEncoderQueue();
+ }
+ if (log_stats_) {
+ RTC_LOG(LS_INFO) << "Number of frames: captured "
+ << video_stream_encoder_->captured_frame_count_
+ << ", dropped (due to encoder blocked) "
+ << video_stream_encoder_->dropped_frame_count_
+ << ", interval_ms " << kFrameLogIntervalMs;
+ video_stream_encoder_->captured_frame_count_ = 0;
+ video_stream_encoder_->dropped_frame_count_ = 0;
+ }
+ return true;
+ }
+ VideoFrame frame_;
+ VideoStreamEncoder* const video_stream_encoder_;
+ const int64_t time_when_posted_us_;
+ const bool log_stats_;
+};
+
+// VideoSourceProxy is responsible ensuring thread safety between calls to
+// VideoStreamEncoder::SetSource that will happen on libjingle's worker thread
+// when a video capturer is connected to the encoder and the encoder task queue
+// (encoder_queue_) where the encoder reports its VideoSinkWants.
+class VideoStreamEncoder::VideoSourceProxy {
+ public:
+ explicit VideoSourceProxy(VideoStreamEncoder* video_stream_encoder)
+ : video_stream_encoder_(video_stream_encoder),
+ degradation_preference_(
+ VideoSendStream::DegradationPreference::kDegradationDisabled),
+ source_(nullptr) {}
+
+ void SetSource(
+ rtc::VideoSourceInterface<VideoFrame>* source,
+ const VideoSendStream::DegradationPreference& degradation_preference) {
+ // Called on libjingle's worker thread.
+ RTC_DCHECK_CALLED_SEQUENTIALLY(&main_checker_);
+ rtc::VideoSourceInterface<VideoFrame>* old_source = nullptr;
+ rtc::VideoSinkWants wants;
+ {
+ rtc::CritScope lock(&crit_);
+ degradation_preference_ = degradation_preference;
+ old_source = source_;
+ source_ = source;
+ wants = GetActiveSinkWantsInternal();
+ }
+
+ if (old_source != source && old_source != nullptr) {
+ old_source->RemoveSink(video_stream_encoder_);
+ }
+
+ if (!source) {
+ return;
+ }
+
+ source->AddOrUpdateSink(video_stream_encoder_, wants);
+ }
+
+ void SetWantsRotationApplied(bool rotation_applied) {
+ rtc::CritScope lock(&crit_);
+ sink_wants_.rotation_applied = rotation_applied;
+ if (source_)
+ source_->AddOrUpdateSink(video_stream_encoder_, sink_wants_);
+ }
+
+ rtc::VideoSinkWants GetActiveSinkWants() {
+ rtc::CritScope lock(&crit_);
+ return GetActiveSinkWantsInternal();
+ }
+
+ void ResetPixelFpsCount() {
+ rtc::CritScope lock(&crit_);
+ sink_wants_.max_pixel_count = std::numeric_limits<int>::max();
+ sink_wants_.target_pixel_count.reset();
+ sink_wants_.max_framerate_fps = std::numeric_limits<int>::max();
+ if (source_)
+ source_->AddOrUpdateSink(video_stream_encoder_, sink_wants_);
+ }
+
+ bool RequestResolutionLowerThan(int pixel_count,
+ int min_pixels_per_frame,
+ bool* min_pixels_reached) {
+ // Called on the encoder task queue.
+ rtc::CritScope lock(&crit_);
+ if (!source_ || !IsResolutionScalingEnabled(degradation_preference_)) {
+ // This can happen since |degradation_preference_| is set on libjingle's
+ // worker thread but the adaptation is done on the encoder task queue.
+ return false;
+ }
+ // The input video frame size will have a resolution less than or equal to
+ // |max_pixel_count| depending on how the source can scale the frame size.
+ const int pixels_wanted = (pixel_count * 3) / 5;
+ if (pixels_wanted >= sink_wants_.max_pixel_count) {
+ return false;
+ }
+ if (pixels_wanted < min_pixels_per_frame) {
+ *min_pixels_reached = true;
+ return false;
+ }
+ RTC_LOG(LS_INFO) << "Scaling down resolution, max pixels: "
+ << pixels_wanted;
+ sink_wants_.max_pixel_count = pixels_wanted;
+ sink_wants_.target_pixel_count = rtc::Optional<int>();
+ source_->AddOrUpdateSink(video_stream_encoder_,
+ GetActiveSinkWantsInternal());
+ return true;
+ }
+
+ int RequestFramerateLowerThan(int fps) {
+ // Called on the encoder task queue.
+ // The input video frame rate will be scaled down to 2/3, rounding down.
+ int framerate_wanted = (fps * 2) / 3;
+ return RestrictFramerate(framerate_wanted) ? framerate_wanted : -1;
+ }
+
+ bool RequestHigherResolutionThan(int pixel_count) {
+ // Called on the encoder task queue.
+ rtc::CritScope lock(&crit_);
+ if (!source_ || !IsResolutionScalingEnabled(degradation_preference_)) {
+ // This can happen since |degradation_preference_| is set on libjingle's
+ // worker thread but the adaptation is done on the encoder task queue.
+ return false;
+ }
+ int max_pixels_wanted = pixel_count;
+ if (max_pixels_wanted != std::numeric_limits<int>::max())
+ max_pixels_wanted = pixel_count * 4;
+
+ if (max_pixels_wanted <= sink_wants_.max_pixel_count)
+ return false;
+
+ sink_wants_.max_pixel_count = max_pixels_wanted;
+ if (max_pixels_wanted == std::numeric_limits<int>::max()) {
+ // Remove any constraints.
+ sink_wants_.target_pixel_count.reset();
+ } else {
+ // On step down we request at most 3/5 the pixel count of the previous
+ // resolution, so in order to take "one step up" we request a resolution
+ // as close as possible to 5/3 of the current resolution. The actual pixel
+ // count selected depends on the capabilities of the source. In order to
+ // not take a too large step up, we cap the requested pixel count to be at
+ // most four time the current number of pixels.
+ sink_wants_.target_pixel_count =
+ rtc::Optional<int>((pixel_count * 5) / 3);
+ }
+ RTC_LOG(LS_INFO) << "Scaling up resolution, max pixels: "
+ << max_pixels_wanted;
+ source_->AddOrUpdateSink(video_stream_encoder_,
+ GetActiveSinkWantsInternal());
+ return true;
+ }
+
+ // Request upgrade in framerate. Returns the new requested frame, or -1 if
+ // no change requested. Note that maxint may be returned if limits due to
+ // adaptation requests are removed completely. In that case, consider
+ // |max_framerate_| to be the current limit (assuming the capturer complies).
+ int RequestHigherFramerateThan(int fps) {
+ // Called on the encoder task queue.
+ // The input frame rate will be scaled up to the last step, with rounding.
+ int framerate_wanted = fps;
+ if (fps != std::numeric_limits<int>::max())
+ framerate_wanted = (fps * 3) / 2;
+
+ return IncreaseFramerate(framerate_wanted) ? framerate_wanted : -1;
+ }
+
+ bool RestrictFramerate(int fps) {
+ // Called on the encoder task queue.
+ rtc::CritScope lock(&crit_);
+ if (!source_ || !IsFramerateScalingEnabled(degradation_preference_))
+ return false;
+
+ const int fps_wanted = std::max(kMinFramerateFps, fps);
+ if (fps_wanted >= sink_wants_.max_framerate_fps)
+ return false;
+
+ RTC_LOG(LS_INFO) << "Scaling down framerate: " << fps_wanted;
+ sink_wants_.max_framerate_fps = fps_wanted;
+ source_->AddOrUpdateSink(video_stream_encoder_,
+ GetActiveSinkWantsInternal());
+ return true;
+ }
+
+ bool IncreaseFramerate(int fps) {
+ // Called on the encoder task queue.
+ rtc::CritScope lock(&crit_);
+ if (!source_ || !IsFramerateScalingEnabled(degradation_preference_))
+ return false;
+
+ const int fps_wanted = std::max(kMinFramerateFps, fps);
+ if (fps_wanted <= sink_wants_.max_framerate_fps)
+ return false;
+
+ RTC_LOG(LS_INFO) << "Scaling up framerate: " << fps_wanted;
+ sink_wants_.max_framerate_fps = fps_wanted;
+ source_->AddOrUpdateSink(video_stream_encoder_,
+ GetActiveSinkWantsInternal());
+ return true;
+ }
+
+ private:
+ rtc::VideoSinkWants GetActiveSinkWantsInternal()
+ RTC_EXCLUSIVE_LOCKS_REQUIRED(&crit_) {
+ rtc::VideoSinkWants wants = sink_wants_;
+ // Clear any constraints from the current sink wants that don't apply to
+ // the used degradation_preference.
+ switch (degradation_preference_) {
+ case VideoSendStream::DegradationPreference::kBalanced:
+ break;
+ case VideoSendStream::DegradationPreference::kMaintainFramerate:
+ wants.max_framerate_fps = std::numeric_limits<int>::max();
+ break;
+ case VideoSendStream::DegradationPreference::kMaintainResolution:
+ wants.max_pixel_count = std::numeric_limits<int>::max();
+ wants.target_pixel_count.reset();
+ break;
+ case VideoSendStream::DegradationPreference::kDegradationDisabled:
+ wants.max_pixel_count = std::numeric_limits<int>::max();
+ wants.target_pixel_count.reset();
+ wants.max_framerate_fps = std::numeric_limits<int>::max();
+ }
+ return wants;
+ }
+
+ rtc::CriticalSection crit_;
+ rtc::SequencedTaskChecker main_checker_;
+ VideoStreamEncoder* const video_stream_encoder_;
+ rtc::VideoSinkWants sink_wants_ RTC_GUARDED_BY(&crit_);
+ VideoSendStream::DegradationPreference degradation_preference_
+ RTC_GUARDED_BY(&crit_);
+ rtc::VideoSourceInterface<VideoFrame>* source_ RTC_GUARDED_BY(&crit_);
+
+ RTC_DISALLOW_COPY_AND_ASSIGN(VideoSourceProxy);
+};
+
+VideoStreamEncoder::VideoStreamEncoder(
+ uint32_t number_of_cores,
+ SendStatisticsProxy* stats_proxy,
+ const VideoSendStream::Config::EncoderSettings& settings,
+ rtc::VideoSinkInterface<VideoFrame>* pre_encode_callback,
+ EncodedFrameObserver* encoder_timing,
+ std::unique_ptr<OveruseFrameDetector> overuse_detector)
+ : shutdown_event_(true /* manual_reset */, false),
+ number_of_cores_(number_of_cores),
+ initial_rampup_(0),
+ source_proxy_(new VideoSourceProxy(this)),
+ sink_(nullptr),
+ settings_(settings),
+ codec_type_(PayloadStringToCodecType(settings.payload_name)),
+ video_sender_(Clock::GetRealTimeClock(), this),
+ overuse_detector_(
+ overuse_detector.get()
+ ? overuse_detector.release()
+ : new OveruseFrameDetector(
+ GetCpuOveruseOptions(settings.full_overuse_time),
+ this,
+ encoder_timing,
+ stats_proxy)),
+ stats_proxy_(stats_proxy),
+ pre_encode_callback_(pre_encode_callback),
+ max_framerate_(-1),
+ pending_encoder_reconfiguration_(false),
+ encoder_start_bitrate_bps_(0),
+ max_data_payload_length_(0),
+ nack_enabled_(false),
+ last_observed_bitrate_bps_(0),
+ encoder_paused_and_dropped_frame_(false),
+ clock_(Clock::GetRealTimeClock()),
+ degradation_preference_(
+ VideoSendStream::DegradationPreference::kDegradationDisabled),
+ posted_frames_waiting_for_encode_(0),
+ last_captured_timestamp_(0),
+ delta_ntp_internal_ms_(clock_->CurrentNtpInMilliseconds() -
+ clock_->TimeInMilliseconds()),
+ last_frame_log_ms_(clock_->TimeInMilliseconds()),
+ captured_frame_count_(0),
+ dropped_frame_count_(0),
+ bitrate_observer_(nullptr),
+ encoder_queue_("EncoderQueue") {
+ RTC_DCHECK(stats_proxy);
+ encoder_queue_.PostTask([this] {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ overuse_detector_->StartCheckForOveruse();
+ video_sender_.RegisterExternalEncoder(
+ settings_.encoder, settings_.payload_type, settings_.internal_source);
+ });
+}
+
+VideoStreamEncoder::~VideoStreamEncoder() {
+ RTC_DCHECK_RUN_ON(&thread_checker_);
+ RTC_DCHECK(shutdown_event_.Wait(0))
+ << "Must call ::Stop() before destruction.";
+}
+
+// TODO(pbos): Lower these thresholds (to closer to 100%) when we handle
+// pipelining encoders better (multiple input frames before something comes
+// out). This should effectively turn off CPU adaptations for systems that
+// remotely cope with the load right now.
+CpuOveruseOptions VideoStreamEncoder::GetCpuOveruseOptions(
+ bool full_overuse_time) {
+ CpuOveruseOptions options;
+ if (full_overuse_time) {
+ options.low_encode_usage_threshold_percent = 150;
+ options.high_encode_usage_threshold_percent = 200;
+ }
+ return options;
+}
+
+void VideoStreamEncoder::Stop() {
+ RTC_DCHECK_RUN_ON(&thread_checker_);
+ source_proxy_->SetSource(nullptr, VideoSendStream::DegradationPreference());
+ encoder_queue_.PostTask([this] {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ overuse_detector_->StopCheckForOveruse();
+ rate_allocator_.reset();
+ bitrate_observer_ = nullptr;
+ video_sender_.RegisterExternalEncoder(nullptr, settings_.payload_type,
+ false);
+ quality_scaler_ = nullptr;
+ shutdown_event_.Set();
+ });
+
+ shutdown_event_.Wait(rtc::Event::kForever);
+}
+
+void VideoStreamEncoder::SetBitrateObserver(
+ VideoBitrateAllocationObserver* bitrate_observer) {
+ RTC_DCHECK_RUN_ON(&thread_checker_);
+ encoder_queue_.PostTask([this, bitrate_observer] {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ RTC_DCHECK(!bitrate_observer_);
+ bitrate_observer_ = bitrate_observer;
+ });
+}
+
+void VideoStreamEncoder::SetSource(
+ rtc::VideoSourceInterface<VideoFrame>* source,
+ const VideoSendStream::DegradationPreference& degradation_preference) {
+ RTC_DCHECK_RUN_ON(&thread_checker_);
+ source_proxy_->SetSource(source, degradation_preference);
+ encoder_queue_.PostTask([this, degradation_preference] {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ if (degradation_preference_ != degradation_preference) {
+ // Reset adaptation state, so that we're not tricked into thinking there's
+ // an already pending request of the same type.
+ last_adaptation_request_.reset();
+ if (degradation_preference ==
+ VideoSendStream::DegradationPreference::kBalanced ||
+ degradation_preference_ ==
+ VideoSendStream::DegradationPreference::kBalanced) {
+ // TODO(asapersson): Consider removing |adapt_counters_| map and use one
+ // AdaptCounter for all modes.
+ source_proxy_->ResetPixelFpsCount();
+ adapt_counters_.clear();
+ }
+ }
+ degradation_preference_ = degradation_preference;
+ bool allow_scaling = IsResolutionScalingEnabled(degradation_preference_);
+ initial_rampup_ = allow_scaling ? 0 : kMaxInitialFramedrop;
+ ConfigureQualityScaler();
+ if (!IsFramerateScalingEnabled(degradation_preference) &&
+ max_framerate_ != -1) {
+ // If frame rate scaling is no longer allowed, remove any potential
+ // allowance for longer frame intervals.
+ overuse_detector_->OnTargetFramerateUpdated(max_framerate_);
+ }
+ });
+}
+
+void VideoStreamEncoder::SetSink(EncoderSink* sink, bool rotation_applied) {
+ source_proxy_->SetWantsRotationApplied(rotation_applied);
+ encoder_queue_.PostTask([this, sink] {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ sink_ = sink;
+ });
+}
+
+void VideoStreamEncoder::SetStartBitrate(int start_bitrate_bps) {
+ encoder_queue_.PostTask([this, start_bitrate_bps] {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ encoder_start_bitrate_bps_ = start_bitrate_bps;
+ });
+}
+
+void VideoStreamEncoder::ConfigureEncoder(VideoEncoderConfig config,
+ size_t max_data_payload_length,
+ bool nack_enabled) {
+ encoder_queue_.PostTask(
+ std::unique_ptr<rtc::QueuedTask>(new ConfigureEncoderTask(
+ this, std::move(config), max_data_payload_length, nack_enabled)));
+}
+
+void VideoStreamEncoder::ConfigureEncoderOnTaskQueue(
+ VideoEncoderConfig config,
+ size_t max_data_payload_length,
+ bool nack_enabled) {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ RTC_DCHECK(sink_);
+ RTC_LOG(LS_INFO) << "ConfigureEncoder requested.";
+
+ max_data_payload_length_ = max_data_payload_length;
+ nack_enabled_ = nack_enabled;
+ encoder_config_ = std::move(config);
+ pending_encoder_reconfiguration_ = true;
+
+ // Reconfigure the encoder now if the encoder has an internal source or
+ // if the frame resolution is known. Otherwise, the reconfiguration is
+ // deferred until the next frame to minimize the number of reconfigurations.
+ // The codec configuration depends on incoming video frame size.
+ if (last_frame_info_) {
+ ReconfigureEncoder();
+ } else if (settings_.internal_source) {
+ last_frame_info_ =
+ rtc::Optional<VideoFrameInfo>(VideoFrameInfo(176, 144, false));
+ ReconfigureEncoder();
+ }
+}
+
+void VideoStreamEncoder::ReconfigureEncoder() {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ RTC_DCHECK(pending_encoder_reconfiguration_);
+ std::vector<VideoStream> streams =
+ encoder_config_.video_stream_factory->CreateEncoderStreams(
+ last_frame_info_->width, last_frame_info_->height, encoder_config_);
+
+ // TODO(ilnik): If configured resolution is significantly less than provided,
+ // e.g. because there are not enough SSRCs for all simulcast streams,
+ // signal new resolutions via SinkWants to video source.
+
+ // Stream dimensions may be not equal to given because of a simulcast
+ // restrictions.
+ auto highest_stream = std::max_element(
+ streams.begin(), streams.end(),
+ [](const webrtc::VideoStream& a, const webrtc::VideoStream& b) {
+ return std::tie(a.width, a.height) < std::tie(b.width, b.height);
+ });
+ int highest_stream_width = static_cast<int>(highest_stream->width);
+ int highest_stream_height = static_cast<int>(highest_stream->height);
+ // Dimension may be reduced to be, e.g. divisible by 4.
+ RTC_CHECK_GE(last_frame_info_->width, highest_stream_width);
+ RTC_CHECK_GE(last_frame_info_->height, highest_stream_height);
+ crop_width_ = last_frame_info_->width - highest_stream_width;
+ crop_height_ = last_frame_info_->height - highest_stream_height;
+
+ VideoCodec codec;
+ if (!VideoCodecInitializer::SetupCodec(encoder_config_, settings_, streams,
+ nack_enabled_, &codec,
+ &rate_allocator_)) {
+ RTC_LOG(LS_ERROR) << "Failed to create encoder configuration.";
+ }
+
+ codec.startBitrate =
+ std::max(encoder_start_bitrate_bps_ / 1000, codec.minBitrate);
+ codec.startBitrate = std::min(codec.startBitrate, codec.maxBitrate);
+ codec.expect_encode_from_texture = last_frame_info_->is_texture;
+ max_framerate_ = codec.maxFramerate;
+ RTC_DCHECK_LE(max_framerate_, kMaxFramerateFps);
+
+ bool success = video_sender_.RegisterSendCodec(
+ &codec, number_of_cores_,
+ static_cast<uint32_t>(max_data_payload_length_)) == VCM_OK;
+ if (!success) {
+ RTC_LOG(LS_ERROR) << "Failed to configure encoder.";
+ rate_allocator_.reset();
+ }
+
+ video_sender_.UpdateChannelParemeters(rate_allocator_.get(),
+ bitrate_observer_);
+
+ // Get the current actual framerate, as measured by the stats proxy. This is
+ // used to get the correct bitrate layer allocation.
+ int current_framerate = stats_proxy_->GetSendFrameRate();
+ if (current_framerate == 0)
+ current_framerate = codec.maxFramerate;
+ stats_proxy_->OnEncoderReconfigured(
+ encoder_config_,
+ rate_allocator_.get()
+ ? rate_allocator_->GetPreferredBitrateBps(current_framerate)
+ : codec.maxBitrate);
+
+ pending_encoder_reconfiguration_ = false;
+
+ sink_->OnEncoderConfigurationChanged(
+ std::move(streams), encoder_config_.min_transmit_bitrate_bps);
+
+ // Get the current target framerate, ie the maximum framerate as specified by
+ // the current codec configuration, or any limit imposed by cpu adaption in
+ // maintain-resolution or balanced mode. This is used to make sure overuse
+ // detection doesn't needlessly trigger in low and/or variable framerate
+ // scenarios.
+ int target_framerate = std::min(
+ max_framerate_, source_proxy_->GetActiveSinkWants().max_framerate_fps);
+ overuse_detector_->OnTargetFramerateUpdated(target_framerate);
+
+ ConfigureQualityScaler();
+}
+
+void VideoStreamEncoder::ConfigureQualityScaler() {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ const auto scaling_settings = settings_.encoder->GetScalingSettings();
+ const bool quality_scaling_allowed =
+ IsResolutionScalingEnabled(degradation_preference_) &&
+ scaling_settings.enabled;
+
+ if (quality_scaling_allowed) {
+ if (quality_scaler_.get() == nullptr) {
+ // Quality scaler has not already been configured.
+ // Drop frames and scale down until desired quality is achieved.
+ if (scaling_settings.thresholds) {
+ quality_scaler_.reset(
+ new QualityScaler(this, *(scaling_settings.thresholds)));
+ } else {
+ quality_scaler_.reset(new QualityScaler(this, codec_type_));
+ }
+ }
+ } else {
+ quality_scaler_.reset(nullptr);
+ initial_rampup_ = kMaxInitialFramedrop;
+ }
+
+ stats_proxy_->SetAdaptationStats(GetActiveCounts(kCpu),
+ GetActiveCounts(kQuality));
+}
+
+void VideoStreamEncoder::OnFrame(const VideoFrame& video_frame) {
+ RTC_DCHECK_RUNS_SERIALIZED(&incoming_frame_race_checker_);
+ VideoFrame incoming_frame = video_frame;
+
+ // Local time in webrtc time base.
+ int64_t current_time_us = clock_->TimeInMicroseconds();
+ int64_t current_time_ms = current_time_us / rtc::kNumMicrosecsPerMillisec;
+ // In some cases, e.g., when the frame from decoder is fed to encoder,
+ // the timestamp may be set to the future. As the encoding pipeline assumes
+ // capture time to be less than present time, we should reset the capture
+ // timestamps here. Otherwise there may be issues with RTP send stream.
+
+ if (incoming_frame.timestamp_us() > current_time_us)
+ incoming_frame.set_timestamp_us(current_time_us);
+
+ // Capture time may come from clock with an offset and drift from clock_.
+ int64_t capture_ntp_time_ms;
+ if (video_frame.ntp_time_ms() > 0) {
+ capture_ntp_time_ms = video_frame.ntp_time_ms();
+ } else if (video_frame.render_time_ms() != 0) {
+ capture_ntp_time_ms = video_frame.render_time_ms() + delta_ntp_internal_ms_;
+ } else {
+ capture_ntp_time_ms = current_time_ms + delta_ntp_internal_ms_;
+ }
+ incoming_frame.set_ntp_time_ms(capture_ntp_time_ms);
+
+ // Convert NTP time, in ms, to RTP timestamp.
+ const int kMsToRtpTimestamp = 90;
+ incoming_frame.set_timestamp(
+ kMsToRtpTimestamp * static_cast<uint32_t>(incoming_frame.ntp_time_ms()));
+
+ if (incoming_frame.ntp_time_ms() <= last_captured_timestamp_) {
+ // We don't allow the same capture time for two frames, drop this one.
+ RTC_LOG(LS_WARNING) << "Same/old NTP timestamp ("
+ << incoming_frame.ntp_time_ms()
+ << " <= " << last_captured_timestamp_
+ << ") for incoming frame. Dropping.";
+ return;
+ }
+
+ bool log_stats = false;
+ if (current_time_ms - last_frame_log_ms_ > kFrameLogIntervalMs) {
+ last_frame_log_ms_ = current_time_ms;
+ log_stats = true;
+ }
+
+ last_captured_timestamp_ = incoming_frame.ntp_time_ms();
+ encoder_queue_.PostTask(std::unique_ptr<rtc::QueuedTask>(new EncodeTask(
+ incoming_frame, this, rtc::TimeMicros(), log_stats)));
+}
+
+void VideoStreamEncoder::OnDiscardedFrame() {
+ stats_proxy_->OnFrameDroppedBySource();
+}
+
+bool VideoStreamEncoder::EncoderPaused() const {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ // Pause video if paused by caller or as long as the network is down or the
+ // pacer queue has grown too large in buffered mode.
+ // If the pacer queue has grown too large or the network is down,
+ // last_observed_bitrate_bps_ will be 0.
+ return last_observed_bitrate_bps_ == 0;
+}
+
+void VideoStreamEncoder::TraceFrameDropStart() {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ // Start trace event only on the first frame after encoder is paused.
+ if (!encoder_paused_and_dropped_frame_) {
+ TRACE_EVENT_ASYNC_BEGIN0("webrtc", "EncoderPaused", this);
+ }
+ encoder_paused_and_dropped_frame_ = true;
+}
+
+void VideoStreamEncoder::TraceFrameDropEnd() {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ // End trace event on first frame after encoder resumes, if frame was dropped.
+ if (encoder_paused_and_dropped_frame_) {
+ TRACE_EVENT_ASYNC_END0("webrtc", "EncoderPaused", this);
+ }
+ encoder_paused_and_dropped_frame_ = false;
+}
+
+void VideoStreamEncoder::EncodeVideoFrame(const VideoFrame& video_frame,
+ int64_t time_when_posted_us) {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+
+ if (pre_encode_callback_)
+ pre_encode_callback_->OnFrame(video_frame);
+
+ if (!last_frame_info_ || video_frame.width() != last_frame_info_->width ||
+ video_frame.height() != last_frame_info_->height ||
+ video_frame.is_texture() != last_frame_info_->is_texture) {
+ pending_encoder_reconfiguration_ = true;
+ last_frame_info_ = rtc::Optional<VideoFrameInfo>(VideoFrameInfo(
+ video_frame.width(), video_frame.height(), video_frame.is_texture()));
+ RTC_LOG(LS_INFO) << "Video frame parameters changed: dimensions="
+ << last_frame_info_->width << "x"
+ << last_frame_info_->height
+ << ", texture=" << last_frame_info_->is_texture << ".";
+ }
+
+ if (initial_rampup_ < kMaxInitialFramedrop &&
+ video_frame.size() >
+ MaximumFrameSizeForBitrate(encoder_start_bitrate_bps_ / 1000)) {
+ RTC_LOG(LS_INFO) << "Dropping frame. Too large for target bitrate.";
+ AdaptDown(kQuality);
+ ++initial_rampup_;
+ return;
+ }
+ initial_rampup_ = kMaxInitialFramedrop;
+
+ int64_t now_ms = clock_->TimeInMilliseconds();
+ if (pending_encoder_reconfiguration_) {
+ ReconfigureEncoder();
+ last_parameters_update_ms_.emplace(now_ms);
+ } else if (!last_parameters_update_ms_ ||
+ now_ms - *last_parameters_update_ms_ >=
+ vcm::VCMProcessTimer::kDefaultProcessIntervalMs) {
+ video_sender_.UpdateChannelParemeters(rate_allocator_.get(),
+ bitrate_observer_);
+ last_parameters_update_ms_.emplace(now_ms);
+ }
+
+ if (EncoderPaused()) {
+ TraceFrameDropStart();
+ return;
+ }
+ TraceFrameDropEnd();
+
+ VideoFrame out_frame(video_frame);
+ // Crop frame if needed.
+ if (crop_width_ > 0 || crop_height_ > 0) {
+ int cropped_width = video_frame.width() - crop_width_;
+ int cropped_height = video_frame.height() - crop_height_;
+ rtc::scoped_refptr<I420Buffer> cropped_buffer =
+ I420Buffer::Create(cropped_width, cropped_height);
+ // TODO(ilnik): Remove scaling if cropping is too big, as it should never
+ // happen after SinkWants signaled correctly from ReconfigureEncoder.
+ if (crop_width_ < 4 && crop_height_ < 4) {
+ cropped_buffer->CropAndScaleFrom(
+ *video_frame.video_frame_buffer()->ToI420(), crop_width_ / 2,
+ crop_height_ / 2, cropped_width, cropped_height);
+ } else {
+ cropped_buffer->ScaleFrom(
+ *video_frame.video_frame_buffer()->ToI420().get());
+ }
+ out_frame =
+ VideoFrame(cropped_buffer, video_frame.timestamp(),
+ video_frame.render_time_ms(), video_frame.rotation());
+ out_frame.set_ntp_time_ms(video_frame.ntp_time_ms());
+ }
+
+ TRACE_EVENT_ASYNC_STEP0("webrtc", "Video", video_frame.render_time_ms(),
+ "Encode");
+
+ overuse_detector_->FrameCaptured(out_frame, time_when_posted_us);
+
+ video_sender_.AddVideoFrame(out_frame, nullptr);
+}
+
+void VideoStreamEncoder::SendKeyFrame() {
+ if (!encoder_queue_.IsCurrent()) {
+ encoder_queue_.PostTask([this] { SendKeyFrame(); });
+ return;
+ }
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ video_sender_.IntraFrameRequest(0);
+}
+
+EncodedImageCallback::Result VideoStreamEncoder::OnEncodedImage(
+ const EncodedImage& encoded_image,
+ const CodecSpecificInfo* codec_specific_info,
+ const RTPFragmentationHeader* fragmentation) {
+ // Encoded is called on whatever thread the real encoder implementation run
+ // on. In the case of hardware encoders, there might be several encoders
+ // running in parallel on different threads.
+ stats_proxy_->OnSendEncodedImage(encoded_image, codec_specific_info);
+
+ EncodedImageCallback::Result result =
+ sink_->OnEncodedImage(encoded_image, codec_specific_info, fragmentation);
+
+ int64_t time_sent_us = rtc::TimeMicros();
+ uint32_t timestamp = encoded_image._timeStamp;
+ const int qp = encoded_image.qp_;
+ encoder_queue_.PostTask([this, timestamp, time_sent_us, qp] {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ overuse_detector_->FrameSent(timestamp, time_sent_us);
+ if (quality_scaler_ && qp >= 0)
+ quality_scaler_->ReportQP(qp);
+ });
+
+ return result;
+}
+
+void VideoStreamEncoder::OnDroppedFrame(DropReason reason) {
+ switch (reason) {
+ case DropReason::kDroppedByMediaOptimizations:
+ stats_proxy_->OnFrameDroppedByMediaOptimizations();
+ encoder_queue_.PostTask([this] {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ if (quality_scaler_)
+ quality_scaler_->ReportDroppedFrame();
+ });
+ break;
+ case DropReason::kDroppedByEncoder:
+ stats_proxy_->OnFrameDroppedByEncoder();
+ break;
+ }
+}
+
+void VideoStreamEncoder::OnReceivedIntraFrameRequest(size_t stream_index) {
+ if (!encoder_queue_.IsCurrent()) {
+ encoder_queue_.PostTask(
+ [this, stream_index] { OnReceivedIntraFrameRequest(stream_index); });
+ return;
+ }
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ // Key frame request from remote side, signal to VCM.
+ TRACE_EVENT0("webrtc", "OnKeyFrameRequest");
+ video_sender_.IntraFrameRequest(stream_index);
+}
+
+void VideoStreamEncoder::OnBitrateUpdated(uint32_t bitrate_bps,
+ uint8_t fraction_lost,
+ int64_t round_trip_time_ms) {
+ if (!encoder_queue_.IsCurrent()) {
+ encoder_queue_.PostTask(
+ [this, bitrate_bps, fraction_lost, round_trip_time_ms] {
+ OnBitrateUpdated(bitrate_bps, fraction_lost, round_trip_time_ms);
+ });
+ return;
+ }
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ RTC_DCHECK(sink_) << "sink_ must be set before the encoder is active.";
+
+ RTC_LOG(LS_VERBOSE) << "OnBitrateUpdated, bitrate " << bitrate_bps
+ << " packet loss " << static_cast<int>(fraction_lost)
+ << " rtt " << round_trip_time_ms;
+
+ video_sender_.SetChannelParameters(bitrate_bps, fraction_lost,
+ round_trip_time_ms, rate_allocator_.get(),
+ bitrate_observer_);
+
+ encoder_start_bitrate_bps_ =
+ bitrate_bps != 0 ? bitrate_bps : encoder_start_bitrate_bps_;
+ bool video_is_suspended = bitrate_bps == 0;
+ bool video_suspension_changed = video_is_suspended != EncoderPaused();
+ last_observed_bitrate_bps_ = bitrate_bps;
+
+ if (video_suspension_changed) {
+ RTC_LOG(LS_INFO) << "Video suspend state changed to: "
+ << (video_is_suspended ? "suspended" : "not suspended");
+ stats_proxy_->OnSuspendChange(video_is_suspended);
+ }
+}
+
+void VideoStreamEncoder::AdaptDown(AdaptReason reason) {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+ AdaptationRequest adaptation_request = {
+ last_frame_info_->pixel_count(),
+ stats_proxy_->GetStats().input_frame_rate,
+ AdaptationRequest::Mode::kAdaptDown};
+
+ bool downgrade_requested =
+ last_adaptation_request_ &&
+ last_adaptation_request_->mode_ == AdaptationRequest::Mode::kAdaptDown;
+
+ switch (degradation_preference_) {
+ case VideoSendStream::DegradationPreference::kBalanced:
+ break;
+ case VideoSendStream::DegradationPreference::kMaintainFramerate:
+ if (downgrade_requested &&
+ adaptation_request.input_pixel_count_ >=
+ last_adaptation_request_->input_pixel_count_) {
+ // Don't request lower resolution if the current resolution is not
+ // lower than the last time we asked for the resolution to be lowered.
+ return;
+ }
+ break;
+ case VideoSendStream::DegradationPreference::kMaintainResolution:
+ if (adaptation_request.framerate_fps_ <= 0 ||
+ (downgrade_requested &&
+ adaptation_request.framerate_fps_ < kMinFramerateFps)) {
+ // If no input fps estimate available, can't determine how to scale down
+ // framerate. Otherwise, don't request lower framerate if we don't have
+ // a valid frame rate. Since framerate, unlike resolution, is a measure
+ // we have to estimate, and can fluctuate naturally over time, don't
+ // make the same kind of limitations as for resolution, but trust the
+ // overuse detector to not trigger too often.
+ return;
+ }
+ break;
+ case VideoSendStream::DegradationPreference::kDegradationDisabled:
+ return;
+ }
+
+ if (reason == kCpu) {
+ if (GetConstAdaptCounter().ResolutionCount(kCpu) >=
+ kMaxCpuResolutionDowngrades ||
+ GetConstAdaptCounter().FramerateCount(kCpu) >=
+ kMaxCpuFramerateDowngrades) {
+ return;
+ }
+ }
+
+ switch (degradation_preference_) {
+ case VideoSendStream::DegradationPreference::kBalanced: {
+ // Try scale down framerate, if lower.
+ int fps = MinFps(last_frame_info_->pixel_count());
+ if (source_proxy_->RestrictFramerate(fps)) {
+ GetAdaptCounter().IncrementFramerate(reason);
+ break;
+ }
+ // Scale down resolution.
+ FALLTHROUGH();
+ }
+ case VideoSendStream::DegradationPreference::kMaintainFramerate: {
+ // Scale down resolution.
+ bool min_pixels_reached = false;
+ if (!source_proxy_->RequestResolutionLowerThan(
+ adaptation_request.input_pixel_count_,
+ settings_.encoder->GetScalingSettings().min_pixels_per_frame,
+ &min_pixels_reached)) {
+ if (min_pixels_reached)
+ stats_proxy_->OnMinPixelLimitReached();
+ return;
+ }
+ GetAdaptCounter().IncrementResolution(reason);
+ break;
+ }
+ case VideoSendStream::DegradationPreference::kMaintainResolution: {
+ // Scale down framerate.
+ const int requested_framerate = source_proxy_->RequestFramerateLowerThan(
+ adaptation_request.framerate_fps_);
+ if (requested_framerate == -1)
+ return;
+ RTC_DCHECK_NE(max_framerate_, -1);
+ overuse_detector_->OnTargetFramerateUpdated(
+ std::min(max_framerate_, requested_framerate));
+ GetAdaptCounter().IncrementFramerate(reason);
+ break;
+ }
+ case VideoSendStream::DegradationPreference::kDegradationDisabled:
+ RTC_NOTREACHED();
+ }
+
+ last_adaptation_request_.emplace(adaptation_request);
+
+ UpdateAdaptationStats(reason);
+
+ RTC_LOG(LS_INFO) << GetConstAdaptCounter().ToString();
+}
+
+void VideoStreamEncoder::AdaptUp(AdaptReason reason) {
+ RTC_DCHECK_RUN_ON(&encoder_queue_);
+
+ const AdaptCounter& adapt_counter = GetConstAdaptCounter();
+ int num_downgrades = adapt_counter.TotalCount(reason);
+ if (num_downgrades == 0)
+ return;
+ RTC_DCHECK_GT(num_downgrades, 0);
+
+ AdaptationRequest adaptation_request = {
+ last_frame_info_->pixel_count(),
+ stats_proxy_->GetStats().input_frame_rate,
+ AdaptationRequest::Mode::kAdaptUp};
+
+ bool adapt_up_requested =
+ last_adaptation_request_ &&
+ last_adaptation_request_->mode_ == AdaptationRequest::Mode::kAdaptUp;
+
+ if (degradation_preference_ ==
+ VideoSendStream::DegradationPreference::kMaintainFramerate) {
+ if (adapt_up_requested &&
+ adaptation_request.input_pixel_count_ <=
+ last_adaptation_request_->input_pixel_count_) {
+ // Don't request higher resolution if the current resolution is not
+ // higher than the last time we asked for the resolution to be higher.
+ return;
+ }
+ }
+
+ switch (degradation_preference_) {
+ case VideoSendStream::DegradationPreference::kBalanced: {
+ // Try scale up framerate, if higher.
+ int fps = MaxFps(last_frame_info_->pixel_count());
+ if (source_proxy_->IncreaseFramerate(fps)) {
+ GetAdaptCounter().DecrementFramerate(reason, fps);
+ // Reset framerate in case of fewer fps steps down than up.
+ if (adapt_counter.FramerateCount() == 0 &&
+ fps != std::numeric_limits<int>::max()) {
+ RTC_LOG(LS_INFO) << "Removing framerate down-scaling setting.";
+ source_proxy_->IncreaseFramerate(std::numeric_limits<int>::max());
+ }
+ break;
+ }
+ // Scale up resolution.
+ FALLTHROUGH();
+ }
+ case VideoSendStream::DegradationPreference::kMaintainFramerate: {
+ // Scale up resolution.
+ int pixel_count = adaptation_request.input_pixel_count_;
+ if (adapt_counter.ResolutionCount() == 1) {
+ RTC_LOG(LS_INFO) << "Removing resolution down-scaling setting.";
+ pixel_count = std::numeric_limits<int>::max();
+ }
+ if (!source_proxy_->RequestHigherResolutionThan(pixel_count))
+ return;
+ GetAdaptCounter().DecrementResolution(reason);
+ break;
+ }
+ case VideoSendStream::DegradationPreference::kMaintainResolution: {
+ // Scale up framerate.
+ int fps = adaptation_request.framerate_fps_;
+ if (adapt_counter.FramerateCount() == 1) {
+ RTC_LOG(LS_INFO) << "Removing framerate down-scaling setting.";
+ fps = std::numeric_limits<int>::max();
+ }
+
+ const int requested_framerate =
+ source_proxy_->RequestHigherFramerateThan(fps);
+ if (requested_framerate == -1) {
+ overuse_detector_->OnTargetFramerateUpdated(max_framerate_);
+ return;
+ }
+ overuse_detector_->OnTargetFramerateUpdated(
+ std::min(max_framerate_, requested_framerate));
+ GetAdaptCounter().DecrementFramerate(reason);
+ break;
+ }
+ case VideoSendStream::DegradationPreference::kDegradationDisabled:
+ return;
+ }
+
+ last_adaptation_request_.emplace(adaptation_request);
+
+ UpdateAdaptationStats(reason);
+
+ RTC_LOG(LS_INFO) << adapt_counter.ToString();
+}
+
+void VideoStreamEncoder::UpdateAdaptationStats(AdaptReason reason) {
+ switch (reason) {
+ case kCpu:
+ stats_proxy_->OnCpuAdaptationChanged(GetActiveCounts(kCpu),
+ GetActiveCounts(kQuality));
+ break;
+ case kQuality:
+ stats_proxy_->OnQualityAdaptationChanged(GetActiveCounts(kCpu),
+ GetActiveCounts(kQuality));
+ break;
+ }
+}
+
+VideoStreamEncoder::AdaptCounts VideoStreamEncoder::GetActiveCounts(
+ AdaptReason reason) {
+ VideoStreamEncoder::AdaptCounts counts =
+ GetConstAdaptCounter().Counts(reason);
+ switch (reason) {
+ case kCpu:
+ if (!IsFramerateScalingEnabled(degradation_preference_))
+ counts.fps = -1;
+ if (!IsResolutionScalingEnabled(degradation_preference_))
+ counts.resolution = -1;
+ break;
+ case kQuality:
+ if (!IsFramerateScalingEnabled(degradation_preference_) ||
+ !quality_scaler_) {
+ counts.fps = -1;
+ }
+ if (!IsResolutionScalingEnabled(degradation_preference_) ||
+ !quality_scaler_) {
+ counts.resolution = -1;
+ }
+ break;
+ }
+ return counts;
+}
+
+VideoStreamEncoder::AdaptCounter& VideoStreamEncoder::GetAdaptCounter() {
+ return adapt_counters_[degradation_preference_];
+}
+
+const VideoStreamEncoder::AdaptCounter&
+VideoStreamEncoder::GetConstAdaptCounter() {
+ return adapt_counters_[degradation_preference_];
+}
+
+// Class holding adaptation information.
+VideoStreamEncoder::AdaptCounter::AdaptCounter() {
+ fps_counters_.resize(kScaleReasonSize);
+ resolution_counters_.resize(kScaleReasonSize);
+ static_assert(kScaleReasonSize == 2, "Update MoveCount.");
+}
+
+VideoStreamEncoder::AdaptCounter::~AdaptCounter() {}
+
+std::string VideoStreamEncoder::AdaptCounter::ToString() const {
+ std::stringstream ss;
+ ss << "Downgrade counts: fps: {" << ToString(fps_counters_);
+ ss << "}, resolution: {" << ToString(resolution_counters_) << "}";
+ return ss.str();
+}
+
+VideoStreamEncoder::AdaptCounts VideoStreamEncoder::AdaptCounter::Counts(
+ int reason) const {
+ AdaptCounts counts;
+ counts.fps = fps_counters_[reason];
+ counts.resolution = resolution_counters_[reason];
+ return counts;
+}
+
+void VideoStreamEncoder::AdaptCounter::IncrementFramerate(int reason) {
+ ++(fps_counters_[reason]);
+}
+
+void VideoStreamEncoder::AdaptCounter::IncrementResolution(int reason) {
+ ++(resolution_counters_[reason]);
+}
+
+void VideoStreamEncoder::AdaptCounter::DecrementFramerate(int reason) {
+ if (fps_counters_[reason] == 0) {
+ // Balanced mode: Adapt up is in a different order, switch reason.
+ // E.g. framerate adapt down: quality (2), framerate adapt up: cpu (3).
+ // 1. Down resolution (cpu): res={quality:0,cpu:1}, fps={quality:0,cpu:0}
+ // 2. Down fps (quality): res={quality:0,cpu:1}, fps={quality:1,cpu:0}
+ // 3. Up fps (cpu): res={quality:1,cpu:0}, fps={quality:0,cpu:0}
+ // 4. Up resolution (quality): res={quality:0,cpu:0}, fps={quality:0,cpu:0}
+ RTC_DCHECK_GT(TotalCount(reason), 0) << "No downgrade for reason.";
+ RTC_DCHECK_GT(FramerateCount(), 0) << "Framerate not downgraded.";
+ MoveCount(&resolution_counters_, reason);
+ MoveCount(&fps_counters_, (reason + 1) % kScaleReasonSize);
+ }
+ --(fps_counters_[reason]);
+ RTC_DCHECK_GE(fps_counters_[reason], 0);
+}
+
+void VideoStreamEncoder::AdaptCounter::DecrementResolution(int reason) {
+ if (resolution_counters_[reason] == 0) {
+ // Balanced mode: Adapt up is in a different order, switch reason.
+ RTC_DCHECK_GT(TotalCount(reason), 0) << "No downgrade for reason.";
+ RTC_DCHECK_GT(ResolutionCount(), 0) << "Resolution not downgraded.";
+ MoveCount(&fps_counters_, reason);
+ MoveCount(&resolution_counters_, (reason + 1) % kScaleReasonSize);
+ }
+ --(resolution_counters_[reason]);
+ RTC_DCHECK_GE(resolution_counters_[reason], 0);
+}
+
+void VideoStreamEncoder::AdaptCounter::DecrementFramerate(int reason,
+ int cur_fps) {
+ DecrementFramerate(reason);
+ // Reset if at max fps (i.e. in case of fewer steps up than down).
+ if (cur_fps == std::numeric_limits<int>::max())
+ std::fill(fps_counters_.begin(), fps_counters_.end(), 0);
+}
+
+int VideoStreamEncoder::AdaptCounter::FramerateCount() const {
+ return Count(fps_counters_);
+}
+
+int VideoStreamEncoder::AdaptCounter::ResolutionCount() const {
+ return Count(resolution_counters_);
+}
+
+int VideoStreamEncoder::AdaptCounter::FramerateCount(int reason) const {
+ return fps_counters_[reason];
+}
+
+int VideoStreamEncoder::AdaptCounter::ResolutionCount(int reason) const {
+ return resolution_counters_[reason];
+}
+
+int VideoStreamEncoder::AdaptCounter::TotalCount(int reason) const {
+ return FramerateCount(reason) + ResolutionCount(reason);
+}
+
+int VideoStreamEncoder::AdaptCounter::Count(
+ const std::vector<int>& counters) const {
+ return std::accumulate(counters.begin(), counters.end(), 0);
+}
+
+void VideoStreamEncoder::AdaptCounter::MoveCount(std::vector<int>* counters,
+ int from_reason) {
+ int to_reason = (from_reason + 1) % kScaleReasonSize;
+ ++((*counters)[to_reason]);
+ --((*counters)[from_reason]);
+}
+
+std::string VideoStreamEncoder::AdaptCounter::ToString(
+ const std::vector<int>& counters) const {
+ std::stringstream ss;
+ for (size_t reason = 0; reason < kScaleReasonSize; ++reason) {
+ ss << (reason ? " cpu" : "quality") << ":" << counters[reason];
+ }
+ return ss.str();
+}
+
+} // namespace webrtc