blob: e25eac8a186e7f41e4a63ec69d91eae6710ede48 (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
|
/*
* Copyright (c) 2021 The WebRTC project authors. All Rights Reserved.
*
* Use of this source code is governed by a BSD-style license
* that can be found in the LICENSE file in the root of the source
* tree. An additional intellectual property rights grant can be found
* in the file PATENTS. All contributing project authors may
* be found in the AUTHORS file in the root of the source tree.
*/
#include "modules/video_coding/frame_helpers.h"
#include <utility>
#include "rtc_base/logging.h"
namespace webrtc {
namespace {
constexpr TimeDelta kMaxVideoDelay = TimeDelta::Millis(10000);
}
bool FrameHasBadRenderTiming(Timestamp render_time, Timestamp now) {
// Zero render time means render immediately.
if (render_time.IsZero()) {
return false;
}
if (render_time < Timestamp::Zero()) {
return true;
}
TimeDelta frame_delay = render_time - now;
if (frame_delay.Abs() > kMaxVideoDelay) {
RTC_LOG(LS_WARNING) << "Frame has bad render timing because it is out of "
"the delay bounds (frame_delay_ms="
<< frame_delay.ms()
<< ", kMaxVideoDelay_ms=" << kMaxVideoDelay.ms() << ")";
return true;
}
return false;
}
bool TargetVideoDelayIsTooLarge(TimeDelta target_video_delay) {
if (target_video_delay > kMaxVideoDelay) {
RTC_LOG(LS_WARNING)
<< "Target video delay is too large. (target_video_delay_ms="
<< target_video_delay.ms()
<< ", kMaxVideoDelay_ms=" << kMaxVideoDelay.ms() << ")";
return true;
}
return false;
}
std::unique_ptr<EncodedFrame> CombineAndDeleteFrames(
absl::InlinedVector<std::unique_ptr<EncodedFrame>, 4> frames) {
RTC_DCHECK(!frames.empty());
if (frames.size() == 1) {
return std::move(frames[0]);
}
size_t total_length = 0;
for (const auto& frame : frames) {
total_length += frame->size();
}
const EncodedFrame& last_frame = *frames.back();
std::unique_ptr<EncodedFrame> first_frame = std::move(frames[0]);
auto encoded_image_buffer = EncodedImageBuffer::Create(total_length);
uint8_t* buffer = encoded_image_buffer->data();
first_frame->SetSpatialLayerFrameSize(first_frame->SpatialIndex().value_or(0),
first_frame->size());
memcpy(buffer, first_frame->data(), first_frame->size());
buffer += first_frame->size();
// Spatial index of combined frame is set equal to spatial index of its top
// spatial layer.
first_frame->SetSpatialIndex(last_frame.SpatialIndex().value_or(0));
first_frame->video_timing_mutable()->network2_timestamp_ms =
last_frame.video_timing().network2_timestamp_ms;
first_frame->video_timing_mutable()->receive_finish_ms =
last_frame.video_timing().receive_finish_ms;
// Append all remaining frames to the first one.
for (size_t i = 1; i < frames.size(); ++i) {
// Let |next_frame| fall out of scope so it is deleted after copying.
std::unique_ptr<EncodedFrame> next_frame = std::move(frames[i]);
first_frame->SetSpatialLayerFrameSize(
next_frame->SpatialIndex().value_or(0), next_frame->size());
memcpy(buffer, next_frame->data(), next_frame->size());
buffer += next_frame->size();
}
first_frame->SetEncodedData(encoded_image_buffer);
return first_frame;
}
} // namespace webrtc
|