summaryrefslogtreecommitdiffstats
path: root/dom/media/webrtc/MediaEngineRemoteVideoSource.cpp
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-07 09:22:09 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-07 09:22:09 +0000
commit43a97878ce14b72f0981164f87f2e35e14151312 (patch)
tree620249daf56c0258faa40cbdcf9cfba06de2a846 /dom/media/webrtc/MediaEngineRemoteVideoSource.cpp
parentInitial commit. (diff)
downloadfirefox-43a97878ce14b72f0981164f87f2e35e14151312.tar.xz
firefox-43a97878ce14b72f0981164f87f2e35e14151312.zip
Adding upstream version 110.0.1.upstream/110.0.1upstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'dom/media/webrtc/MediaEngineRemoteVideoSource.cpp')
-rw-r--r--dom/media/webrtc/MediaEngineRemoteVideoSource.cpp906
1 files changed, 906 insertions, 0 deletions
diff --git a/dom/media/webrtc/MediaEngineRemoteVideoSource.cpp b/dom/media/webrtc/MediaEngineRemoteVideoSource.cpp
new file mode 100644
index 0000000000..59c9b74a9a
--- /dev/null
+++ b/dom/media/webrtc/MediaEngineRemoteVideoSource.cpp
@@ -0,0 +1,906 @@
+/* -*- Mode: C++; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*-
+ * This Source Code Form is subject to the terms of the Mozilla Public
+ * License, v. 2.0. If a copy of the MPL was not distributed with this file,
+ * You can obtain one at http://mozilla.org/MPL/2.0/. */
+
+#include "MediaEngineRemoteVideoSource.h"
+
+#include "CamerasChild.h"
+#include "MediaManager.h"
+#include "MediaTrackConstraints.h"
+#include "mozilla/dom/MediaTrackSettingsBinding.h"
+#include "mozilla/ErrorNames.h"
+#include "mozilla/gfx/Point.h"
+#include "mozilla/RefPtr.h"
+#include "PerformanceRecorder.h"
+#include "Tracing.h"
+#include "VideoFrameUtils.h"
+#include "VideoUtils.h"
+#include "ImageContainer.h"
+#include "common_video/include/video_frame_buffer.h"
+#include "common_video/libyuv/include/webrtc_libyuv.h"
+
+namespace mozilla {
+
+extern LazyLogModule gMediaManagerLog;
+#define LOG(...) MOZ_LOG(gMediaManagerLog, LogLevel::Debug, (__VA_ARGS__))
+#define LOG_FRAME(...) \
+ MOZ_LOG(gMediaManagerLog, LogLevel::Verbose, (__VA_ARGS__))
+
+using dom::ConstrainLongRange;
+using dom::MediaSourceEnum;
+using dom::MediaTrackConstraints;
+using dom::MediaTrackConstraintSet;
+using dom::MediaTrackSettings;
+using dom::VideoFacingModeEnum;
+
+/* static */
+camera::CaptureEngine MediaEngineRemoteVideoSource::CaptureEngine(
+ MediaSourceEnum aMediaSource) {
+ switch (aMediaSource) {
+ case MediaSourceEnum::Browser:
+ return camera::BrowserEngine;
+ case MediaSourceEnum::Camera:
+ return camera::CameraEngine;
+ case MediaSourceEnum::Screen:
+ return camera::ScreenEngine;
+ case MediaSourceEnum::Window:
+ return camera::WinEngine;
+ default:
+ MOZ_CRASH();
+ }
+}
+
+static Maybe<VideoFacingModeEnum> GetFacingMode(const nsString& aDeviceName) {
+ // Set facing mode based on device name.
+#if defined(ANDROID)
+ // Names are generated. Example: "Camera 0, Facing back, Orientation 90"
+ //
+ // See media/webrtc/trunk/webrtc/modules/video_capture/android/java/src/org/
+ // webrtc/videoengine/VideoCaptureDeviceInfoAndroid.java
+
+ if (aDeviceName.Find(u"Facing back"_ns) != kNotFound) {
+ return Some(VideoFacingModeEnum::Environment);
+ }
+ if (aDeviceName.Find(u"Facing front"_ns) != kNotFound) {
+ return Some(VideoFacingModeEnum::User);
+ }
+#endif // ANDROID
+#ifdef XP_MACOSX
+ // Kludge to test user-facing cameras on OSX.
+ if (aDeviceName.Find(u"Face"_ns) != -1) {
+ return Some(VideoFacingModeEnum::User);
+ }
+#endif
+#ifdef XP_WIN
+ // The cameras' name of Surface book are "Microsoft Camera Front" and
+ // "Microsoft Camera Rear" respectively.
+
+ if (aDeviceName.Find(u"Front"_ns) != kNotFound) {
+ return Some(VideoFacingModeEnum::User);
+ }
+ if (aDeviceName.Find(u"Rear"_ns) != kNotFound) {
+ return Some(VideoFacingModeEnum::Environment);
+ }
+#endif // WINDOWS
+
+ return Nothing();
+}
+
+MediaEngineRemoteVideoSource::MediaEngineRemoteVideoSource(
+ const MediaDevice* aMediaDevice)
+ : mCapEngine(CaptureEngine(aMediaDevice->mMediaSource)),
+ mTrackingId(CaptureEngineToTrackingSourceStr(mCapEngine), 0),
+ mMutex("MediaEngineRemoteVideoSource::mMutex"),
+ mRescalingBufferPool(/* zero_initialize */ false,
+ /* max_number_of_buffers */ 1),
+ mSettingsUpdatedByFrame(MakeAndAddRef<media::Refcountable<AtomicBool>>()),
+ mSettings(MakeAndAddRef<media::Refcountable<MediaTrackSettings>>()),
+ mFirstFramePromise(mFirstFramePromiseHolder.Ensure(__func__)),
+ mMediaDevice(aMediaDevice),
+ mDeviceUUID(NS_ConvertUTF16toUTF8(aMediaDevice->mRawID)) {
+ LOG("%s", __PRETTY_FUNCTION__);
+ mSettings->mWidth.Construct(0);
+ mSettings->mHeight.Construct(0);
+ mSettings->mFrameRate.Construct(0);
+ if (mCapEngine == camera::CameraEngine) {
+ // Only cameras can have a facing mode.
+ Maybe<VideoFacingModeEnum> facingMode =
+ GetFacingMode(mMediaDevice->mRawName);
+ if (facingMode.isSome()) {
+ NS_ConvertASCIItoUTF16 facingString(
+ dom::VideoFacingModeEnumValues::GetString(*facingMode));
+ mSettings->mFacingMode.Construct(facingString);
+ mFacingMode.emplace(facingString);
+ }
+ }
+}
+
+MediaEngineRemoteVideoSource::~MediaEngineRemoteVideoSource() {
+ mFirstFramePromiseHolder.RejectIfExists(NS_ERROR_ABORT, __func__);
+}
+
+nsresult MediaEngineRemoteVideoSource::Allocate(
+ const MediaTrackConstraints& aConstraints, const MediaEnginePrefs& aPrefs,
+ uint64_t aWindowID, const char** aOutBadConstraint) {
+ LOG("%s", __PRETTY_FUNCTION__);
+ AssertIsOnOwningThread();
+
+ MOZ_ASSERT(mState == kReleased);
+
+ NormalizedConstraints constraints(aConstraints);
+ webrtc::CaptureCapability newCapability;
+ LOG("ChooseCapability(kFitness) for mCapability (Allocate) ++");
+ if (!ChooseCapability(constraints, aPrefs, newCapability, kFitness)) {
+ *aOutBadConstraint =
+ MediaConstraintsHelper::FindBadConstraint(constraints, mMediaDevice);
+ return NS_ERROR_FAILURE;
+ }
+ LOG("ChooseCapability(kFitness) for mCapability (Allocate) --");
+
+ mCaptureId =
+ camera::GetChildAndCall(&camera::CamerasChild::AllocateCapture,
+ mCapEngine, mDeviceUUID.get(), aWindowID);
+ if (mCaptureId < 0) {
+ return NS_ERROR_FAILURE;
+ }
+
+ {
+ MutexAutoLock lock(mMutex);
+ mState = kAllocated;
+ mCapability = newCapability;
+ mTrackingId =
+ TrackingId(CaptureEngineToTrackingSourceStr(mCapEngine), mCaptureId);
+ }
+
+ LOG("Video device %d allocated", mCaptureId);
+ return NS_OK;
+}
+
+nsresult MediaEngineRemoteVideoSource::Deallocate() {
+ LOG("%s", __PRETTY_FUNCTION__);
+ AssertIsOnOwningThread();
+
+ MOZ_ASSERT(mState == kStopped || mState == kAllocated);
+
+ if (mTrack) {
+ mTrack->End();
+ }
+
+ {
+ MutexAutoLock lock(mMutex);
+
+ mTrack = nullptr;
+ mPrincipal = PRINCIPAL_HANDLE_NONE;
+ mState = kReleased;
+ }
+
+ // Stop() has stopped capture synchronously on the media thread before we get
+ // here, so there are no longer any callbacks on an IPC thread accessing
+ // mImageContainer or mRescalingBufferPool.
+ mImageContainer = nullptr;
+ mRescalingBufferPool.Release();
+
+ LOG("Video device %d deallocated", mCaptureId);
+
+ if (camera::GetChildAndCall(&camera::CamerasChild::ReleaseCapture, mCapEngine,
+ mCaptureId)) {
+ MOZ_ASSERT_UNREACHABLE("Couldn't release allocated device");
+ }
+ return NS_OK;
+}
+
+void MediaEngineRemoteVideoSource::SetTrack(const RefPtr<MediaTrack>& aTrack,
+ const PrincipalHandle& aPrincipal) {
+ LOG("%s", __PRETTY_FUNCTION__);
+ AssertIsOnOwningThread();
+
+ MOZ_ASSERT(mState == kAllocated);
+ MOZ_ASSERT(!mTrack);
+ MOZ_ASSERT(aTrack);
+ MOZ_ASSERT(aTrack->AsSourceTrack());
+
+ if (!mImageContainer) {
+ mImageContainer = MakeAndAddRef<layers::ImageContainer>(
+ layers::ImageContainer::ASYNCHRONOUS);
+ }
+
+ {
+ MutexAutoLock lock(mMutex);
+ mTrack = aTrack->AsSourceTrack();
+ mPrincipal = aPrincipal;
+ }
+}
+
+nsresult MediaEngineRemoteVideoSource::Start() {
+ LOG("%s", __PRETTY_FUNCTION__);
+ AssertIsOnOwningThread();
+
+ MOZ_ASSERT(mState == kAllocated || mState == kStopped);
+ MOZ_ASSERT(mTrack);
+
+ {
+ MutexAutoLock lock(mMutex);
+ mState = kStarted;
+ }
+
+ mSettingsUpdatedByFrame->mValue = false;
+
+ if (camera::GetChildAndCall(&camera::CamerasChild::StartCapture, mCapEngine,
+ mCaptureId, mCapability, this)) {
+ LOG("StartCapture failed");
+ MutexAutoLock lock(mMutex);
+ mState = kStopped;
+ return NS_ERROR_FAILURE;
+ }
+
+ NS_DispatchToMainThread(NS_NewRunnableFunction(
+ "MediaEngineRemoteVideoSource::SetLastCapability",
+ [settings = mSettings, updated = mSettingsUpdatedByFrame,
+ capEngine = mCapEngine, cap = mCapability]() mutable {
+ switch (capEngine) {
+ case camera::ScreenEngine:
+ case camera::WinEngine:
+ // Undo the hack where ideal and max constraints are crammed
+ // together in mCapability for consumption by low-level code. We
+ // don't actually know the real resolution yet, so report min(ideal,
+ // max) for now.
+ // TODO: This can be removed in bug 1453269.
+ cap.width = std::min(cap.width >> 16, cap.width & 0xffff);
+ cap.height = std::min(cap.height >> 16, cap.height & 0xffff);
+ break;
+ default:
+ break;
+ }
+
+ if (!updated->mValue) {
+ settings->mWidth.Value() = cap.width;
+ settings->mHeight.Value() = cap.height;
+ }
+ settings->mFrameRate.Value() = cap.maxFPS;
+ }));
+
+ return NS_OK;
+}
+
+nsresult MediaEngineRemoteVideoSource::FocusOnSelectedSource() {
+ LOG("%s", __PRETTY_FUNCTION__);
+ AssertIsOnOwningThread();
+
+ int result;
+ result = camera::GetChildAndCall(&camera::CamerasChild::FocusOnSelectedSource,
+ mCapEngine, mCaptureId);
+ return result == 0 ? NS_OK : NS_ERROR_FAILURE;
+}
+
+nsresult MediaEngineRemoteVideoSource::Stop() {
+ LOG("%s", __PRETTY_FUNCTION__);
+ AssertIsOnOwningThread();
+
+ if (mState == kStopped || mState == kAllocated) {
+ return NS_OK;
+ }
+
+ MOZ_ASSERT(mState == kStarted);
+
+ if (camera::GetChildAndCall(&camera::CamerasChild::StopCapture, mCapEngine,
+ mCaptureId)) {
+ MOZ_DIAGNOSTIC_ASSERT(false, "Stopping a started capture failed");
+ return NS_ERROR_FAILURE;
+ }
+
+ {
+ MutexAutoLock lock(mMutex);
+ mState = kStopped;
+ }
+
+ return NS_OK;
+}
+
+nsresult MediaEngineRemoteVideoSource::Reconfigure(
+ const MediaTrackConstraints& aConstraints, const MediaEnginePrefs& aPrefs,
+ const char** aOutBadConstraint) {
+ LOG("%s", __PRETTY_FUNCTION__);
+ AssertIsOnOwningThread();
+
+ NormalizedConstraints constraints(aConstraints);
+ webrtc::CaptureCapability newCapability;
+ LOG("ChooseCapability(kFitness) for mTargetCapability (Reconfigure) ++");
+ if (!ChooseCapability(constraints, aPrefs, newCapability, kFitness)) {
+ *aOutBadConstraint =
+ MediaConstraintsHelper::FindBadConstraint(constraints, mMediaDevice);
+ return NS_ERROR_INVALID_ARG;
+ }
+ LOG("ChooseCapability(kFitness) for mTargetCapability (Reconfigure) --");
+
+ if (mCapability == newCapability) {
+ return NS_OK;
+ }
+
+ bool started = mState == kStarted;
+ if (started) {
+ nsresult rv = Stop();
+ if (NS_WARN_IF(NS_FAILED(rv))) {
+ nsAutoCString name;
+ GetErrorName(rv, name);
+ LOG("Video source %p for video device %d Reconfigure() failed "
+ "unexpectedly in Stop(). rv=%s",
+ this, mCaptureId, name.Data());
+ return NS_ERROR_UNEXPECTED;
+ }
+ }
+
+ {
+ MutexAutoLock lock(mMutex);
+ // Start() applies mCapability on the device.
+ mCapability = newCapability;
+ }
+
+ if (started) {
+ nsresult rv = Start();
+ if (NS_WARN_IF(NS_FAILED(rv))) {
+ nsAutoCString name;
+ GetErrorName(rv, name);
+ LOG("Video source %p for video device %d Reconfigure() failed "
+ "unexpectedly in Start(). rv=%s",
+ this, mCaptureId, name.Data());
+ return NS_ERROR_UNEXPECTED;
+ }
+ }
+
+ return NS_OK;
+}
+
+size_t MediaEngineRemoteVideoSource::NumCapabilities() const {
+ AssertIsOnOwningThread();
+
+ if (!mCapabilities.IsEmpty()) {
+ return mCapabilities.Length();
+ }
+
+ int num = camera::GetChildAndCall(&camera::CamerasChild::NumberOfCapabilities,
+ mCapEngine, mDeviceUUID.get());
+ if (num > 0) {
+ mCapabilities.SetLength(num);
+ } else {
+ // The default for devices that don't return discrete capabilities: treat
+ // them as supporting all capabilities orthogonally. E.g. screensharing.
+ // CaptureCapability defaults key values to 0, which means accept any value.
+ mCapabilities.AppendElement(MakeUnique<webrtc::CaptureCapability>());
+ mCapabilitiesAreHardcoded = true;
+ }
+
+ return mCapabilities.Length();
+}
+
+webrtc::CaptureCapability& MediaEngineRemoteVideoSource::GetCapability(
+ size_t aIndex) const {
+ AssertIsOnOwningThread();
+ MOZ_RELEASE_ASSERT(aIndex < mCapabilities.Length());
+ if (!mCapabilities[aIndex]) {
+ mCapabilities[aIndex] = MakeUnique<webrtc::CaptureCapability>();
+ camera::GetChildAndCall(&camera::CamerasChild::GetCaptureCapability,
+ mCapEngine, mDeviceUUID.get(), aIndex,
+ mCapabilities[aIndex].get());
+ }
+ return *mCapabilities[aIndex];
+}
+
+const TrackingId& MediaEngineRemoteVideoSource::GetTrackingId() const {
+ AssertIsOnOwningThread();
+ MOZ_ASSERT(mState != kReleased);
+ return mTrackingId;
+}
+
+int MediaEngineRemoteVideoSource::DeliverFrame(
+ uint8_t* aBuffer, const camera::VideoFrameProperties& aProps) {
+ // Cameras IPC thread - take great care with accessing members!
+
+ Maybe<int32_t> req_max_width;
+ Maybe<int32_t> req_max_height;
+ Maybe<int32_t> req_ideal_width;
+ Maybe<int32_t> req_ideal_height;
+ {
+ MutexAutoLock lock(mMutex);
+ MOZ_ASSERT(mState == kStarted);
+ // TODO: These can be removed in bug 1453269.
+ const int32_t max_width = mCapability.width & 0xffff;
+ const int32_t max_height = mCapability.height & 0xffff;
+ const int32_t ideal_width = (mCapability.width >> 16) & 0xffff;
+ const int32_t ideal_height = (mCapability.height >> 16) & 0xffff;
+
+ req_max_width = max_width ? Some(max_width) : Nothing();
+ req_max_height = max_height ? Some(max_height) : Nothing();
+ req_ideal_width = ideal_width ? Some(ideal_width) : Nothing();
+ req_ideal_height = ideal_height ? Some(ideal_height) : Nothing();
+ if (!mFrameDeliveringTrackingId) {
+ mFrameDeliveringTrackingId = Some(mTrackingId);
+ }
+ }
+
+ // This is only used in the case of screen sharing, see bug 1453269.
+
+ if (aProps.rotation() == 90 || aProps.rotation() == 270) {
+ // This frame is rotated, so what was negotiated as width is now height,
+ // and vice versa.
+ std::swap(req_max_width, req_max_height);
+ std::swap(req_ideal_width, req_ideal_height);
+ }
+
+ int32_t dst_max_width =
+ std::min(aProps.width(), req_max_width.valueOr(aProps.width()));
+ int32_t dst_max_height =
+ std::min(aProps.height(), req_max_height.valueOr(aProps.height()));
+ // This logic works for both camera and screen sharing case.
+ // for camera case, req_ideal_width and req_ideal_height are absent.
+ int32_t dst_width = req_ideal_width.valueOr(aProps.width());
+ int32_t dst_height = req_ideal_height.valueOr(aProps.height());
+
+ if (!req_ideal_width && req_ideal_height) {
+ dst_width = *req_ideal_height * aProps.width() / aProps.height();
+ } else if (!req_ideal_height && req_ideal_width) {
+ dst_height = *req_ideal_width * aProps.height() / aProps.width();
+ }
+ dst_width = std::min(dst_width, dst_max_width);
+ dst_height = std::min(dst_height, dst_max_height);
+
+ // Apply scaling for screen sharing, see bug 1453269.
+ switch (mCapEngine) {
+ case camera::ScreenEngine:
+ case camera::WinEngine: {
+ // scale to average of portrait and landscape
+ float scale_width = (float)dst_width / (float)aProps.width();
+ float scale_height = (float)dst_height / (float)aProps.height();
+ float scale = (scale_width + scale_height) / 2;
+ // If both req_ideal_width & req_ideal_height are absent, scale is 1, but
+ // if one is present and the other not, scale precisely to the one present
+ if (!req_ideal_width) {
+ scale = scale_height;
+ } else if (!req_ideal_height) {
+ scale = scale_width;
+ }
+ dst_width = int32_t(scale * (float)aProps.width());
+ dst_height = int32_t(scale * (float)aProps.height());
+
+ // if scaled rectangle exceeds max rectangle, scale to minimum of portrait
+ // and landscape
+ if (dst_width > dst_max_width || dst_height > dst_max_height) {
+ scale_width = (float)dst_max_width / (float)dst_width;
+ scale_height = (float)dst_max_height / (float)dst_height;
+ scale = std::min(scale_width, scale_height);
+ dst_width = int32_t(scale * dst_width);
+ dst_height = int32_t(scale * dst_height);
+ }
+ break;
+ }
+ default: {
+ break;
+ }
+ }
+
+ // Ensure width and height are at least two. Smaller frames can lead to
+ // problems with scaling and video encoding.
+ dst_width = std::max(2, dst_width);
+ dst_height = std::max(2, dst_height);
+
+ std::function<void()> callback_unused = []() {};
+ rtc::scoped_refptr<webrtc::I420BufferInterface> buffer =
+ webrtc::WrapI420Buffer(
+ aProps.width(), aProps.height(), aBuffer, aProps.yStride(),
+ aBuffer + aProps.yAllocatedSize(), aProps.uStride(),
+ aBuffer + aProps.yAllocatedSize() + aProps.uAllocatedSize(),
+ aProps.vStride(), callback_unused);
+
+ if ((dst_width != aProps.width() || dst_height != aProps.height()) &&
+ dst_width <= aProps.width() && dst_height <= aProps.height()) {
+ PerformanceRecorder<CopyVideoStage> rec("MERVS::CropAndScale"_ns,
+ *mFrameDeliveringTrackingId,
+ dst_width, dst_height);
+ // Destination resolution is smaller than source buffer. We'll rescale.
+ rtc::scoped_refptr<webrtc::I420Buffer> scaledBuffer =
+ mRescalingBufferPool.CreateI420Buffer(dst_width, dst_height);
+ if (!scaledBuffer) {
+ MOZ_ASSERT_UNREACHABLE(
+ "We might fail to allocate a buffer, but with this "
+ "being a recycling pool that shouldn't happen");
+ return 0;
+ }
+ scaledBuffer->CropAndScaleFrom(*buffer);
+ buffer = scaledBuffer;
+ rec.Record();
+ }
+
+ layers::PlanarYCbCrData data;
+ data.mYChannel = const_cast<uint8_t*>(buffer->DataY());
+ data.mYStride = buffer->StrideY();
+ MOZ_ASSERT(buffer->StrideU() == buffer->StrideV());
+ data.mCbCrStride = buffer->StrideU();
+ data.mCbChannel = const_cast<uint8_t*>(buffer->DataU());
+ data.mCrChannel = const_cast<uint8_t*>(buffer->DataV());
+ data.mPictureRect = gfx::IntRect(0, 0, buffer->width(), buffer->height());
+ data.mYUVColorSpace = gfx::YUVColorSpace::BT601;
+ data.mChromaSubsampling = gfx::ChromaSubsampling::HALF_WIDTH_AND_HEIGHT;
+
+ RefPtr<layers::PlanarYCbCrImage> image;
+ {
+ PerformanceRecorder<CopyVideoStage> rec(
+ "MERVS::Copy"_ns, *mFrameDeliveringTrackingId, dst_width, dst_height);
+ image = mImageContainer->CreatePlanarYCbCrImage();
+ if (!image->CopyData(data)) {
+ MOZ_ASSERT_UNREACHABLE(
+ "We might fail to allocate a buffer, but with this "
+ "being a recycling container that shouldn't happen");
+ return 0;
+ }
+ rec.Record();
+ }
+
+#ifdef DEBUG
+ static uint32_t frame_num = 0;
+ LOG_FRAME(
+ "frame %d (%dx%d)->(%dx%d); rotation %d, timeStamp %u, ntpTimeMs %" PRIu64
+ ", renderTimeMs %" PRIu64,
+ frame_num++, aProps.width(), aProps.height(), dst_width, dst_height,
+ aProps.rotation(), aProps.timeStamp(), aProps.ntpTimeMs(),
+ aProps.renderTimeMs());
+#endif
+
+ if (mImageSize.width != dst_width || mImageSize.height != dst_height) {
+ NS_DispatchToMainThread(NS_NewRunnableFunction(
+ "MediaEngineRemoteVideoSource::FrameSizeChange",
+ [settings = mSettings, updated = mSettingsUpdatedByFrame,
+ holder = std::move(mFirstFramePromiseHolder), dst_width,
+ dst_height]() mutable {
+ settings->mWidth.Value() = dst_width;
+ settings->mHeight.Value() = dst_height;
+ updated->mValue = true;
+ // Since mImageSize was initialized to (0,0), we end up here on the
+ // arrival of the first frame. We resolve the promise representing
+ // arrival of first frame, after correct settings values have been
+ // made available (Resolve() is idempotent if already resolved).
+ holder.ResolveIfExists(true, __func__);
+ }));
+ }
+
+ {
+ MutexAutoLock lock(mMutex);
+ MOZ_ASSERT(mState == kStarted);
+ VideoSegment segment;
+ mImageSize = image->GetSize();
+ segment.AppendFrame(image.forget(), mImageSize, mPrincipal);
+ mTrack->AppendData(&segment);
+ }
+
+ return 0;
+}
+
+uint32_t MediaEngineRemoteVideoSource::GetDistance(
+ const webrtc::CaptureCapability& aCandidate,
+ const NormalizedConstraintSet& aConstraints,
+ const DistanceCalculation aCalculate) const {
+ if (aCalculate == kFeasibility) {
+ return GetFeasibilityDistance(aCandidate, aConstraints);
+ }
+ return GetFitnessDistance(aCandidate, aConstraints);
+}
+
+uint32_t MediaEngineRemoteVideoSource::GetFitnessDistance(
+ const webrtc::CaptureCapability& aCandidate,
+ const NormalizedConstraintSet& aConstraints) const {
+ AssertIsOnOwningThread();
+
+ // Treat width|height|frameRate == 0 on capability as "can do any".
+ // This allows for orthogonal capabilities that are not in discrete steps.
+
+ typedef MediaConstraintsHelper H;
+ uint64_t distance =
+ uint64_t(H::FitnessDistance(mFacingMode, aConstraints.mFacingMode)) +
+ uint64_t(aCandidate.width ? H::FitnessDistance(int32_t(aCandidate.width),
+ aConstraints.mWidth)
+ : 0) +
+ uint64_t(aCandidate.height
+ ? H::FitnessDistance(int32_t(aCandidate.height),
+ aConstraints.mHeight)
+ : 0) +
+ uint64_t(aCandidate.maxFPS ? H::FitnessDistance(double(aCandidate.maxFPS),
+ aConstraints.mFrameRate)
+ : 0);
+ return uint32_t(std::min(distance, uint64_t(UINT32_MAX)));
+}
+
+uint32_t MediaEngineRemoteVideoSource::GetFeasibilityDistance(
+ const webrtc::CaptureCapability& aCandidate,
+ const NormalizedConstraintSet& aConstraints) const {
+ AssertIsOnOwningThread();
+
+ // Treat width|height|frameRate == 0 on capability as "can do any".
+ // This allows for orthogonal capabilities that are not in discrete steps.
+
+ typedef MediaConstraintsHelper H;
+ uint64_t distance =
+ uint64_t(H::FitnessDistance(mFacingMode, aConstraints.mFacingMode)) +
+ uint64_t(aCandidate.width
+ ? H::FeasibilityDistance(int32_t(aCandidate.width),
+ aConstraints.mWidth)
+ : 0) +
+ uint64_t(aCandidate.height
+ ? H::FeasibilityDistance(int32_t(aCandidate.height),
+ aConstraints.mHeight)
+ : 0) +
+ uint64_t(aCandidate.maxFPS
+ ? H::FeasibilityDistance(double(aCandidate.maxFPS),
+ aConstraints.mFrameRate)
+ : 0);
+ return uint32_t(std::min(distance, uint64_t(UINT32_MAX)));
+}
+
+// Find best capability by removing inferiors. May leave >1 of equal distance
+
+/* static */
+void MediaEngineRemoteVideoSource::TrimLessFitCandidates(
+ nsTArray<CapabilityCandidate>& aSet) {
+ uint32_t best = UINT32_MAX;
+ for (auto& candidate : aSet) {
+ if (best > candidate.mDistance) {
+ best = candidate.mDistance;
+ }
+ }
+ aSet.RemoveElementsBy(
+ [best](const auto& set) { return set.mDistance > best; });
+ MOZ_ASSERT(aSet.Length());
+}
+
+uint32_t MediaEngineRemoteVideoSource::GetBestFitnessDistance(
+ const nsTArray<const NormalizedConstraintSet*>& aConstraintSets) const {
+ AssertIsOnOwningThread();
+
+ size_t num = NumCapabilities();
+ nsTArray<CapabilityCandidate> candidateSet;
+ for (size_t i = 0; i < num; i++) {
+ candidateSet.AppendElement(CapabilityCandidate(GetCapability(i)));
+ }
+
+ bool first = true;
+ for (const NormalizedConstraintSet* ns : aConstraintSets) {
+ for (size_t i = 0; i < candidateSet.Length();) {
+ auto& candidate = candidateSet[i];
+ uint32_t distance = GetFitnessDistance(candidate.mCapability, *ns);
+ if (distance == UINT32_MAX) {
+ candidateSet.RemoveElementAt(i);
+ } else {
+ ++i;
+ if (first) {
+ candidate.mDistance = distance;
+ }
+ }
+ }
+ first = false;
+ }
+ if (!candidateSet.Length()) {
+ return UINT32_MAX;
+ }
+ TrimLessFitCandidates(candidateSet);
+ return candidateSet[0].mDistance;
+}
+
+static const char* ConvertVideoTypeToCStr(webrtc::VideoType aType) {
+ switch (aType) {
+ case webrtc::VideoType::kI420:
+ return "I420";
+ case webrtc::VideoType::kIYUV:
+ case webrtc::VideoType::kYV12:
+ return "YV12";
+ case webrtc::VideoType::kRGB24:
+ return "24BG";
+ case webrtc::VideoType::kABGR:
+ return "ABGR";
+ case webrtc::VideoType::kARGB:
+ return "ARGB";
+ case webrtc::VideoType::kARGB4444:
+ return "R444";
+ case webrtc::VideoType::kRGB565:
+ return "RGBP";
+ case webrtc::VideoType::kARGB1555:
+ return "RGBO";
+ case webrtc::VideoType::kYUY2:
+ return "YUY2";
+ case webrtc::VideoType::kUYVY:
+ return "UYVY";
+ case webrtc::VideoType::kMJPEG:
+ return "MJPG";
+ case webrtc::VideoType::kNV21:
+ return "NV21";
+ case webrtc::VideoType::kNV12:
+ return "NV12";
+ case webrtc::VideoType::kBGRA:
+ return "BGRA";
+ case webrtc::VideoType::kUnknown:
+ default:
+ return "unknown";
+ }
+}
+
+static void LogCapability(const char* aHeader,
+ const webrtc::CaptureCapability& aCapability,
+ uint32_t aDistance) {
+ LOG("%s: %4u x %4u x %2u maxFps, %s. Distance = %" PRIu32, aHeader,
+ aCapability.width, aCapability.height, aCapability.maxFPS,
+ ConvertVideoTypeToCStr(aCapability.videoType), aDistance);
+}
+
+bool MediaEngineRemoteVideoSource::ChooseCapability(
+ const NormalizedConstraints& aConstraints, const MediaEnginePrefs& aPrefs,
+ webrtc::CaptureCapability& aCapability,
+ const DistanceCalculation aCalculate) {
+ LOG("%s", __PRETTY_FUNCTION__);
+ AssertIsOnOwningThread();
+
+ if (MOZ_LOG_TEST(gMediaManagerLog, LogLevel::Debug)) {
+ LOG("ChooseCapability: prefs: %dx%d @%dfps", aPrefs.GetWidth(),
+ aPrefs.GetHeight(), aPrefs.mFPS);
+ MediaConstraintsHelper::LogConstraints(aConstraints);
+ if (!aConstraints.mAdvanced.empty()) {
+ LOG("Advanced array[%zu]:", aConstraints.mAdvanced.size());
+ for (auto& advanced : aConstraints.mAdvanced) {
+ MediaConstraintsHelper::LogConstraints(advanced);
+ }
+ }
+ }
+
+ switch (mCapEngine) {
+ case camera::ScreenEngine:
+ case camera::WinEngine: {
+ FlattenedConstraints c(aConstraints);
+ // The actual resolution to constrain around is not easy to find ahead of
+ // time (and may in fact change over time), so as a hack, we push ideal
+ // and max constraints down to desktop_capture_impl.cc and finish the
+ // algorithm there.
+ // TODO: This can be removed in bug 1453269.
+ aCapability.width =
+ (std::min(0xffff, c.mWidth.mIdeal.valueOr(0)) & 0xffff) << 16 |
+ (std::min(0xffff, c.mWidth.mMax) & 0xffff);
+ aCapability.height =
+ (std::min(0xffff, c.mHeight.mIdeal.valueOr(0)) & 0xffff) << 16 |
+ (std::min(0xffff, c.mHeight.mMax) & 0xffff);
+ aCapability.maxFPS =
+ c.mFrameRate.Clamp(c.mFrameRate.mIdeal.valueOr(aPrefs.mFPS));
+ return true;
+ }
+ case camera::BrowserEngine: {
+ FlattenedConstraints c(aConstraints);
+ aCapability.maxFPS =
+ c.mFrameRate.Clamp(c.mFrameRate.mIdeal.valueOr(aPrefs.mFPS));
+ return true;
+ }
+ default:
+ break;
+ }
+
+ nsTArray<CapabilityCandidate> candidateSet;
+ size_t num = NumCapabilities();
+ for (size_t i = 0; i < num; i++) {
+ candidateSet.AppendElement(CapabilityCandidate(GetCapability(i)));
+ }
+
+ if (mCapabilitiesAreHardcoded && mCapEngine == camera::CameraEngine) {
+ // We have a hardcoded capability, which means this camera didn't report
+ // discrete capabilities. It might still allow a ranged capability, so we
+ // add a couple of default candidates based on prefs and constraints.
+ // The chosen candidate will be propagated to StartCapture() which will fail
+ // for an invalid candidate.
+ MOZ_DIAGNOSTIC_ASSERT(mCapabilities.Length() == 1);
+ MOZ_DIAGNOSTIC_ASSERT(candidateSet.Length() == 1);
+ candidateSet.Clear();
+
+ FlattenedConstraints c(aConstraints);
+ // Reuse the code across both the low-definition (`false`) pref and
+ // the high-definition (`true`) pref.
+ // If there are constraints we try to satisfy them but we default to prefs.
+ // Note that since constraints are from content and can literally be
+ // anything we put (rather generous) caps on them.
+ for (bool isHd : {false, true}) {
+ webrtc::CaptureCapability cap;
+ int32_t prefWidth = aPrefs.GetWidth(isHd);
+ int32_t prefHeight = aPrefs.GetHeight(isHd);
+
+ cap.width = c.mWidth.Get(prefWidth);
+ cap.width = std::max(0, std::min(cap.width, 7680));
+
+ cap.height = c.mHeight.Get(prefHeight);
+ cap.height = std::max(0, std::min(cap.height, 4320));
+
+ cap.maxFPS = c.mFrameRate.Get(aPrefs.mFPS);
+ cap.maxFPS = std::max(0, std::min(cap.maxFPS, 480));
+
+ if (cap.width != prefWidth) {
+ // Width was affected by constraints.
+ // We'll adjust the height too so the aspect ratio is retained.
+ cap.height = cap.width * prefHeight / prefWidth;
+ } else if (cap.height != prefHeight) {
+ // Height was affected by constraints but not width.
+ // We'll adjust the width too so the aspect ratio is retained.
+ cap.width = cap.height * prefWidth / prefHeight;
+ }
+
+ if (candidateSet.Contains(cap, CapabilityComparator())) {
+ continue;
+ }
+ LogCapability("Hardcoded capability", cap, 0);
+ candidateSet.AppendElement(cap);
+ }
+ }
+
+ // First, filter capabilities by required constraints (min, max, exact).
+
+ for (size_t i = 0; i < candidateSet.Length();) {
+ auto& candidate = candidateSet[i];
+ candidate.mDistance =
+ GetDistance(candidate.mCapability, aConstraints, aCalculate);
+ LogCapability("Capability", candidate.mCapability, candidate.mDistance);
+ if (candidate.mDistance == UINT32_MAX) {
+ candidateSet.RemoveElementAt(i);
+ } else {
+ ++i;
+ }
+ }
+
+ if (candidateSet.IsEmpty()) {
+ LOG("failed to find capability match from %zu choices",
+ candidateSet.Length());
+ return false;
+ }
+
+ // Filter further with all advanced constraints (that don't overconstrain).
+
+ for (const auto& cs : aConstraints.mAdvanced) {
+ nsTArray<CapabilityCandidate> rejects;
+ for (size_t i = 0; i < candidateSet.Length();) {
+ if (GetDistance(candidateSet[i].mCapability, cs, aCalculate) ==
+ UINT32_MAX) {
+ rejects.AppendElement(candidateSet[i]);
+ candidateSet.RemoveElementAt(i);
+ } else {
+ ++i;
+ }
+ }
+ if (!candidateSet.Length()) {
+ candidateSet.AppendElements(std::move(rejects));
+ }
+ }
+ MOZ_ASSERT(
+ candidateSet.Length(),
+ "advanced constraints filtering step can't reduce candidates to zero");
+
+ // Remaining algorithm is up to the UA.
+
+ TrimLessFitCandidates(candidateSet);
+
+ // Any remaining multiples all have the same distance. A common case of this
+ // occurs when no ideal is specified. Lean toward defaults.
+ uint32_t sameDistance = candidateSet[0].mDistance;
+ {
+ MediaTrackConstraintSet prefs;
+ prefs.mWidth.Construct().SetAsLong() = aPrefs.GetWidth();
+ prefs.mHeight.Construct().SetAsLong() = aPrefs.GetHeight();
+ prefs.mFrameRate.Construct().SetAsDouble() = aPrefs.mFPS;
+ NormalizedConstraintSet normPrefs(prefs, false);
+
+ for (auto& candidate : candidateSet) {
+ candidate.mDistance =
+ GetDistance(candidate.mCapability, normPrefs, aCalculate);
+ }
+ TrimLessFitCandidates(candidateSet);
+ }
+
+ aCapability = candidateSet[0].mCapability;
+
+ LogCapability("Chosen capability", aCapability, sameDistance);
+ return true;
+}
+
+void MediaEngineRemoteVideoSource::GetSettings(
+ MediaTrackSettings& aOutSettings) const {
+ aOutSettings = *mSettings;
+}
+
+} // namespace mozilla