diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-07 09:22:09 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-07 09:22:09 +0000 |
commit | 43a97878ce14b72f0981164f87f2e35e14151312 (patch) | |
tree | 620249daf56c0258faa40cbdcf9cfba06de2a846 /dom/media/MediaRecorder.cpp | |
parent | Initial commit. (diff) | |
download | firefox-upstream.tar.xz firefox-upstream.zip |
Adding upstream version 110.0.1.upstream/110.0.1upstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'dom/media/MediaRecorder.cpp')
-rw-r--r-- | dom/media/MediaRecorder.cpp | 1894 |
1 files changed, 1894 insertions, 0 deletions
diff --git a/dom/media/MediaRecorder.cpp b/dom/media/MediaRecorder.cpp new file mode 100644 index 0000000000..9f8c2ad425 --- /dev/null +++ b/dom/media/MediaRecorder.cpp @@ -0,0 +1,1894 @@ +/* -*- Mode: C++; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*- */ +/* vim:set ts=2 sw=2 sts=2 et cindent: */ +/* This Source Code Form is subject to the terms of the Mozilla Public + * License, v. 2.0. If a copy of the MPL was not distributed with this + * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ + +#include "MediaRecorder.h" + +#include "AudioNodeEngine.h" +#include "AudioNodeTrack.h" +#include "DOMMediaStream.h" +#include "MediaDecoder.h" +#include "MediaEncoder.h" +#include "MediaTrackGraphImpl.h" +#include "VideoUtils.h" +#include "mozilla/DOMEventTargetHelper.h" +#include "mozilla/dom/AudioStreamTrack.h" +#include "mozilla/dom/BlobEvent.h" +#include "mozilla/dom/EmptyBlobImpl.h" +#include "mozilla/dom/File.h" +#include "mozilla/dom/MediaRecorderErrorEvent.h" +#include "mozilla/dom/VideoStreamTrack.h" +#include "mozilla/media/MediaUtils.h" +#include "mozilla/MemoryReporting.h" +#include "mozilla/Preferences.h" +#include "mozilla/StaticPtr.h" +#include "mozilla/TaskQueue.h" +#include "nsContentTypeParser.h" +#include "nsContentUtils.h" +#include "nsDocShell.h" +#include "nsError.h" +#include "mozilla/dom/Document.h" +#include "nsIPrincipal.h" +#include "nsIScriptError.h" +#include "nsMimeTypes.h" +#include "nsProxyRelease.h" +#include "nsServiceManagerUtils.h" +#include "nsTArray.h" + +mozilla::LazyLogModule gMediaRecorderLog("MediaRecorder"); +#define LOG(type, msg) MOZ_LOG(gMediaRecorderLog, type, msg) + +constexpr int MIN_VIDEO_BITRATE_BPS = 10e3; // 10kbps +constexpr int DEFAULT_VIDEO_BITRATE_BPS = 2500e3; // 2.5Mbps +constexpr int MAX_VIDEO_BITRATE_BPS = 100e6; // 100Mbps + +constexpr int MIN_AUDIO_BITRATE_BPS = 500; // 500bps +constexpr int DEFAULT_AUDIO_BITRATE_BPS = 128e3; // 128kbps +constexpr int MAX_AUDIO_BITRATE_BPS = 512e3; // 512kbps + +namespace mozilla::dom { + +using namespace mozilla::media; + +/** + * MediaRecorderReporter measures memory being used by the Media Recorder. + * + * It is a singleton reporter and the single class object lives as long as at + * least one Recorder is registered. In MediaRecorder, the reporter is + * unregistered when it is destroyed. + */ +class MediaRecorderReporter final : public nsIMemoryReporter { + public: + static void AddMediaRecorder(MediaRecorder* aRecorder) { + if (!sUniqueInstance) { + sUniqueInstance = MakeAndAddRef<MediaRecorderReporter>(); + RegisterWeakAsyncMemoryReporter(sUniqueInstance); + } + sUniqueInstance->mRecorders.AppendElement(aRecorder); + } + + static void RemoveMediaRecorder(MediaRecorder* aRecorder) { + if (!sUniqueInstance) { + return; + } + + sUniqueInstance->mRecorders.RemoveElement(aRecorder); + if (sUniqueInstance->mRecorders.IsEmpty()) { + UnregisterWeakMemoryReporter(sUniqueInstance); + sUniqueInstance = nullptr; + } + } + + NS_DECL_THREADSAFE_ISUPPORTS + + MediaRecorderReporter() = default; + + NS_IMETHOD + CollectReports(nsIHandleReportCallback* aHandleReport, nsISupports* aData, + bool aAnonymize) override { + nsTArray<RefPtr<MediaRecorder::SizeOfPromise>> promises; + for (const RefPtr<MediaRecorder>& recorder : mRecorders) { + promises.AppendElement(recorder->SizeOfExcludingThis(MallocSizeOf)); + } + + nsCOMPtr<nsIHandleReportCallback> handleReport = aHandleReport; + nsCOMPtr<nsISupports> data = aData; + MediaRecorder::SizeOfPromise::All(GetCurrentSerialEventTarget(), promises) + ->Then( + GetCurrentSerialEventTarget(), __func__, + [handleReport, data](const nsTArray<size_t>& sizes) { + nsCOMPtr<nsIMemoryReporterManager> manager = + do_GetService("@mozilla.org/memory-reporter-manager;1"); + if (!manager) { + return; + } + + size_t sum = 0; + for (const size_t& size : sizes) { + sum += size; + } + + handleReport->Callback(""_ns, "explicit/media/recorder"_ns, + KIND_HEAP, UNITS_BYTES, sum, + "Memory used by media recorder."_ns, data); + + manager->EndReport(); + }, + [](size_t) { MOZ_CRASH("Unexpected reject"); }); + + return NS_OK; + } + + private: + MOZ_DEFINE_MALLOC_SIZE_OF(MallocSizeOf) + + virtual ~MediaRecorderReporter() { + MOZ_ASSERT(mRecorders.IsEmpty(), "All recorders must have been removed"); + } + + static StaticRefPtr<MediaRecorderReporter> sUniqueInstance; + + nsTArray<RefPtr<MediaRecorder>> mRecorders; +}; +NS_IMPL_ISUPPORTS(MediaRecorderReporter, nsIMemoryReporter); + +NS_IMPL_CYCLE_COLLECTION_CLASS(MediaRecorder) + +NS_IMPL_CYCLE_COLLECTION_TRAVERSE_BEGIN_INHERITED(MediaRecorder, + DOMEventTargetHelper) + NS_IMPL_CYCLE_COLLECTION_TRAVERSE(mStream) + NS_IMPL_CYCLE_COLLECTION_TRAVERSE(mAudioNode) + NS_IMPL_CYCLE_COLLECTION_TRAVERSE(mOtherDomException) + NS_IMPL_CYCLE_COLLECTION_TRAVERSE(mSecurityDomException) + NS_IMPL_CYCLE_COLLECTION_TRAVERSE(mUnknownDomException) + NS_IMPL_CYCLE_COLLECTION_TRAVERSE(mDocument) +NS_IMPL_CYCLE_COLLECTION_TRAVERSE_END + +NS_IMPL_CYCLE_COLLECTION_UNLINK_BEGIN_INHERITED(MediaRecorder, + DOMEventTargetHelper) + NS_IMPL_CYCLE_COLLECTION_UNLINK(mStream) + NS_IMPL_CYCLE_COLLECTION_UNLINK(mAudioNode) + NS_IMPL_CYCLE_COLLECTION_UNLINK(mOtherDomException) + NS_IMPL_CYCLE_COLLECTION_UNLINK(mSecurityDomException) + NS_IMPL_CYCLE_COLLECTION_UNLINK(mUnknownDomException) + tmp->UnRegisterActivityObserver(); + NS_IMPL_CYCLE_COLLECTION_UNLINK(mDocument) +NS_IMPL_CYCLE_COLLECTION_UNLINK_END + +NS_INTERFACE_MAP_BEGIN_CYCLE_COLLECTION(MediaRecorder) + NS_INTERFACE_MAP_ENTRY(nsIDocumentActivity) +NS_INTERFACE_MAP_END_INHERITING(DOMEventTargetHelper) + +NS_IMPL_ADDREF_INHERITED(MediaRecorder, DOMEventTargetHelper) +NS_IMPL_RELEASE_INHERITED(MediaRecorder, DOMEventTargetHelper) + +namespace { +bool PrincipalSubsumes(MediaRecorder* aRecorder, nsIPrincipal* aPrincipal) { + if (!aRecorder->GetOwner()) { + return false; + } + nsCOMPtr<Document> doc = aRecorder->GetOwner()->GetExtantDoc(); + if (!doc) { + return false; + } + if (!aPrincipal) { + return false; + } + bool subsumes; + if (NS_FAILED(doc->NodePrincipal()->Subsumes(aPrincipal, &subsumes))) { + return false; + } + return subsumes; +} + +bool MediaStreamTracksPrincipalSubsumes( + MediaRecorder* aRecorder, + const nsTArray<RefPtr<MediaStreamTrack>>& aTracks) { + nsCOMPtr<nsIPrincipal> principal = nullptr; + for (const auto& track : aTracks) { + nsContentUtils::CombineResourcePrincipals(&principal, + track->GetPrincipal()); + } + return PrincipalSubsumes(aRecorder, principal); +} + +bool AudioNodePrincipalSubsumes(MediaRecorder* aRecorder, + AudioNode* aAudioNode) { + MOZ_ASSERT(aAudioNode); + Document* doc = + aAudioNode->GetOwner() ? aAudioNode->GetOwner()->GetExtantDoc() : nullptr; + nsCOMPtr<nsIPrincipal> principal = doc ? doc->NodePrincipal() : nullptr; + return PrincipalSubsumes(aRecorder, principal); +} + +// This list is sorted so that lesser failures are later, so that +// IsTypeSupportedImpl() can report the error from audio or video types that +// is closer to being supported. +enum class TypeSupport { + MediaTypeInvalid, + NoVideoWithAudioType, + ContainersDisabled, + CodecsDisabled, + ContainerUnsupported, + CodecUnsupported, + CodecDuplicated, + Supported, +}; + +nsCString TypeSupportToCString(TypeSupport aSupport, + const nsAString& aMimeType) { + nsAutoCString mime = NS_ConvertUTF16toUTF8(aMimeType); + switch (aSupport) { + case TypeSupport::Supported: + return nsPrintfCString("%s is supported", mime.get()); + case TypeSupport::MediaTypeInvalid: + return nsPrintfCString("%s is not a valid media type", mime.get()); + case TypeSupport::NoVideoWithAudioType: + return nsPrintfCString( + "Video cannot be recorded with %s as it is an audio type", + mime.get()); + case TypeSupport::ContainersDisabled: + return "All containers are disabled"_ns; + case TypeSupport::CodecsDisabled: + return "All codecs are disabled"_ns; + case TypeSupport::ContainerUnsupported: + return nsPrintfCString("%s indicates an unsupported container", + mime.get()); + case TypeSupport::CodecUnsupported: + return nsPrintfCString("%s indicates an unsupported codec", mime.get()); + case TypeSupport::CodecDuplicated: + return nsPrintfCString("%s contains the same codec multiple times", + mime.get()); + default: + MOZ_ASSERT_UNREACHABLE("Unknown TypeSupport"); + return "Unknown error"_ns; + } +} + +TypeSupport CanRecordAudioTrackWith(const Maybe<MediaContainerType>& aMimeType, + const nsAString& aMimeTypeString) { + if (aMimeTypeString.IsEmpty()) { + // For the empty string we just need to check whether we have support for an + // audio container and an audio codec. + if (!MediaEncoder::IsWebMEncoderEnabled() && + !MediaDecoder::IsOggEnabled()) { + // No container support for audio. + return TypeSupport::ContainersDisabled; + } + + if (!MediaDecoder::IsOpusEnabled()) { + // No codec support for audio. + return TypeSupport::CodecsDisabled; + } + + return TypeSupport::Supported; + } + + if (!aMimeType) { + // A mime type string was set, but it couldn't be parsed to a valid + // MediaContainerType. + return TypeSupport::MediaTypeInvalid; + } + + if (aMimeType->Type() != MEDIAMIMETYPE(VIDEO_WEBM) && + aMimeType->Type() != MEDIAMIMETYPE(AUDIO_WEBM) && + aMimeType->Type() != MEDIAMIMETYPE(AUDIO_OGG)) { + // Any currently supported container can record audio. + return TypeSupport::ContainerUnsupported; + } + + if (aMimeType->Type() == MEDIAMIMETYPE(VIDEO_WEBM) && + !MediaEncoder::IsWebMEncoderEnabled()) { + return TypeSupport::ContainerUnsupported; + } + + if (aMimeType->Type() == MEDIAMIMETYPE(AUDIO_WEBM) && + !MediaEncoder::IsWebMEncoderEnabled()) { + return TypeSupport::ContainerUnsupported; + } + + if (aMimeType->Type() == MEDIAMIMETYPE(AUDIO_OGG) && + !MediaDecoder::IsOggEnabled()) { + return TypeSupport::ContainerUnsupported; + } + + if (!MediaDecoder::IsOpusEnabled()) { + return TypeSupport::CodecUnsupported; + } + + if (!aMimeType->ExtendedType().HaveCodecs()) { + // No codecs constrained, we can pick opus. + return TypeSupport::Supported; + } + + size_t opus = 0; + size_t unknown = 0; + for (const auto& codec : aMimeType->ExtendedType().Codecs().Range()) { + // Ignore video codecs. + if (codec.EqualsLiteral("vp8")) { + continue; + } + if (codec.EqualsLiteral("vp8.0")) { + continue; + } + if (codec.EqualsLiteral("opus")) { + // All containers support opus + opus++; + continue; + } + unknown++; + } + + if (unknown > 0) { + // Unsupported codec. + return TypeSupport::CodecUnsupported; + } + + if (opus == 0) { + // Codecs specified but not opus. Unsupported for audio. + return TypeSupport::CodecUnsupported; + } + + if (opus > 1) { + // Opus specified more than once. Bad form. + return TypeSupport::CodecDuplicated; + } + + return TypeSupport::Supported; +} + +TypeSupport CanRecordVideoTrackWith(const Maybe<MediaContainerType>& aMimeType, + const nsAString& aMimeTypeString) { + if (aMimeTypeString.IsEmpty()) { + // For the empty string we just need to check whether we have support for a + // video container and a video codec. The VP8 encoder is always available. + if (!MediaEncoder::IsWebMEncoderEnabled()) { + // No container support for video. + return TypeSupport::ContainersDisabled; + } + + return TypeSupport::Supported; + } + + if (!aMimeType) { + // A mime type string was set, but it couldn't be parsed to a valid + // MediaContainerType. + return TypeSupport::MediaTypeInvalid; + } + + if (!aMimeType->Type().HasVideoMajorType()) { + return TypeSupport::NoVideoWithAudioType; + } + + if (aMimeType->Type() != MEDIAMIMETYPE(VIDEO_WEBM)) { + return TypeSupport::ContainerUnsupported; + } + + if (!MediaEncoder::IsWebMEncoderEnabled()) { + return TypeSupport::ContainerUnsupported; + } + + if (!aMimeType->ExtendedType().HaveCodecs()) { + // No codecs constrained, we can pick vp8. + return TypeSupport::Supported; + } + + size_t vp8 = 0; + size_t unknown = 0; + for (const auto& codec : aMimeType->ExtendedType().Codecs().Range()) { + if (codec.EqualsLiteral("opus")) { + // Ignore audio codecs. + continue; + } + if (codec.EqualsLiteral("vp8")) { + vp8++; + continue; + } + if (codec.EqualsLiteral("vp8.0")) { + vp8++; + continue; + } + unknown++; + } + + if (unknown > 0) { + // Unsupported codec. + return TypeSupport::CodecUnsupported; + } + + if (vp8 == 0) { + // Codecs specified but not vp8. Unsupported for video. + return TypeSupport::CodecUnsupported; + } + + if (vp8 > 1) { + // Vp8 specified more than once. Bad form. + return TypeSupport::CodecDuplicated; + } + + return TypeSupport::Supported; +} + +TypeSupport CanRecordWith(MediaStreamTrack* aTrack, + const Maybe<MediaContainerType>& aMimeType, + const nsAString& aMimeTypeString) { + if (aTrack->AsAudioStreamTrack()) { + return CanRecordAudioTrackWith(aMimeType, aMimeTypeString); + } + + if (aTrack->AsVideoStreamTrack()) { + return CanRecordVideoTrackWith(aMimeType, aMimeTypeString); + } + + MOZ_CRASH("Unexpected track type"); +} + +TypeSupport IsTypeSupportedImpl(const nsAString& aMIMEType) { + if (aMIMEType.IsEmpty()) { + // Lie and return true even if no container/codec support is enabled, + // because the spec mandates it. + return TypeSupport::Supported; + } + Maybe<MediaContainerType> mime = MakeMediaContainerType(aMIMEType); + TypeSupport audioSupport = CanRecordAudioTrackWith(mime, aMIMEType); + TypeSupport videoSupport = CanRecordVideoTrackWith(mime, aMIMEType); + return std::max(audioSupport, videoSupport); +} + +nsString SelectMimeType(bool aHasVideo, bool aHasAudio, + const nsString& aConstrainedMimeType) { + MOZ_ASSERT(aHasVideo || aHasAudio); + + Maybe<MediaContainerType> constrainedType = + MakeMediaContainerType(aConstrainedMimeType); + + // If we are recording video, Start() should have rejected any non-video mime + // types. + MOZ_ASSERT_IF(constrainedType && aHasVideo, + constrainedType->Type().HasVideoMajorType()); + // IsTypeSupported() rejects application mime types. + MOZ_ASSERT_IF(constrainedType, + !constrainedType->Type().HasApplicationMajorType()); + + nsString result; + if (constrainedType && constrainedType->ExtendedType().HaveCodecs()) { + // The constrained mime type is fully defined (it has codecs!). No need to + // select anything. + CopyUTF8toUTF16(constrainedType->OriginalString(), result); + } else { + // There is no constrained mime type, or there is and it is not fully + // defined but still valid. Select what's missing, so that we have major + // type, container and codecs. + + // If there is a constrained mime type it should not have codecs defined, + // because then it is fully defined and used unchanged (covered earlier). + MOZ_ASSERT_IF(constrainedType, + !constrainedType->ExtendedType().HaveCodecs()); + + nsCString majorType; + { + if (constrainedType) { + // There is a constrained type. It has both major type and container in + // order to be valid. Use them as is. + majorType = constrainedType->Type().AsString(); + } else if (aHasVideo) { + majorType = nsLiteralCString(VIDEO_WEBM); + } else { + majorType = nsLiteralCString(AUDIO_OGG); + } + } + + nsCString codecs; + { + if (aHasVideo && aHasAudio) { + codecs = "\"vp8, opus\""_ns; + } else if (aHasVideo) { + codecs = "vp8"_ns; + } else { + codecs = "opus"_ns; + } + } + result = NS_ConvertUTF8toUTF16( + nsPrintfCString("%s; codecs=%s", majorType.get(), codecs.get())); + } + + MOZ_ASSERT_IF(aHasAudio, + CanRecordAudioTrackWith(MakeMediaContainerType(result), + result) == TypeSupport::Supported); + MOZ_ASSERT_IF(aHasVideo, + CanRecordVideoTrackWith(MakeMediaContainerType(result), + result) == TypeSupport::Supported); + return result; +} + +void SelectBitrates(uint32_t aBitsPerSecond, uint8_t aNumVideoTracks, + uint32_t* aOutVideoBps, uint8_t aNumAudioTracks, + uint32_t* aOutAudioBps) { + uint32_t vbps = 0; + uint32_t abps = 0; + + const uint32_t minVideoBps = MIN_VIDEO_BITRATE_BPS * aNumVideoTracks; + const uint32_t maxVideoBps = MAX_VIDEO_BITRATE_BPS * aNumVideoTracks; + + const uint32_t minAudioBps = MIN_AUDIO_BITRATE_BPS * aNumAudioTracks; + const uint32_t maxAudioBps = MAX_AUDIO_BITRATE_BPS * aNumAudioTracks; + + if (aNumVideoTracks == 0) { + MOZ_DIAGNOSTIC_ASSERT(aNumAudioTracks > 0); + abps = std::min(maxAudioBps, std::max(minAudioBps, aBitsPerSecond)); + } else if (aNumAudioTracks == 0) { + vbps = std::min(maxVideoBps, std::max(minVideoBps, aBitsPerSecond)); + } else { + // Scale the bits so that video gets 20 times the bits of audio. + // Since we must account for varying number of tracks of each type we weight + // them by type; video = weight 20, audio = weight 1. + const uint32_t videoWeight = aNumVideoTracks * 20; + const uint32_t audioWeight = aNumAudioTracks; + const uint32_t totalWeights = audioWeight + videoWeight; + const uint32_t videoBitrate = + uint64_t(aBitsPerSecond) * videoWeight / totalWeights; + const uint32_t audioBitrate = + uint64_t(aBitsPerSecond) * audioWeight / totalWeights; + vbps = std::min(maxVideoBps, std::max(minVideoBps, videoBitrate)); + abps = std::min(maxAudioBps, std::max(minAudioBps, audioBitrate)); + } + + *aOutVideoBps = vbps; + *aOutAudioBps = abps; +} +} // namespace + +/** + * Session is an object to represent a single recording event. + * In original design, all recording context is stored in MediaRecorder, which + * causes a problem if someone calls MediaRecorder::Stop and + * MediaRecorder::Start quickly. To prevent blocking main thread, media encoding + * is executed in a second thread, named encoder thread. For the same reason, we + * do not await encoder thread shutdown in MediaRecorder::Stop. + * If someone calls MediaRecorder::Start before encoder thread shutdown, the + * same recording context in MediaRecorder might be accessed by two distinct + * encoder threads, which would be racy. With the recording context, including + * the encoder thread, in a Session object the problem is solved. + * + * Lifetime of MediaRecorder and Session objects. + * 1) MediaRecorder creates a Session in MediaRecorder::Start() and holds + * a reference to it. Then the Session registers itself to a ShutdownBlocker + * and also holds a reference to MediaRecorder. + * Therefore, the reference dependency in gecko is: + * ShutdownBlocker -> Session <-> MediaRecorder, note that there is a cycle + * reference between Session and MediaRecorder. + * 2) A Session is destroyed after Session::DoSessionEndTask() has been called + * _and_ all encoded media data has been passed to OnDataAvailable handler. + * In some cases the encoded media can be discarded before being passed to + * the OnDataAvailable handler. + * 3) Session::DoSessionEndTask is called by an application through + * MediaRecorder::Stop(), from a MediaEncoder Shutdown notification, from the + * document going inactive or invisible, or from the ShutdownBlocker. + */ +class MediaRecorder::Session : public PrincipalChangeObserver<MediaStreamTrack>, + public DOMMediaStream::TrackListener { + NS_INLINE_DECL_THREADSAFE_REFCOUNTING(Session) + + struct TrackTypeComparator { + enum Type { + AUDIO, + VIDEO, + }; + static bool Equals(const RefPtr<MediaStreamTrack>& aTrack, Type aType) { + return (aType == AUDIO && aTrack->AsAudioStreamTrack()) || + (aType == VIDEO && aTrack->AsVideoStreamTrack()); + } + }; + + public: + Session(MediaRecorder* aRecorder, + nsTArray<RefPtr<MediaStreamTrack>> aMediaStreamTracks, + uint32_t aVideoBitsPerSecond, uint32_t aAudioBitsPerSecond) + : mRecorder(aRecorder), + mMediaStreamTracks(std::move(aMediaStreamTracks)), + mMainThread(mRecorder->GetOwner()->EventTargetFor(TaskCategory::Other)), + mMimeType(SelectMimeType( + mMediaStreamTracks.Contains(TrackTypeComparator::VIDEO, + TrackTypeComparator()), + mRecorder->mAudioNode || + mMediaStreamTracks.Contains(TrackTypeComparator::AUDIO, + TrackTypeComparator()), + mRecorder->mConstrainedMimeType)), + mVideoBitsPerSecond(aVideoBitsPerSecond), + mAudioBitsPerSecond(aAudioBitsPerSecond), + mStartTime(TimeStamp::Now()), + mRunningState(RunningState::Idling) { + MOZ_ASSERT(NS_IsMainThread()); + Telemetry::ScalarAdd(Telemetry::ScalarID::MEDIARECORDER_RECORDING_COUNT, 1); + } + + void PrincipalChanged(MediaStreamTrack* aTrack) override { + NS_ASSERTION(mMediaStreamTracks.Contains(aTrack), + "Principal changed for unrecorded track"); + if (!MediaStreamTracksPrincipalSubsumes(mRecorder, mMediaStreamTracks)) { + DoSessionEndTask(NS_ERROR_DOM_SECURITY_ERR); + } + } + + void NotifyTrackAdded(const RefPtr<MediaStreamTrack>& aTrack) override { + LOG(LogLevel::Warning, + ("Session.NotifyTrackAdded %p Raising error due to track set change", + this)); + // There's a chance we have a sensible JS stack here. + if (!mRecorder->mOtherDomException) { + mRecorder->mOtherDomException = DOMException::Create( + NS_ERROR_DOM_INVALID_MODIFICATION_ERR, + "An attempt was made to add a track to the recorded MediaStream " + "during the recording"_ns); + } + DoSessionEndTask(NS_ERROR_DOM_INVALID_MODIFICATION_ERR); + } + + void NotifyTrackRemoved(const RefPtr<MediaStreamTrack>& aTrack) override { + if (aTrack->Ended()) { + // TrackEncoder will pickup tracks that end itself. + return; + } + LOG(LogLevel::Warning, + ("Session.NotifyTrackRemoved %p Raising error due to track set change", + this)); + // There's a chance we have a sensible JS stack here. + if (!mRecorder->mOtherDomException) { + mRecorder->mOtherDomException = DOMException::Create( + NS_ERROR_DOM_INVALID_MODIFICATION_ERR, + "An attempt was made to remove a track from the recorded MediaStream " + "during the recording"_ns); + } + DoSessionEndTask(NS_ERROR_DOM_INVALID_MODIFICATION_ERR); + } + + void Start(TimeDuration aTimeslice) { + LOG(LogLevel::Debug, ("Session.Start %p", this)); + MOZ_ASSERT(NS_IsMainThread()); + + if (mRecorder->mStream) { + // The TrackListener reports back when tracks are added or removed from + // the MediaStream. + mMediaStream = mRecorder->mStream; + mMediaStream->RegisterTrackListener(this); + + uint8_t trackTypes = 0; + for (const auto& track : mMediaStreamTracks) { + if (track->AsAudioStreamTrack()) { + trackTypes |= ContainerWriter::CREATE_AUDIO_TRACK; + } else if (track->AsVideoStreamTrack()) { + trackTypes |= ContainerWriter::CREATE_VIDEO_TRACK; + } else { + MOZ_CRASH("Unexpected track type"); + } + } + + for (const auto& t : mMediaStreamTracks) { + t->AddPrincipalChangeObserver(this); + } + + LOG(LogLevel::Debug, ("Session.Start track types = (%d)", trackTypes)); + InitEncoder(trackTypes, mMediaStreamTracks[0]->Graph()->GraphRate(), + aTimeslice); + return; + } + + if (mRecorder->mAudioNode) { + TrackRate trackRate = + mRecorder->mAudioNode->Context()->Graph()->GraphRate(); + + // Web Audio node has only audio. + InitEncoder(ContainerWriter::CREATE_AUDIO_TRACK, trackRate, aTimeslice); + return; + } + + MOZ_ASSERT(false, "Unknown source"); + } + + void Stop() { + LOG(LogLevel::Debug, ("Session.Stop %p", this)); + MOZ_ASSERT(NS_IsMainThread()); + + if (mEncoder) { + mEncoder->DisconnectTracks(); + } + + // Remove main thread state added in Start(). + if (mMediaStream) { + mMediaStream->UnregisterTrackListener(this); + mMediaStream = nullptr; + } + + { + for (const auto& track : mMediaStreamTracks) { + track->RemovePrincipalChangeObserver(this); + } + } + + if (mRunningState.isOk() && + mRunningState.inspect() == RunningState::Idling) { + LOG(LogLevel::Debug, ("Session.Stop Explicit end task %p", this)); + // End the Session directly if there is no encoder. + DoSessionEndTask(NS_OK); + } else if (mRunningState.isOk() && + (mRunningState.inspect() == RunningState::Starting || + mRunningState.inspect() == RunningState::Running)) { + if (mRunningState.inspect() == RunningState::Starting) { + // The MediaEncoder might not report started, but by spec we must fire + // "start". + mStartedListener.DisconnectIfExists(); + NS_DispatchToMainThread(NewRunnableMethod( + "MediaRecorder::Session::Stop", this, &Session::OnStarted)); + } + mRunningState = RunningState::Stopping; + } + } + + void Pause() { + LOG(LogLevel::Debug, ("Session.Pause")); + MOZ_ASSERT(NS_IsMainThread()); + MOZ_ASSERT_IF(mRunningState.isOk(), + mRunningState.unwrap() != RunningState::Idling); + if (mRunningState.isErr() || + mRunningState.unwrap() == RunningState::Stopping || + mRunningState.unwrap() == RunningState::Stopped) { + return; + } + MOZ_ASSERT(mEncoder); + mEncoder->Suspend(); + } + + void Resume() { + LOG(LogLevel::Debug, ("Session.Resume")); + MOZ_ASSERT(NS_IsMainThread()); + MOZ_ASSERT_IF(mRunningState.isOk(), + mRunningState.unwrap() != RunningState::Idling); + if (mRunningState.isErr() || + mRunningState.unwrap() == RunningState::Stopping || + mRunningState.unwrap() == RunningState::Stopped) { + return; + } + MOZ_ASSERT(mEncoder); + mEncoder->Resume(); + } + + void RequestData() { + LOG(LogLevel::Debug, ("Session.RequestData")); + MOZ_ASSERT(NS_IsMainThread()); + MOZ_ASSERT(mEncoder); + + InvokeAsync(mEncoderThread, mEncoder.get(), __func__, + &MediaEncoder::RequestData) + ->Then( + mMainThread, __func__, + [this, self = RefPtr<Session>(this)]( + const MediaEncoder::BlobPromise::ResolveOrRejectValue& aRrv) { + if (aRrv.IsReject()) { + LOG(LogLevel::Warning, ("RequestData failed")); + DoSessionEndTask(aRrv.RejectValue()); + return; + } + + nsresult rv = + mRecorder->CreateAndDispatchBlobEvent(aRrv.ResolveValue()); + if (NS_FAILED(rv)) { + DoSessionEndTask(NS_OK); + } + }); + } + + public: + RefPtr<SizeOfPromise> SizeOfExcludingThis( + mozilla::MallocSizeOf aMallocSizeOf) { + MOZ_ASSERT(NS_IsMainThread()); + if (!mEncoder) { + return SizeOfPromise::CreateAndResolve(0, __func__); + } + + return mEncoder->SizeOfExcludingThis(aMallocSizeOf); + } + + private: + virtual ~Session() { + MOZ_ASSERT(NS_IsMainThread()); + MOZ_ASSERT(mShutdownPromise); + MOZ_ASSERT(!mShutdownBlocker); + LOG(LogLevel::Debug, ("Session.~Session (%p)", this)); + } + + void InitEncoder(uint8_t aTrackTypes, TrackRate aTrackRate, + TimeDuration aTimeslice) { + LOG(LogLevel::Debug, ("Session.InitEncoder %p", this)); + MOZ_ASSERT(NS_IsMainThread()); + + if (!mRunningState.isOk() || + mRunningState.inspect() != RunningState::Idling) { + MOZ_ASSERT_UNREACHABLE("Double-init"); + return; + } + + // Create a TaskQueue to read encode media data from MediaEncoder. + MOZ_RELEASE_ASSERT(!mEncoderThread); + RefPtr<SharedThreadPool> pool = + GetMediaThreadPool(MediaThreadType::WEBRTC_WORKER); + if (!pool) { + LOG(LogLevel::Debug, ("Session.InitEncoder %p Failed to create " + "MediaRecorderReadThread thread pool", + this)); + DoSessionEndTask(NS_ERROR_FAILURE); + return; + } + + mEncoderThread = + TaskQueue::Create(pool.forget(), "MediaRecorderReadThread"); + + MOZ_DIAGNOSTIC_ASSERT(!mShutdownBlocker); + // Add a shutdown blocker so mEncoderThread can be shutdown async. + class Blocker : public ShutdownBlocker { + const RefPtr<Session> mSession; + + public: + Blocker(RefPtr<Session> aSession, const nsString& aName) + : ShutdownBlocker(aName), mSession(std::move(aSession)) {} + + NS_IMETHOD BlockShutdown(nsIAsyncShutdownClient*) override { + mSession->DoSessionEndTask(NS_ERROR_ABORT); + return NS_OK; + } + }; + + nsCOMPtr<nsIAsyncShutdownClient> barrier = GetShutdownBarrier(); + if (!barrier) { + LOG(LogLevel::Error, + ("Session.InitEncoder %p Failed to get shutdown barrier", this)); + DoSessionEndTask(NS_ERROR_FAILURE); + return; + } + + nsString name; + name.AppendPrintf("MediaRecorder::Session %p shutdown", this); + mShutdownBlocker = MakeAndAddRef<Blocker>(this, name); + nsresult rv = barrier->AddBlocker( + mShutdownBlocker, NS_LITERAL_STRING_FROM_CSTRING(__FILE__), __LINE__, + u"MediaRecorder::Session: shutdown"_ns); + MOZ_RELEASE_ASSERT(NS_SUCCEEDED(rv)); + + uint32_t maxMemory = Preferences::GetUint("media.recorder.max_memory", + MAX_ALLOW_MEMORY_BUFFER); + + mEncoder = MediaEncoder::CreateEncoder( + mEncoderThread, mMimeType, mAudioBitsPerSecond, mVideoBitsPerSecond, + aTrackTypes, aTrackRate, maxMemory, aTimeslice); + + if (!mEncoder) { + LOG(LogLevel::Error, ("Session.InitEncoder !mEncoder %p", this)); + DoSessionEndTask(NS_ERROR_ABORT); + return; + } + + mStartedListener = mEncoder->StartedEvent().Connect(mMainThread, this, + &Session::OnStarted); + mDataAvailableListener = mEncoder->DataAvailableEvent().Connect( + mMainThread, this, &Session::OnDataAvailable); + mErrorListener = + mEncoder->ErrorEvent().Connect(mMainThread, this, &Session::OnError); + mShutdownListener = mEncoder->ShutdownEvent().Connect(mMainThread, this, + &Session::OnShutdown); + + if (mRecorder->mAudioNode) { + mEncoder->ConnectAudioNode(mRecorder->mAudioNode, + mRecorder->mAudioNodeOutput); + } + + for (const auto& track : mMediaStreamTracks) { + mEncoder->ConnectMediaStreamTrack(track); + } + + // Set mRunningState to Running so that DoSessionEndTask will + // take the responsibility to end the session. + mRunningState = RunningState::Starting; + } + + // This is the task that will stop recording per spec: + // - If rv is NS_ERROR_ABORT or NS_ERROR_DOM_SECURITY_ERR, cancel the encoders + // - Otherwise, stop the encoders gracefully, this still encodes buffered data + // - Set state to "inactive" + // - Fire an error event, if NS_FAILED(rv) + // - Discard blob data if rv is NS_ERROR_DOM_SECURITY_ERR + // - Fire a Blob event + // - Fire an event named stop + void DoSessionEndTask(nsresult rv) { + MOZ_ASSERT(NS_IsMainThread()); + if (mRunningState.isErr()) { + // We have already ended with an error. + return; + } + + if (mRunningState.isOk() && + mRunningState.inspect() == RunningState::Stopped) { + // We have already ended gracefully. + return; + } + + bool needsStartEvent = false; + if (mRunningState.isOk() && + (mRunningState.inspect() == RunningState::Idling || + mRunningState.inspect() == RunningState::Starting)) { + needsStartEvent = true; + } + + // Set a terminated running state. Future DoSessionEnd tasks will exit + // early. + if (rv == NS_OK) { + mRunningState = RunningState::Stopped; + } else { + mRunningState = Err(rv); + } + + RefPtr<MediaEncoder::BlobPromise> blobPromise; + if (!mEncoder) { + blobPromise = MediaEncoder::BlobPromise::CreateAndReject(NS_OK, __func__); + } else { + blobPromise = + (rv == NS_ERROR_ABORT || rv == NS_ERROR_DOM_SECURITY_ERR + ? mEncoder->Cancel() + : mEncoder->Stop()) + ->Then(mEncoderThread, __func__, + [encoder = mEncoder]( + const GenericNonExclusivePromise::ResolveOrRejectValue& + aValue) { + MOZ_DIAGNOSTIC_ASSERT(aValue.IsResolve()); + return encoder->RequestData(); + }); + } + + blobPromise + ->Then( + mMainThread, __func__, + [this, self = RefPtr<Session>(this), rv, needsStartEvent]( + const MediaEncoder::BlobPromise::ResolveOrRejectValue& aRv) { + if (mRecorder->mSessions.LastElement() == this) { + // Set state to inactive, but only if the recorder is not + // controlled by another session already. + mRecorder->Inactivate(); + } + + if (needsStartEvent) { + mRecorder->DispatchSimpleEvent(u"start"_ns); + } + + // If there was an error, Fire the appropriate one + if (NS_FAILED(rv)) { + mRecorder->NotifyError(rv); + } + + // Fire a blob event named dataavailable + RefPtr<BlobImpl> blobImpl; + if (rv == NS_ERROR_DOM_SECURITY_ERR || aRv.IsReject()) { + // In case of SecurityError, the blob data must be discarded. + // We create a new empty one and throw the blob with its data + // away. + // In case we failed to gather blob data, we create an empty + // memory blob instead. + blobImpl = new EmptyBlobImpl(mMimeType); + } else { + blobImpl = aRv.ResolveValue(); + } + if (NS_FAILED(mRecorder->CreateAndDispatchBlobEvent(blobImpl))) { + // Failed to dispatch blob event. That's unexpected. It's + // probably all right to fire an error event if we haven't + // already. + if (NS_SUCCEEDED(rv)) { + mRecorder->NotifyError(NS_ERROR_FAILURE); + } + } + + // Fire an event named stop + mRecorder->DispatchSimpleEvent(u"stop"_ns); + + // And finally, Shutdown and destroy the Session + return Shutdown(); + }) + ->Then(mMainThread, __func__, [this, self = RefPtr<Session>(this)] { + // Guard against the case where we fail to add a blocker due to being + // in XPCOM shutdown. If we're in this state we shouldn't try and get + // a shutdown barrier as we'll fail. + if (!mShutdownBlocker) { + return; + } + MustGetShutdownBarrier()->RemoveBlocker(mShutdownBlocker); + mShutdownBlocker = nullptr; + }); + } + + void OnStarted() { + MOZ_ASSERT(NS_IsMainThread()); + if (mRunningState.isErr()) { + return; + } + RunningState state = mRunningState.inspect(); + if (state == RunningState::Starting || state == RunningState::Stopping) { + if (state == RunningState::Starting) { + // We set it to Running in the runnable since we can only assign + // mRunningState on main thread. We set it before running the start + // event runnable since that dispatches synchronously (and may cause + // js calls to methods depending on mRunningState). + mRunningState = RunningState::Running; + + mRecorder->mMimeType = mEncoder->mMimeType; + } + mRecorder->DispatchSimpleEvent(u"start"_ns); + } + } + + void OnDataAvailable(const RefPtr<BlobImpl>& aBlob) { + if (mRunningState.isErr() && + mRunningState.unwrapErr() == NS_ERROR_DOM_SECURITY_ERR) { + return; + } + if (NS_WARN_IF(NS_FAILED(mRecorder->CreateAndDispatchBlobEvent(aBlob)))) { + LOG(LogLevel::Warning, + ("MediaRecorder %p Creating or dispatching BlobEvent failed", this)); + DoSessionEndTask(NS_OK); + } + } + + void OnError() { + MOZ_ASSERT(NS_IsMainThread()); + DoSessionEndTask(NS_ERROR_FAILURE); + } + + void OnShutdown() { + MOZ_ASSERT(NS_IsMainThread()); + DoSessionEndTask(NS_OK); + } + + RefPtr<ShutdownPromise> Shutdown() { + MOZ_ASSERT(NS_IsMainThread()); + LOG(LogLevel::Debug, ("Session Shutdown %p", this)); + + if (mShutdownPromise) { + return mShutdownPromise; + } + + // This is a coarse calculation and does not reflect the duration of the + // final recording for reasons such as pauses. However it allows us an + // idea of how long people are running their recorders for. + TimeDuration timeDelta = TimeStamp::Now() - mStartTime; + Telemetry::Accumulate(Telemetry::MEDIA_RECORDER_RECORDING_DURATION, + timeDelta.ToSeconds()); + + mShutdownPromise = ShutdownPromise::CreateAndResolve(true, __func__); + + if (mEncoder) { + mShutdownPromise = + mShutdownPromise + ->Then(mMainThread, __func__, + [this, self = RefPtr<Session>(this)] { + mStartedListener.DisconnectIfExists(); + mDataAvailableListener.DisconnectIfExists(); + mErrorListener.DisconnectIfExists(); + mShutdownListener.DisconnectIfExists(); + return mEncoder->Cancel(); + }) + ->Then(mEncoderThread, __func__, [] { + // Meh, this is just to convert the promise type to match + // mShutdownPromise. + return ShutdownPromise::CreateAndResolve(true, __func__); + }); + } + + // Remove main thread state. This could be needed if Stop() wasn't called. + if (mMediaStream) { + mMediaStream->UnregisterTrackListener(this); + mMediaStream = nullptr; + } + + { + auto tracks(std::move(mMediaStreamTracks)); + for (RefPtr<MediaStreamTrack>& track : tracks) { + track->RemovePrincipalChangeObserver(this); + } + } + + // Break the cycle reference between Session and MediaRecorder. + mShutdownPromise = mShutdownPromise->Then( + mMainThread, __func__, + [self = RefPtr<Session>(this)]() { + self->mRecorder->RemoveSession(self); + return ShutdownPromise::CreateAndResolve(true, __func__); + }, + []() { + MOZ_ASSERT_UNREACHABLE("Unexpected reject"); + return ShutdownPromise::CreateAndReject(false, __func__); + }); + + if (mEncoderThread) { + mShutdownPromise = mShutdownPromise->Then( + mMainThread, __func__, + [encoderThread = mEncoderThread]() { + return encoderThread->BeginShutdown(); + }, + []() { + MOZ_ASSERT_UNREACHABLE("Unexpected reject"); + return ShutdownPromise::CreateAndReject(false, __func__); + }); + } + + return mShutdownPromise; + } + + private: + enum class RunningState { + Idling, // Session has been created + Starting, // MediaEncoder started, waiting for data + Running, // MediaEncoder has received data + Stopping, // Stop() has been called + Stopped, // Session has stopped without any error + }; + + // Our associated MediaRecorder. + const RefPtr<MediaRecorder> mRecorder; + + // Stream currently recorded. + RefPtr<DOMMediaStream> mMediaStream; + + // Tracks currently recorded. This should be a subset of mMediaStream's track + // set. + nsTArray<RefPtr<MediaStreamTrack>> mMediaStreamTracks; + + // Main thread used for MozPromise operations. + const RefPtr<nsISerialEventTarget> mMainThread; + // Runnable thread for reading data from MediaEncoder. + RefPtr<TaskQueue> mEncoderThread; + // MediaEncoder pipeline. + RefPtr<MediaEncoder> mEncoder; + // Listener connected to mMediaEncoder::StartedEvent(). + MediaEventListener mStartedListener; + // Listener connected to mMediaEncoder::DataAvailableEvent(). + MediaEventListener mDataAvailableListener; + // Listener connected to mMediaEncoder::ErrorEvent(). + MediaEventListener mErrorListener; + // Listener connected to mMediaEncoder::ShutdownEvent(). + MediaEventListener mShutdownListener; + // Set in Shutdown() and resolved when shutdown is complete. + RefPtr<ShutdownPromise> mShutdownPromise; + // Session mimeType + const nsString mMimeType; + // The video bitrate the recorder was configured with. + const uint32_t mVideoBitsPerSecond; + // The audio bitrate the recorder was configured with. + const uint32_t mAudioBitsPerSecond; + // The time this session started, for telemetry. + const TimeStamp mStartTime; + // The session's current main thread state. The error type gets set when + // ending a recording with an error. An NS_OK error is invalid. + // Main thread only. + Result<RunningState, nsresult> mRunningState; + // Shutdown blocker unique for this Session. Main thread only. + RefPtr<ShutdownBlocker> mShutdownBlocker; +}; + +MediaRecorder::~MediaRecorder() { + LOG(LogLevel::Debug, ("~MediaRecorder (%p)", this)); + UnRegisterActivityObserver(); +} + +MediaRecorder::MediaRecorder(nsPIDOMWindowInner* aOwnerWindow) + : DOMEventTargetHelper(aOwnerWindow) { + MOZ_ASSERT(aOwnerWindow); + RegisterActivityObserver(); +} + +void MediaRecorder::RegisterActivityObserver() { + if (nsPIDOMWindowInner* window = GetOwner()) { + mDocument = window->GetExtantDoc(); + if (mDocument) { + mDocument->RegisterActivityObserver( + NS_ISUPPORTS_CAST(nsIDocumentActivity*, this)); + } + } +} + +void MediaRecorder::UnRegisterActivityObserver() { + if (mDocument) { + mDocument->UnregisterActivityObserver( + NS_ISUPPORTS_CAST(nsIDocumentActivity*, this)); + } +} + +void MediaRecorder::GetMimeType(nsString& aMimeType) { aMimeType = mMimeType; } + +void MediaRecorder::Start(const Optional<uint32_t>& aTimeslice, + ErrorResult& aResult) { + LOG(LogLevel::Debug, ("MediaRecorder.Start %p", this)); + + InitializeDomExceptions(); + + // When a MediaRecorder object’s start() method is invoked, the UA MUST run + // the following steps: + + // 1. Let recorder be the MediaRecorder object on which the method was + // invoked. + + // 2. Let timeslice be the method’s first argument, if provided, or undefined. + TimeDuration timeslice = + aTimeslice.WasPassed() + ? TimeDuration::FromMilliseconds(aTimeslice.Value()) + : TimeDuration::Forever(); + + // 3. Let stream be the value of recorder’s stream attribute. + + // 4. Let tracks be the set of live tracks in stream’s track set. + nsTArray<RefPtr<MediaStreamTrack>> tracks; + if (mStream) { + mStream->GetTracks(tracks); + } + tracks.RemoveLastElements( + tracks.end() - std::remove_if(tracks.begin(), tracks.end(), + [](const auto& t) { return t->Ended(); })); + + // 5. If the value of recorder’s state attribute is not inactive, throw an + // InvalidStateError DOMException and abort these steps. + if (mState != RecordingState::Inactive) { + aResult.ThrowInvalidStateError( + "The MediaRecorder has already been started"); + return; + } + + // 6. If the isolation properties of stream disallow access from recorder, + // throw a SecurityError DOMException and abort these steps. + if (mStream) { + RefPtr<nsIPrincipal> streamPrincipal = mStream->GetPrincipal(); + if (!streamPrincipal) { + // This is more or less part of the step 7, see below. + aResult.ThrowNotSupportedError("The MediaStream is inactive"); + return; + } + + if (!PrincipalSubsumes(this, streamPrincipal)) { + aResult.ThrowSecurityError( + "The MediaStream's isolation properties disallow access from " + "MediaRecorder"); + return; + } + } + if (mAudioNode && !AudioNodePrincipalSubsumes(this, mAudioNode)) { + LOG(LogLevel::Warning, + ("MediaRecorder %p Start AudioNode principal check failed", this)); + aResult.ThrowSecurityError( + "The AudioNode's isolation properties disallow access from " + "MediaRecorder"); + return; + } + + // 7. If stream is inactive, throw a NotSupportedError DOMException and abort + // these steps. + if (mStream && !mStream->Active()) { + aResult.ThrowNotSupportedError("The MediaStream is inactive"); + return; + } + + // 8. If the [[ConstrainedMimeType]] slot specifies a media type, container, + // or codec, then run the following sub steps: + // 1. Constrain the configuration of recorder to the media type, container, + // and codec specified in the [[ConstrainedMimeType]] slot. + // 2. For each track in tracks, if the User Agent cannot record the track + // using the current configuration, then throw a NotSupportedError + // DOMException and abort all steps. + Maybe<MediaContainerType> mime; + if (mConstrainedMimeType.Length() > 0) { + mime = MakeMediaContainerType(mConstrainedMimeType); + MOZ_DIAGNOSTIC_ASSERT( + mime, + "Invalid media MIME type should have been caught by IsTypeSupported"); + } + for (const auto& track : tracks) { + TypeSupport support = CanRecordWith(track, mime, mConstrainedMimeType); + if (support != TypeSupport::Supported) { + nsString id; + track->GetId(id); + aResult.ThrowNotSupportedError(nsPrintfCString( + "%s track cannot be recorded: %s", + track->AsAudioStreamTrack() ? "An audio" : "A video", + TypeSupportToCString(support, mConstrainedMimeType).get())); + return; + } + } + if (mAudioNode) { + TypeSupport support = CanRecordAudioTrackWith(mime, mConstrainedMimeType); + if (support != TypeSupport::Supported) { + aResult.ThrowNotSupportedError(nsPrintfCString( + "An AudioNode cannot be recorded: %s", + TypeSupportToCString(support, mConstrainedMimeType).get())); + return; + } + } + + // 9. If recorder’s [[ConstrainedBitsPerSecond]] slot is not undefined, set + // recorder’s videoBitsPerSecond and audioBitsPerSecond attributes to + // values the User Agent deems reasonable for the respective media types, + // for recording all tracks in tracks, such that the sum of + // videoBitsPerSecond and audioBitsPerSecond is close to the value of + // recorder’s + // [[ConstrainedBitsPerSecond]] slot. + uint8_t numVideoTracks = 0; + uint8_t numAudioTracks = 0; + for (const auto& t : tracks) { + if (t->AsVideoStreamTrack() && numVideoTracks < UINT8_MAX) { + ++numVideoTracks; + } else if (t->AsAudioStreamTrack() && numAudioTracks < UINT8_MAX) { + ++numAudioTracks; + } + } + if (mAudioNode) { + MOZ_DIAGNOSTIC_ASSERT(!mStream); + ++numAudioTracks; + } + if (mConstrainedBitsPerSecond) { + SelectBitrates(*mConstrainedBitsPerSecond, numVideoTracks, + &mVideoBitsPerSecond, numAudioTracks, &mAudioBitsPerSecond); + } + + // 10. Let videoBitrate be the value of recorder’s videoBitsPerSecond + // attribute, and constrain the configuration of recorder to target an + // aggregate bitrate of videoBitrate bits per second for all video tracks + // recorder will be recording. videoBitrate is a hint for the encoder and + // the value might be surpassed, not achieved, or only be achieved over a + // long period of time. + const uint32_t videoBitrate = mVideoBitsPerSecond; + + // 11. Let audioBitrate be the value of recorder’s audioBitsPerSecond + // attribute, and constrain the configuration of recorder to target an + // aggregate bitrate of audioBitrate bits per second for all audio tracks + // recorder will be recording. audioBitrate is a hint for the encoder and + // the value might be surpassed, not achieved, or only be achieved over a + // long period of time. + const uint32_t audioBitrate = mAudioBitsPerSecond; + + // 12. Constrain the configuration of recorder to encode using the BitrateMode + // specified by the value of recorder’s audioBitrateMode attribute for all + // audio tracks recorder will be recording. + // -- NOT IMPLEMENTED + + // 13. For each track in tracks, if the User Agent cannot record the track + // using the current configuration, then throw a NotSupportedError + // DOMException and abort these steps. + if (numVideoTracks > 1) { + aResult.ThrowNotSupportedError( + "MediaRecorder does not support recording more than one video track"_ns); + return; + } + if (numAudioTracks > 1) { + aResult.ThrowNotSupportedError( + "MediaRecorder does not support recording more than one audio track"_ns); + return; + } + + // 14. Set recorder’s state to recording + mState = RecordingState::Recording; + + MediaRecorderReporter::AddMediaRecorder(this); + // Start a session. + mSessions.AppendElement(); + mSessions.LastElement() = + new Session(this, std::move(tracks), videoBitrate, audioBitrate); + mSessions.LastElement()->Start(timeslice); +} + +void MediaRecorder::Stop(ErrorResult& aResult) { + LOG(LogLevel::Debug, ("MediaRecorder.Stop %p", this)); + MediaRecorderReporter::RemoveMediaRecorder(this); + + // When a MediaRecorder object’s stop() method is invoked, the UA MUST run the + // following steps: + + // 1. Let recorder be the MediaRecorder object on which the method was + // invoked. + + // 2. If recorder’s state attribute is inactive, abort these steps. + if (mState == RecordingState::Inactive) { + return; + } + + // 3. Inactivate the recorder with recorder. + Inactivate(); + + // 4. Queue a task, using the DOM manipulation task source, that runs the + // following steps: + // 1. Stop gathering data. + // 2. Let blob be the Blob of collected data so far, then fire a blob event + // named dataavailable at recorder with blob. + // 3. Fire an event named stop at recorder. + MOZ_ASSERT(mSessions.Length() > 0); + mSessions.LastElement()->Stop(); + + // 5. return undefined. +} + +void MediaRecorder::Pause(ErrorResult& aResult) { + LOG(LogLevel::Debug, ("MediaRecorder.Pause %p", this)); + + // When a MediaRecorder object’s pause() method is invoked, the UA MUST run + // the following steps: + + // 1. If state is inactive, throw an InvalidStateError DOMException and abort + // these steps. + if (mState == RecordingState::Inactive) { + aResult.ThrowInvalidStateError("The MediaRecorder is inactive"); + return; + } + + // 2. If state is paused, abort these steps. + if (mState == RecordingState::Paused) { + return; + } + + // 3. Set state to paused, and queue a task, using the DOM manipulation task + // source, that runs the following steps: + mState = RecordingState::Paused; + + // XXX - We pause synchronously pending spec issue + // https://github.com/w3c/mediacapture-record/issues/131 + // 1. Stop gathering data into blob (but keep it available so that + // recording can be resumed in the future). + MOZ_ASSERT(!mSessions.IsEmpty()); + mSessions.LastElement()->Pause(); + + NS_DispatchToMainThread(NS_NewRunnableFunction( + "MediaRecorder::Pause", [recorder = RefPtr<MediaRecorder>(this)] { + // 2. Let target be the MediaRecorder context object. Fire an event + // named pause at target. + recorder->DispatchSimpleEvent(u"pause"_ns); + })); + + // 4. return undefined. +} + +void MediaRecorder::Resume(ErrorResult& aResult) { + LOG(LogLevel::Debug, ("MediaRecorder.Resume %p", this)); + + // When a MediaRecorder object’s resume() method is invoked, the UA MUST run + // the following steps: + + // 1. If state is inactive, throw an InvalidStateError DOMException and abort + // these steps. + if (mState == RecordingState::Inactive) { + aResult.ThrowInvalidStateError("The MediaRecorder is inactive"); + return; + } + + // 2. If state is recording, abort these steps. + if (mState == RecordingState::Recording) { + return; + } + + // 3. Set state to recording, and queue a task, using the DOM manipulation + // task source, that runs the following steps: + mState = RecordingState::Recording; + + // XXX - We resume synchronously pending spec issue + // https://github.com/w3c/mediacapture-record/issues/131 + // 1. Resume (or continue) gathering data into the current blob. + MOZ_ASSERT(!mSessions.IsEmpty()); + mSessions.LastElement()->Resume(); + + NS_DispatchToMainThread(NS_NewRunnableFunction( + "MediaRecorder::Resume", [recorder = RefPtr<MediaRecorder>(this)] { + // 2. Let target be the MediaRecorder context object. Fire an event + // named resume at target. + recorder->DispatchSimpleEvent(u"resume"_ns); + })); + + // 4. return undefined. +} + +void MediaRecorder::RequestData(ErrorResult& aResult) { + LOG(LogLevel::Debug, ("MediaRecorder.RequestData %p", this)); + + // When a MediaRecorder object’s requestData() method is invoked, the UA MUST + // run the following steps: + + // 1. If state is inactive throw an InvalidStateError DOMException and + // terminate these steps. Otherwise the UA MUST queue a task, using the DOM + // manipulation task source, that runs the following steps: + // 1. Let blob be the Blob of collected data so far and let target be the + // MediaRecorder context object, then fire a blob event named + // dataavailable at target with blob. (Note that blob will be empty if no + // data has been gathered yet.) + // 2. Create a new Blob and gather subsequent data into it. + if (mState == RecordingState::Inactive) { + aResult.ThrowInvalidStateError("The MediaRecorder is inactive"); + return; + } + MOZ_ASSERT(mSessions.Length() > 0); + mSessions.LastElement()->RequestData(); + + // 2. return undefined. +} + +JSObject* MediaRecorder::WrapObject(JSContext* aCx, + JS::Handle<JSObject*> aGivenProto) { + return MediaRecorder_Binding::Wrap(aCx, this, aGivenProto); +} + +/* static */ +already_AddRefed<MediaRecorder> MediaRecorder::Constructor( + const GlobalObject& aGlobal, DOMMediaStream& aStream, + const MediaRecorderOptions& aOptions, ErrorResult& aRv) { + nsCOMPtr<nsPIDOMWindowInner> ownerWindow = + do_QueryInterface(aGlobal.GetAsSupports()); + if (!ownerWindow) { + aRv.Throw(NS_ERROR_FAILURE); + return nullptr; + } + + // When the MediaRecorder() constructor is invoked, the User Agent MUST run + // the following steps: + + // 1. Let stream be the constructor’s first argument. + + // 2. Let options be the constructor’s second argument. + + // 3. If invoking is type supported with options’ mimeType member as its + // argument returns false, throw a NotSupportedError DOMException and abort + // these steps. + TypeSupport support = IsTypeSupportedImpl(aOptions.mMimeType); + if (support != TypeSupport::Supported) { + // This catches also the empty string mimeType when support for any encoders + // has been disabled. + aRv.ThrowNotSupportedError( + TypeSupportToCString(support, aOptions.mMimeType)); + return nullptr; + } + + // 4. Let recorder be a newly constructed MediaRecorder object. + RefPtr<MediaRecorder> recorder = new MediaRecorder(ownerWindow); + + // 5. Let recorder have a [[ConstrainedMimeType]] internal slot, initialized + // to the value of options' mimeType member. + recorder->mConstrainedMimeType = aOptions.mMimeType; + + // 6. Let recorder have a [[ConstrainedBitsPerSecond]] internal slot, + // initialized to the value of options’ bitsPerSecond member, if it is + // present, otherwise undefined. + recorder->mConstrainedBitsPerSecond = + aOptions.mBitsPerSecond.WasPassed() + ? Some(aOptions.mBitsPerSecond.Value()) + : Nothing(); + + // 7. Initialize recorder’s stream attribute to stream. + recorder->mStream = &aStream; + + // 8. Initialize recorder’s mimeType attribute to the value of recorder’s + // [[ConstrainedMimeType]] slot. + recorder->mMimeType = recorder->mConstrainedMimeType; + + // 9. Initialize recorder’s state attribute to inactive. + recorder->mState = RecordingState::Inactive; + + // 10. Initialize recorder’s videoBitsPerSecond attribute to the value of + // options’ videoBitsPerSecond member, if it is present. Otherwise, choose + // a target value the User Agent deems reasonable for video. + recorder->mVideoBitsPerSecond = aOptions.mVideoBitsPerSecond.WasPassed() + ? aOptions.mVideoBitsPerSecond.Value() + : DEFAULT_VIDEO_BITRATE_BPS; + + // 11. Initialize recorder’s audioBitsPerSecond attribute to the value of + // options’ audioBitsPerSecond member, if it is present. Otherwise, choose + // a target value the User Agent deems reasonable for audio. + recorder->mAudioBitsPerSecond = aOptions.mAudioBitsPerSecond.WasPassed() + ? aOptions.mAudioBitsPerSecond.Value() + : DEFAULT_AUDIO_BITRATE_BPS; + + // 12. If recorder’s [[ConstrainedBitsPerSecond]] slot is not undefined, set + // recorder’s videoBitsPerSecond and audioBitsPerSecond attributes to + // values the User Agent deems reasonable for the respective media types, + // such that the sum of videoBitsPerSecond and audioBitsPerSecond is close + // to the value of recorder’s [[ConstrainedBitsPerSecond]] slot. + if (recorder->mConstrainedBitsPerSecond) { + SelectBitrates(*recorder->mConstrainedBitsPerSecond, 1, + &recorder->mVideoBitsPerSecond, 1, + &recorder->mAudioBitsPerSecond); + } + + // 13. Return recorder. + return recorder.forget(); +} + +/* static */ +already_AddRefed<MediaRecorder> MediaRecorder::Constructor( + const GlobalObject& aGlobal, AudioNode& aAudioNode, + uint32_t aAudioNodeOutput, const MediaRecorderOptions& aOptions, + ErrorResult& aRv) { + // Allow recording from audio node only when pref is on. + if (!Preferences::GetBool("media.recorder.audio_node.enabled", false)) { + // Pretending that this constructor is not defined. + aRv.ThrowTypeError<MSG_DOES_NOT_IMPLEMENT_INTERFACE>("Argument 1", + "MediaStream"); + return nullptr; + } + + nsCOMPtr<nsPIDOMWindowInner> ownerWindow = + do_QueryInterface(aGlobal.GetAsSupports()); + if (!ownerWindow) { + aRv.Throw(NS_ERROR_FAILURE); + return nullptr; + } + + // aAudioNodeOutput doesn't matter to destination node because it has no + // output. + if (aAudioNode.NumberOfOutputs() > 0 && + aAudioNodeOutput >= aAudioNode.NumberOfOutputs()) { + aRv.ThrowIndexSizeError("Invalid AudioNode output index"); + return nullptr; + } + + // When the MediaRecorder() constructor is invoked, the User Agent MUST run + // the following steps: + + // 1. Let stream be the constructor’s first argument. (we'll let audioNode be + // the first arg, and audioNodeOutput the second) + + // 2. Let options be the constructor’s second argument. (we'll let options be + // the third arg) + + // 3. If invoking is type supported with options’ mimeType member as its + // argument returns false, throw a NotSupportedError DOMException and abort + // these steps. + TypeSupport support = IsTypeSupportedImpl(aOptions.mMimeType); + if (support != TypeSupport::Supported) { + // This catches also the empty string mimeType when support for any encoders + // has been disabled. + aRv.ThrowNotSupportedError( + TypeSupportToCString(support, aOptions.mMimeType)); + return nullptr; + } + + // 4. Let recorder be a newly constructed MediaRecorder object. + RefPtr<MediaRecorder> recorder = new MediaRecorder(ownerWindow); + + // 5. Let recorder have a [[ConstrainedMimeType]] internal slot, initialized + // to the value of options' mimeType member. + recorder->mConstrainedMimeType = aOptions.mMimeType; + + // 6. Let recorder have a [[ConstrainedBitsPerSecond]] internal slot, + // initialized to the value of options’ bitsPerSecond member, if it is + // present, otherwise undefined. + recorder->mConstrainedBitsPerSecond = + aOptions.mBitsPerSecond.WasPassed() + ? Some(aOptions.mBitsPerSecond.Value()) + : Nothing(); + + // 7. Initialize recorder’s stream attribute to stream. (make that the + // audioNode and audioNodeOutput equivalents) + recorder->mAudioNode = &aAudioNode; + recorder->mAudioNodeOutput = aAudioNodeOutput; + + // 8. Initialize recorder’s mimeType attribute to the value of recorder’s + // [[ConstrainedMimeType]] slot. + recorder->mMimeType = recorder->mConstrainedMimeType; + + // 9. Initialize recorder’s state attribute to inactive. + recorder->mState = RecordingState::Inactive; + + // 10. Initialize recorder’s videoBitsPerSecond attribute to the value of + // options’ videoBitsPerSecond member, if it is present. Otherwise, choose + // a target value the User Agent deems reasonable for video. + recorder->mVideoBitsPerSecond = aOptions.mVideoBitsPerSecond.WasPassed() + ? aOptions.mVideoBitsPerSecond.Value() + : DEFAULT_VIDEO_BITRATE_BPS; + + // 11. Initialize recorder’s audioBitsPerSecond attribute to the value of + // options’ audioBitsPerSecond member, if it is present. Otherwise, choose + // a target value the User Agent deems reasonable for audio. + recorder->mAudioBitsPerSecond = aOptions.mAudioBitsPerSecond.WasPassed() + ? aOptions.mAudioBitsPerSecond.Value() + : DEFAULT_AUDIO_BITRATE_BPS; + + // 12. If recorder’s [[ConstrainedBitsPerSecond]] slot is not undefined, set + // recorder’s videoBitsPerSecond and audioBitsPerSecond attributes to + // values the User Agent deems reasonable for the respective media types, + // such that the sum of videoBitsPerSecond and audioBitsPerSecond is close + // to the value of recorder’s [[ConstrainedBitsPerSecond]] slot. + if (recorder->mConstrainedBitsPerSecond) { + SelectBitrates(*recorder->mConstrainedBitsPerSecond, 1, + &recorder->mVideoBitsPerSecond, 1, + &recorder->mAudioBitsPerSecond); + } + + // 13. Return recorder. + return recorder.forget(); +} + +/* static */ +bool MediaRecorder::IsTypeSupported(GlobalObject& aGlobal, + const nsAString& aMIMEType) { + return MediaRecorder::IsTypeSupported(aMIMEType); +} + +/* static */ +bool MediaRecorder::IsTypeSupported(const nsAString& aMIMEType) { + return IsTypeSupportedImpl(aMIMEType) == TypeSupport::Supported; +} + +nsresult MediaRecorder::CreateAndDispatchBlobEvent(BlobImpl* aBlobImpl) { + MOZ_ASSERT(NS_IsMainThread(), "Not running on main thread"); + + if (!GetOwnerGlobal()) { + // This MediaRecorder has been disconnected in the meantime. + return NS_ERROR_FAILURE; + } + + RefPtr<Blob> blob = Blob::Create(GetOwnerGlobal(), aBlobImpl); + if (NS_WARN_IF(!blob)) { + return NS_ERROR_FAILURE; + } + + BlobEventInit init; + init.mBubbles = false; + init.mCancelable = false; + init.mData = blob; + + RefPtr<BlobEvent> event = + BlobEvent::Constructor(this, u"dataavailable"_ns, init); + event->SetTrusted(true); + ErrorResult rv; + DispatchEvent(*event, rv); + return rv.StealNSResult(); +} + +void MediaRecorder::DispatchSimpleEvent(const nsAString& aStr) { + MOZ_ASSERT(NS_IsMainThread(), "Not running on main thread"); + nsresult rv = CheckCurrentGlobalCorrectness(); + if (NS_FAILED(rv)) { + return; + } + + rv = DOMEventTargetHelper::DispatchTrustedEvent(aStr); + if (NS_FAILED(rv)) { + LOG(LogLevel::Error, + ("MediaRecorder.DispatchSimpleEvent: DispatchTrustedEvent failed %p", + this)); + NS_ERROR("Failed to dispatch the event!!!"); + } +} + +void MediaRecorder::NotifyError(nsresult aRv) { + MOZ_ASSERT(NS_IsMainThread(), "Not running on main thread"); + nsresult rv = CheckCurrentGlobalCorrectness(); + if (NS_FAILED(rv)) { + return; + } + MediaRecorderErrorEventInit init; + init.mBubbles = false; + init.mCancelable = false; + // These DOMExceptions have been created earlier so they can contain stack + // traces. We attach the appropriate one here to be fired. We should have + // exceptions here, but defensively check. + switch (aRv) { + case NS_ERROR_DOM_SECURITY_ERR: + if (!mSecurityDomException) { + LOG(LogLevel::Debug, ("MediaRecorder.NotifyError: " + "mSecurityDomException was not initialized")); + mSecurityDomException = DOMException::Create(NS_ERROR_DOM_SECURITY_ERR); + } + init.mError = std::move(mSecurityDomException); + break; + default: + if (mOtherDomException && aRv == mOtherDomException->GetResult()) { + LOG(LogLevel::Debug, ("MediaRecorder.NotifyError: " + "mOtherDomException being fired for aRv: %X", + uint32_t(aRv))); + init.mError = std::move(mOtherDomException); + break; + } + if (!mUnknownDomException) { + LOG(LogLevel::Debug, ("MediaRecorder.NotifyError: " + "mUnknownDomException was not initialized")); + mUnknownDomException = DOMException::Create(NS_ERROR_DOM_UNKNOWN_ERR); + } + LOG(LogLevel::Debug, ("MediaRecorder.NotifyError: " + "mUnknownDomException being fired for aRv: %X", + uint32_t(aRv))); + init.mError = std::move(mUnknownDomException); + break; + } + + RefPtr<MediaRecorderErrorEvent> event = + MediaRecorderErrorEvent::Constructor(this, u"error"_ns, init); + event->SetTrusted(true); + + IgnoredErrorResult res; + DispatchEvent(*event, res); + if (res.Failed()) { + NS_ERROR("Failed to dispatch the error event!!!"); + } +} + +void MediaRecorder::RemoveSession(Session* aSession) { + LOG(LogLevel::Debug, ("MediaRecorder.RemoveSession (%p)", aSession)); + mSessions.RemoveElement(aSession); +} + +void MediaRecorder::NotifyOwnerDocumentActivityChanged() { + nsPIDOMWindowInner* window = GetOwner(); + NS_ENSURE_TRUE_VOID(window); + Document* doc = window->GetExtantDoc(); + NS_ENSURE_TRUE_VOID(doc); + + LOG(LogLevel::Debug, ("MediaRecorder %p NotifyOwnerDocumentActivityChanged " + "IsActive=%d, " + "IsVisible=%d, ", + this, doc->IsActive(), doc->IsVisible())); + if (!doc->IsActive() || !doc->IsVisible()) { + // Stop the session. + ErrorResult result; + Stop(result); + result.SuppressException(); + } +} + +void MediaRecorder::Inactivate() { + LOG(LogLevel::Debug, ("MediaRecorder.Inactivate %p", this)); + // The Inactivate the recorder algorithm given a recorder, is as follows: + + // 1. Set recorder’s mimeType attribute to the value of the + // [[ConstrainedMimeType]] slot. + mMimeType = mConstrainedMimeType; + + // 2. Set recorder’s state attribute to inactive. + mState = RecordingState::Inactive; + + // 3. If recorder’s [[ConstrainedBitsPerSecond]] slot is not undefined, set + // recorder’s videoBitsPerSecond and audioBitsPerSecond attributes to + // values the User Agent deems reasonable for the respective media types, + // such that the sum of videoBitsPerSecond and audioBitsPerSecond is close + // to the value of recorder’s [[ConstrainedBitsPerSecond]] slot. + if (mConstrainedBitsPerSecond) { + SelectBitrates(*mConstrainedBitsPerSecond, 1, &mVideoBitsPerSecond, 1, + &mAudioBitsPerSecond); + } +} + +void MediaRecorder::InitializeDomExceptions() { + mSecurityDomException = DOMException::Create(NS_ERROR_DOM_SECURITY_ERR); + mUnknownDomException = DOMException::Create(NS_ERROR_DOM_UNKNOWN_ERR); +} + +RefPtr<MediaRecorder::SizeOfPromise> MediaRecorder::SizeOfExcludingThis( + mozilla::MallocSizeOf aMallocSizeOf) { + MOZ_ASSERT(NS_IsMainThread()); + + // The return type of a chained MozPromise cannot be changed, so we create a + // holder for our desired return type and resolve that from All()->Then(). + auto holder = MakeRefPtr<Refcountable<MozPromiseHolder<SizeOfPromise>>>(); + RefPtr<SizeOfPromise> promise = holder->Ensure(__func__); + + nsTArray<RefPtr<SizeOfPromise>> promises(mSessions.Length()); + for (const RefPtr<Session>& session : mSessions) { + promises.AppendElement(session->SizeOfExcludingThis(aMallocSizeOf)); + } + + SizeOfPromise::All(GetCurrentSerialEventTarget(), promises) + ->Then( + GetCurrentSerialEventTarget(), __func__, + [holder](const nsTArray<size_t>& sizes) { + size_t total = 0; + for (const size_t& size : sizes) { + total += size; + } + holder->Resolve(total, __func__); + }, + []() { MOZ_CRASH("Unexpected reject"); }); + + return promise; +} + +StaticRefPtr<MediaRecorderReporter> MediaRecorderReporter::sUniqueInstance; + +} // namespace mozilla::dom + +#undef LOG |