/* -*- Mode: C++; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*- */ /* vim:set ts=2 sw=2 sts=2 et cindent: */ /* This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ #include "MediaRecorder.h" #include "AudioNodeEngine.h" #include "AudioNodeTrack.h" #include "DOMMediaStream.h" #include "MediaDecoder.h" #include "MediaEncoder.h" #include "MediaTrackGraphImpl.h" #include "VideoUtils.h" #include "mozilla/DOMEventTargetHelper.h" #include "mozilla/dom/AudioStreamTrack.h" #include "mozilla/dom/BlobEvent.h" #include "mozilla/dom/EmptyBlobImpl.h" #include "mozilla/dom/File.h" #include "mozilla/dom/MediaRecorderErrorEvent.h" #include "mozilla/dom/VideoStreamTrack.h" #include "mozilla/media/MediaUtils.h" #include "mozilla/MemoryReporting.h" #include "mozilla/Preferences.h" #include "mozilla/StaticPtr.h" #include "mozilla/TaskQueue.h" #include "nsContentTypeParser.h" #include "nsContentUtils.h" #include "nsDocShell.h" #include "nsError.h" #include "mozilla/dom/Document.h" #include "nsIPrincipal.h" #include "nsIScriptError.h" #include "nsMimeTypes.h" #include "nsProxyRelease.h" #include "nsServiceManagerUtils.h" #include "nsTArray.h" mozilla::LazyLogModule gMediaRecorderLog("MediaRecorder"); #define LOG(type, msg) MOZ_LOG(gMediaRecorderLog, type, msg) constexpr int MIN_VIDEO_BITRATE_BPS = 10e3; // 10kbps constexpr int DEFAULT_VIDEO_BITRATE_BPS = 2500e3; // 2.5Mbps constexpr int MAX_VIDEO_BITRATE_BPS = 100e6; // 100Mbps constexpr int MIN_AUDIO_BITRATE_BPS = 500; // 500bps constexpr int DEFAULT_AUDIO_BITRATE_BPS = 128e3; // 128kbps constexpr int MAX_AUDIO_BITRATE_BPS = 512e3; // 512kbps namespace mozilla::dom { using namespace mozilla::media; /** * MediaRecorderReporter measures memory being used by the Media Recorder. * * It is a singleton reporter and the single class object lives as long as at * least one Recorder is registered. In MediaRecorder, the reporter is * unregistered when it is destroyed. */ class MediaRecorderReporter final : public nsIMemoryReporter { public: static void AddMediaRecorder(MediaRecorder* aRecorder) { if (!sUniqueInstance) { sUniqueInstance = MakeAndAddRef(); RegisterWeakAsyncMemoryReporter(sUniqueInstance); } sUniqueInstance->mRecorders.AppendElement(aRecorder); } static void RemoveMediaRecorder(MediaRecorder* aRecorder) { if (!sUniqueInstance) { return; } sUniqueInstance->mRecorders.RemoveElement(aRecorder); if (sUniqueInstance->mRecorders.IsEmpty()) { UnregisterWeakMemoryReporter(sUniqueInstance); sUniqueInstance = nullptr; } } NS_DECL_THREADSAFE_ISUPPORTS MediaRecorderReporter() = default; NS_IMETHOD CollectReports(nsIHandleReportCallback* aHandleReport, nsISupports* aData, bool aAnonymize) override { nsTArray> promises; for (const RefPtr& recorder : mRecorders) { promises.AppendElement(recorder->SizeOfExcludingThis(MallocSizeOf)); } nsCOMPtr handleReport = aHandleReport; nsCOMPtr data = aData; MediaRecorder::SizeOfPromise::All(GetCurrentSerialEventTarget(), promises) ->Then( GetCurrentSerialEventTarget(), __func__, [handleReport, data](const nsTArray& sizes) { nsCOMPtr manager = do_GetService("@mozilla.org/memory-reporter-manager;1"); if (!manager) { return; } size_t sum = 0; for (const size_t& size : sizes) { sum += size; } handleReport->Callback(""_ns, "explicit/media/recorder"_ns, KIND_HEAP, UNITS_BYTES, sum, "Memory used by media recorder."_ns, data); manager->EndReport(); }, [](size_t) { MOZ_CRASH("Unexpected reject"); }); return NS_OK; } private: MOZ_DEFINE_MALLOC_SIZE_OF(MallocSizeOf) virtual ~MediaRecorderReporter() { MOZ_ASSERT(mRecorders.IsEmpty(), "All recorders must have been removed"); } static StaticRefPtr sUniqueInstance; nsTArray> mRecorders; }; NS_IMPL_ISUPPORTS(MediaRecorderReporter, nsIMemoryReporter); NS_IMPL_CYCLE_COLLECTION_CLASS(MediaRecorder) NS_IMPL_CYCLE_COLLECTION_TRAVERSE_BEGIN_INHERITED(MediaRecorder, DOMEventTargetHelper) NS_IMPL_CYCLE_COLLECTION_TRAVERSE(mStream) NS_IMPL_CYCLE_COLLECTION_TRAVERSE(mAudioNode) NS_IMPL_CYCLE_COLLECTION_TRAVERSE(mOtherDomException) NS_IMPL_CYCLE_COLLECTION_TRAVERSE(mSecurityDomException) NS_IMPL_CYCLE_COLLECTION_TRAVERSE(mUnknownDomException) NS_IMPL_CYCLE_COLLECTION_TRAVERSE(mDocument) NS_IMPL_CYCLE_COLLECTION_TRAVERSE_END NS_IMPL_CYCLE_COLLECTION_UNLINK_BEGIN_INHERITED(MediaRecorder, DOMEventTargetHelper) NS_IMPL_CYCLE_COLLECTION_UNLINK(mStream) NS_IMPL_CYCLE_COLLECTION_UNLINK(mAudioNode) NS_IMPL_CYCLE_COLLECTION_UNLINK(mOtherDomException) NS_IMPL_CYCLE_COLLECTION_UNLINK(mSecurityDomException) NS_IMPL_CYCLE_COLLECTION_UNLINK(mUnknownDomException) tmp->UnRegisterActivityObserver(); NS_IMPL_CYCLE_COLLECTION_UNLINK(mDocument) NS_IMPL_CYCLE_COLLECTION_UNLINK_END NS_INTERFACE_MAP_BEGIN_CYCLE_COLLECTION(MediaRecorder) NS_INTERFACE_MAP_ENTRY(nsIDocumentActivity) NS_INTERFACE_MAP_END_INHERITING(DOMEventTargetHelper) NS_IMPL_ADDREF_INHERITED(MediaRecorder, DOMEventTargetHelper) NS_IMPL_RELEASE_INHERITED(MediaRecorder, DOMEventTargetHelper) namespace { bool PrincipalSubsumes(MediaRecorder* aRecorder, nsIPrincipal* aPrincipal) { if (!aRecorder->GetOwner()) { return false; } nsCOMPtr doc = aRecorder->GetOwner()->GetExtantDoc(); if (!doc) { return false; } if (!aPrincipal) { return false; } bool subsumes; if (NS_FAILED(doc->NodePrincipal()->Subsumes(aPrincipal, &subsumes))) { return false; } return subsumes; } bool MediaStreamTracksPrincipalSubsumes( MediaRecorder* aRecorder, const nsTArray>& aTracks) { nsCOMPtr principal = nullptr; for (const auto& track : aTracks) { nsContentUtils::CombineResourcePrincipals(&principal, track->GetPrincipal()); } return PrincipalSubsumes(aRecorder, principal); } bool AudioNodePrincipalSubsumes(MediaRecorder* aRecorder, AudioNode* aAudioNode) { MOZ_ASSERT(aAudioNode); Document* doc = aAudioNode->GetOwner() ? aAudioNode->GetOwner()->GetExtantDoc() : nullptr; nsCOMPtr principal = doc ? doc->NodePrincipal() : nullptr; return PrincipalSubsumes(aRecorder, principal); } // This list is sorted so that lesser failures are later, so that // IsTypeSupportedImpl() can report the error from audio or video types that // is closer to being supported. enum class TypeSupport { MediaTypeInvalid, NoVideoWithAudioType, ContainersDisabled, CodecsDisabled, ContainerUnsupported, CodecUnsupported, CodecDuplicated, Supported, }; nsCString TypeSupportToCString(TypeSupport aSupport, const nsAString& aMimeType) { nsAutoCString mime = NS_ConvertUTF16toUTF8(aMimeType); switch (aSupport) { case TypeSupport::Supported: return nsPrintfCString("%s is supported", mime.get()); case TypeSupport::MediaTypeInvalid: return nsPrintfCString("%s is not a valid media type", mime.get()); case TypeSupport::NoVideoWithAudioType: return nsPrintfCString( "Video cannot be recorded with %s as it is an audio type", mime.get()); case TypeSupport::ContainersDisabled: return "All containers are disabled"_ns; case TypeSupport::CodecsDisabled: return "All codecs are disabled"_ns; case TypeSupport::ContainerUnsupported: return nsPrintfCString("%s indicates an unsupported container", mime.get()); case TypeSupport::CodecUnsupported: return nsPrintfCString("%s indicates an unsupported codec", mime.get()); case TypeSupport::CodecDuplicated: return nsPrintfCString("%s contains the same codec multiple times", mime.get()); default: MOZ_ASSERT_UNREACHABLE("Unknown TypeSupport"); return "Unknown error"_ns; } } TypeSupport CanRecordAudioTrackWith(const Maybe& aMimeType, const nsAString& aMimeTypeString) { if (aMimeTypeString.IsEmpty()) { // For the empty string we just need to check whether we have support for an // audio container and an audio codec. if (!MediaEncoder::IsWebMEncoderEnabled() && !MediaDecoder::IsOggEnabled()) { // No container support for audio. return TypeSupport::ContainersDisabled; } if (!MediaDecoder::IsOpusEnabled()) { // No codec support for audio. return TypeSupport::CodecsDisabled; } return TypeSupport::Supported; } if (!aMimeType) { // A mime type string was set, but it couldn't be parsed to a valid // MediaContainerType. return TypeSupport::MediaTypeInvalid; } if (aMimeType->Type() != MEDIAMIMETYPE(VIDEO_WEBM) && aMimeType->Type() != MEDIAMIMETYPE(AUDIO_WEBM) && aMimeType->Type() != MEDIAMIMETYPE(AUDIO_OGG)) { // Any currently supported container can record audio. return TypeSupport::ContainerUnsupported; } if (aMimeType->Type() == MEDIAMIMETYPE(VIDEO_WEBM) && !MediaEncoder::IsWebMEncoderEnabled()) { return TypeSupport::ContainerUnsupported; } if (aMimeType->Type() == MEDIAMIMETYPE(AUDIO_WEBM) && !MediaEncoder::IsWebMEncoderEnabled()) { return TypeSupport::ContainerUnsupported; } if (aMimeType->Type() == MEDIAMIMETYPE(AUDIO_OGG) && !MediaDecoder::IsOggEnabled()) { return TypeSupport::ContainerUnsupported; } if (!MediaDecoder::IsOpusEnabled()) { return TypeSupport::CodecUnsupported; } if (!aMimeType->ExtendedType().HaveCodecs()) { // No codecs constrained, we can pick opus. return TypeSupport::Supported; } size_t opus = 0; size_t unknown = 0; for (const auto& codec : aMimeType->ExtendedType().Codecs().Range()) { // Ignore video codecs. if (codec.EqualsLiteral("vp8")) { continue; } if (codec.EqualsLiteral("vp8.0")) { continue; } if (codec.EqualsLiteral("opus")) { // All containers support opus opus++; continue; } unknown++; } if (unknown > 0) { // Unsupported codec. return TypeSupport::CodecUnsupported; } if (opus == 0) { // Codecs specified but not opus. Unsupported for audio. return TypeSupport::CodecUnsupported; } if (opus > 1) { // Opus specified more than once. Bad form. return TypeSupport::CodecDuplicated; } return TypeSupport::Supported; } TypeSupport CanRecordVideoTrackWith(const Maybe& aMimeType, const nsAString& aMimeTypeString) { if (aMimeTypeString.IsEmpty()) { // For the empty string we just need to check whether we have support for a // video container and a video codec. The VP8 encoder is always available. if (!MediaEncoder::IsWebMEncoderEnabled()) { // No container support for video. return TypeSupport::ContainersDisabled; } return TypeSupport::Supported; } if (!aMimeType) { // A mime type string was set, but it couldn't be parsed to a valid // MediaContainerType. return TypeSupport::MediaTypeInvalid; } if (!aMimeType->Type().HasVideoMajorType()) { return TypeSupport::NoVideoWithAudioType; } if (aMimeType->Type() != MEDIAMIMETYPE(VIDEO_WEBM)) { return TypeSupport::ContainerUnsupported; } if (!MediaEncoder::IsWebMEncoderEnabled()) { return TypeSupport::ContainerUnsupported; } if (!aMimeType->ExtendedType().HaveCodecs()) { // No codecs constrained, we can pick vp8. return TypeSupport::Supported; } size_t vp8 = 0; size_t unknown = 0; for (const auto& codec : aMimeType->ExtendedType().Codecs().Range()) { if (codec.EqualsLiteral("opus")) { // Ignore audio codecs. continue; } if (codec.EqualsLiteral("vp8")) { vp8++; continue; } if (codec.EqualsLiteral("vp8.0")) { vp8++; continue; } unknown++; } if (unknown > 0) { // Unsupported codec. return TypeSupport::CodecUnsupported; } if (vp8 == 0) { // Codecs specified but not vp8. Unsupported for video. return TypeSupport::CodecUnsupported; } if (vp8 > 1) { // Vp8 specified more than once. Bad form. return TypeSupport::CodecDuplicated; } return TypeSupport::Supported; } TypeSupport CanRecordWith(MediaStreamTrack* aTrack, const Maybe& aMimeType, const nsAString& aMimeTypeString) { if (aTrack->AsAudioStreamTrack()) { return CanRecordAudioTrackWith(aMimeType, aMimeTypeString); } if (aTrack->AsVideoStreamTrack()) { return CanRecordVideoTrackWith(aMimeType, aMimeTypeString); } MOZ_CRASH("Unexpected track type"); } TypeSupport IsTypeSupportedImpl(const nsAString& aMIMEType) { if (aMIMEType.IsEmpty()) { // Lie and return true even if no container/codec support is enabled, // because the spec mandates it. return TypeSupport::Supported; } Maybe mime = MakeMediaContainerType(aMIMEType); TypeSupport audioSupport = CanRecordAudioTrackWith(mime, aMIMEType); TypeSupport videoSupport = CanRecordVideoTrackWith(mime, aMIMEType); return std::max(audioSupport, videoSupport); } nsString SelectMimeType(bool aHasVideo, bool aHasAudio, const nsString& aConstrainedMimeType) { MOZ_ASSERT(aHasVideo || aHasAudio); Maybe constrainedType = MakeMediaContainerType(aConstrainedMimeType); // If we are recording video, Start() should have rejected any non-video mime // types. MOZ_ASSERT_IF(constrainedType && aHasVideo, constrainedType->Type().HasVideoMajorType()); // IsTypeSupported() rejects application mime types. MOZ_ASSERT_IF(constrainedType, !constrainedType->Type().HasApplicationMajorType()); nsString result; if (constrainedType && constrainedType->ExtendedType().HaveCodecs()) { // The constrained mime type is fully defined (it has codecs!). No need to // select anything. CopyUTF8toUTF16(constrainedType->OriginalString(), result); } else { // There is no constrained mime type, or there is and it is not fully // defined but still valid. Select what's missing, so that we have major // type, container and codecs. // If there is a constrained mime type it should not have codecs defined, // because then it is fully defined and used unchanged (covered earlier). MOZ_ASSERT_IF(constrainedType, !constrainedType->ExtendedType().HaveCodecs()); nsCString majorType; { if (constrainedType) { // There is a constrained type. It has both major type and container in // order to be valid. Use them as is. majorType = constrainedType->Type().AsString(); } else if (aHasVideo) { majorType = nsLiteralCString(VIDEO_WEBM); } else { majorType = nsLiteralCString(AUDIO_OGG); } } nsCString codecs; { if (aHasVideo && aHasAudio) { codecs = "\"vp8, opus\""_ns; } else if (aHasVideo) { codecs = "vp8"_ns; } else { codecs = "opus"_ns; } } result = NS_ConvertUTF8toUTF16( nsPrintfCString("%s; codecs=%s", majorType.get(), codecs.get())); } MOZ_ASSERT_IF(aHasAudio, CanRecordAudioTrackWith(MakeMediaContainerType(result), result) == TypeSupport::Supported); MOZ_ASSERT_IF(aHasVideo, CanRecordVideoTrackWith(MakeMediaContainerType(result), result) == TypeSupport::Supported); return result; } void SelectBitrates(uint32_t aBitsPerSecond, uint8_t aNumVideoTracks, uint32_t* aOutVideoBps, uint8_t aNumAudioTracks, uint32_t* aOutAudioBps) { uint32_t vbps = 0; uint32_t abps = 0; const uint32_t minVideoBps = MIN_VIDEO_BITRATE_BPS * aNumVideoTracks; const uint32_t maxVideoBps = MAX_VIDEO_BITRATE_BPS * aNumVideoTracks; const uint32_t minAudioBps = MIN_AUDIO_BITRATE_BPS * aNumAudioTracks; const uint32_t maxAudioBps = MAX_AUDIO_BITRATE_BPS * aNumAudioTracks; if (aNumVideoTracks == 0) { MOZ_DIAGNOSTIC_ASSERT(aNumAudioTracks > 0); abps = std::min(maxAudioBps, std::max(minAudioBps, aBitsPerSecond)); } else if (aNumAudioTracks == 0) { vbps = std::min(maxVideoBps, std::max(minVideoBps, aBitsPerSecond)); } else { // Scale the bits so that video gets 20 times the bits of audio. // Since we must account for varying number of tracks of each type we weight // them by type; video = weight 20, audio = weight 1. const uint32_t videoWeight = aNumVideoTracks * 20; const uint32_t audioWeight = aNumAudioTracks; const uint32_t totalWeights = audioWeight + videoWeight; const uint32_t videoBitrate = uint64_t(aBitsPerSecond) * videoWeight / totalWeights; const uint32_t audioBitrate = uint64_t(aBitsPerSecond) * audioWeight / totalWeights; vbps = std::min(maxVideoBps, std::max(minVideoBps, videoBitrate)); abps = std::min(maxAudioBps, std::max(minAudioBps, audioBitrate)); } *aOutVideoBps = vbps; *aOutAudioBps = abps; } } // namespace /** * Session is an object to represent a single recording event. * In original design, all recording context is stored in MediaRecorder, which * causes a problem if someone calls MediaRecorder::Stop and * MediaRecorder::Start quickly. To prevent blocking main thread, media encoding * is executed in a second thread, named encoder thread. For the same reason, we * do not await encoder thread shutdown in MediaRecorder::Stop. * If someone calls MediaRecorder::Start before encoder thread shutdown, the * same recording context in MediaRecorder might be accessed by two distinct * encoder threads, which would be racy. With the recording context, including * the encoder thread, in a Session object the problem is solved. * * Lifetime of MediaRecorder and Session objects. * 1) MediaRecorder creates a Session in MediaRecorder::Start() and holds * a reference to it. Then the Session registers itself to a ShutdownBlocker * and also holds a reference to MediaRecorder. * Therefore, the reference dependency in gecko is: * ShutdownBlocker -> Session <-> MediaRecorder, note that there is a cycle * reference between Session and MediaRecorder. * 2) A Session is destroyed after Session::DoSessionEndTask() has been called * _and_ all encoded media data has been passed to OnDataAvailable handler. * In some cases the encoded media can be discarded before being passed to * the OnDataAvailable handler. * 3) Session::DoSessionEndTask is called by an application through * MediaRecorder::Stop(), from a MediaEncoder Shutdown notification, from the * document going inactive or invisible, or from the ShutdownBlocker. */ class MediaRecorder::Session : public PrincipalChangeObserver, public DOMMediaStream::TrackListener { NS_INLINE_DECL_THREADSAFE_REFCOUNTING(Session) struct TrackTypeComparator { enum Type { AUDIO, VIDEO, }; static bool Equals(const RefPtr& aTrack, Type aType) { return (aType == AUDIO && aTrack->AsAudioStreamTrack()) || (aType == VIDEO && aTrack->AsVideoStreamTrack()); } }; public: Session(MediaRecorder* aRecorder, nsTArray> aMediaStreamTracks, uint32_t aVideoBitsPerSecond, uint32_t aAudioBitsPerSecond) : mRecorder(aRecorder), mMediaStreamTracks(std::move(aMediaStreamTracks)), mMainThread(mRecorder->GetOwner()->EventTargetFor(TaskCategory::Other)), mMimeType(SelectMimeType( mMediaStreamTracks.Contains(TrackTypeComparator::VIDEO, TrackTypeComparator()), mRecorder->mAudioNode || mMediaStreamTracks.Contains(TrackTypeComparator::AUDIO, TrackTypeComparator()), mRecorder->mConstrainedMimeType)), mVideoBitsPerSecond(aVideoBitsPerSecond), mAudioBitsPerSecond(aAudioBitsPerSecond), mStartTime(TimeStamp::Now()), mRunningState(RunningState::Idling) { MOZ_ASSERT(NS_IsMainThread()); Telemetry::ScalarAdd(Telemetry::ScalarID::MEDIARECORDER_RECORDING_COUNT, 1); } void PrincipalChanged(MediaStreamTrack* aTrack) override { NS_ASSERTION(mMediaStreamTracks.Contains(aTrack), "Principal changed for unrecorded track"); if (!MediaStreamTracksPrincipalSubsumes(mRecorder, mMediaStreamTracks)) { DoSessionEndTask(NS_ERROR_DOM_SECURITY_ERR); } } void NotifyTrackAdded(const RefPtr& aTrack) override { LOG(LogLevel::Warning, ("Session.NotifyTrackAdded %p Raising error due to track set change", this)); // There's a chance we have a sensible JS stack here. if (!mRecorder->mOtherDomException) { mRecorder->mOtherDomException = DOMException::Create( NS_ERROR_DOM_INVALID_MODIFICATION_ERR, "An attempt was made to add a track to the recorded MediaStream " "during the recording"_ns); } DoSessionEndTask(NS_ERROR_DOM_INVALID_MODIFICATION_ERR); } void NotifyTrackRemoved(const RefPtr& aTrack) override { if (aTrack->Ended()) { // TrackEncoder will pickup tracks that end itself. return; } LOG(LogLevel::Warning, ("Session.NotifyTrackRemoved %p Raising error due to track set change", this)); // There's a chance we have a sensible JS stack here. if (!mRecorder->mOtherDomException) { mRecorder->mOtherDomException = DOMException::Create( NS_ERROR_DOM_INVALID_MODIFICATION_ERR, "An attempt was made to remove a track from the recorded MediaStream " "during the recording"_ns); } DoSessionEndTask(NS_ERROR_DOM_INVALID_MODIFICATION_ERR); } void Start(TimeDuration aTimeslice) { LOG(LogLevel::Debug, ("Session.Start %p", this)); MOZ_ASSERT(NS_IsMainThread()); if (mRecorder->mStream) { // The TrackListener reports back when tracks are added or removed from // the MediaStream. mMediaStream = mRecorder->mStream; mMediaStream->RegisterTrackListener(this); uint8_t trackTypes = 0; for (const auto& track : mMediaStreamTracks) { if (track->AsAudioStreamTrack()) { trackTypes |= ContainerWriter::CREATE_AUDIO_TRACK; } else if (track->AsVideoStreamTrack()) { trackTypes |= ContainerWriter::CREATE_VIDEO_TRACK; } else { MOZ_CRASH("Unexpected track type"); } } for (const auto& t : mMediaStreamTracks) { t->AddPrincipalChangeObserver(this); } LOG(LogLevel::Debug, ("Session.Start track types = (%d)", trackTypes)); InitEncoder(trackTypes, mMediaStreamTracks[0]->Graph()->GraphRate(), aTimeslice); return; } if (mRecorder->mAudioNode) { TrackRate trackRate = mRecorder->mAudioNode->Context()->Graph()->GraphRate(); // Web Audio node has only audio. InitEncoder(ContainerWriter::CREATE_AUDIO_TRACK, trackRate, aTimeslice); return; } MOZ_ASSERT(false, "Unknown source"); } void Stop() { LOG(LogLevel::Debug, ("Session.Stop %p", this)); MOZ_ASSERT(NS_IsMainThread()); if (mEncoder) { mEncoder->DisconnectTracks(); } // Remove main thread state added in Start(). if (mMediaStream) { mMediaStream->UnregisterTrackListener(this); mMediaStream = nullptr; } { for (const auto& track : mMediaStreamTracks) { track->RemovePrincipalChangeObserver(this); } } if (mRunningState.isOk() && mRunningState.inspect() == RunningState::Idling) { LOG(LogLevel::Debug, ("Session.Stop Explicit end task %p", this)); // End the Session directly if there is no encoder. DoSessionEndTask(NS_OK); } else if (mRunningState.isOk() && (mRunningState.inspect() == RunningState::Starting || mRunningState.inspect() == RunningState::Running)) { if (mRunningState.inspect() == RunningState::Starting) { // The MediaEncoder might not report started, but by spec we must fire // "start". mStartedListener.DisconnectIfExists(); NS_DispatchToMainThread(NewRunnableMethod( "MediaRecorder::Session::Stop", this, &Session::OnStarted)); } mRunningState = RunningState::Stopping; } } void Pause() { LOG(LogLevel::Debug, ("Session.Pause")); MOZ_ASSERT(NS_IsMainThread()); MOZ_ASSERT_IF(mRunningState.isOk(), mRunningState.unwrap() != RunningState::Idling); if (mRunningState.isErr() || mRunningState.unwrap() == RunningState::Stopping || mRunningState.unwrap() == RunningState::Stopped) { return; } MOZ_ASSERT(mEncoder); mEncoder->Suspend(); } void Resume() { LOG(LogLevel::Debug, ("Session.Resume")); MOZ_ASSERT(NS_IsMainThread()); MOZ_ASSERT_IF(mRunningState.isOk(), mRunningState.unwrap() != RunningState::Idling); if (mRunningState.isErr() || mRunningState.unwrap() == RunningState::Stopping || mRunningState.unwrap() == RunningState::Stopped) { return; } MOZ_ASSERT(mEncoder); mEncoder->Resume(); } void RequestData() { LOG(LogLevel::Debug, ("Session.RequestData")); MOZ_ASSERT(NS_IsMainThread()); MOZ_ASSERT(mEncoder); InvokeAsync(mEncoderThread, mEncoder.get(), __func__, &MediaEncoder::RequestData) ->Then( mMainThread, __func__, [this, self = RefPtr(this)]( const MediaEncoder::BlobPromise::ResolveOrRejectValue& aRrv) { if (aRrv.IsReject()) { LOG(LogLevel::Warning, ("RequestData failed")); DoSessionEndTask(aRrv.RejectValue()); return; } nsresult rv = mRecorder->CreateAndDispatchBlobEvent(aRrv.ResolveValue()); if (NS_FAILED(rv)) { DoSessionEndTask(NS_OK); } }); } public: RefPtr SizeOfExcludingThis( mozilla::MallocSizeOf aMallocSizeOf) { MOZ_ASSERT(NS_IsMainThread()); if (!mEncoder) { return SizeOfPromise::CreateAndResolve(0, __func__); } return mEncoder->SizeOfExcludingThis(aMallocSizeOf); } private: virtual ~Session() { MOZ_ASSERT(NS_IsMainThread()); MOZ_ASSERT(mShutdownPromise); MOZ_ASSERT(!mShutdownBlocker); LOG(LogLevel::Debug, ("Session.~Session (%p)", this)); } void InitEncoder(uint8_t aTrackTypes, TrackRate aTrackRate, TimeDuration aTimeslice) { LOG(LogLevel::Debug, ("Session.InitEncoder %p", this)); MOZ_ASSERT(NS_IsMainThread()); if (!mRunningState.isOk() || mRunningState.inspect() != RunningState::Idling) { MOZ_ASSERT_UNREACHABLE("Double-init"); return; } // Create a TaskQueue to read encode media data from MediaEncoder. MOZ_RELEASE_ASSERT(!mEncoderThread); RefPtr pool = GetMediaThreadPool(MediaThreadType::WEBRTC_WORKER); if (!pool) { LOG(LogLevel::Debug, ("Session.InitEncoder %p Failed to create " "MediaRecorderReadThread thread pool", this)); DoSessionEndTask(NS_ERROR_FAILURE); return; } mEncoderThread = TaskQueue::Create(pool.forget(), "MediaRecorderReadThread"); MOZ_DIAGNOSTIC_ASSERT(!mShutdownBlocker); // Add a shutdown blocker so mEncoderThread can be shutdown async. class Blocker : public ShutdownBlocker { const RefPtr mSession; public: Blocker(RefPtr aSession, const nsString& aName) : ShutdownBlocker(aName), mSession(std::move(aSession)) {} NS_IMETHOD BlockShutdown(nsIAsyncShutdownClient*) override { mSession->DoSessionEndTask(NS_ERROR_ABORT); return NS_OK; } }; nsCOMPtr barrier = GetShutdownBarrier(); if (!barrier) { LOG(LogLevel::Error, ("Session.InitEncoder %p Failed to get shutdown barrier", this)); DoSessionEndTask(NS_ERROR_FAILURE); return; } nsString name; name.AppendPrintf("MediaRecorder::Session %p shutdown", this); mShutdownBlocker = MakeAndAddRef(this, name); nsresult rv = barrier->AddBlocker( mShutdownBlocker, NS_LITERAL_STRING_FROM_CSTRING(__FILE__), __LINE__, u"MediaRecorder::Session: shutdown"_ns); MOZ_RELEASE_ASSERT(NS_SUCCEEDED(rv)); uint32_t maxMemory = Preferences::GetUint("media.recorder.max_memory", MAX_ALLOW_MEMORY_BUFFER); mEncoder = MediaEncoder::CreateEncoder( mEncoderThread, mMimeType, mAudioBitsPerSecond, mVideoBitsPerSecond, aTrackTypes, aTrackRate, maxMemory, aTimeslice); if (!mEncoder) { LOG(LogLevel::Error, ("Session.InitEncoder !mEncoder %p", this)); DoSessionEndTask(NS_ERROR_ABORT); return; } mStartedListener = mEncoder->StartedEvent().Connect(mMainThread, this, &Session::OnStarted); mDataAvailableListener = mEncoder->DataAvailableEvent().Connect( mMainThread, this, &Session::OnDataAvailable); mErrorListener = mEncoder->ErrorEvent().Connect(mMainThread, this, &Session::OnError); mShutdownListener = mEncoder->ShutdownEvent().Connect(mMainThread, this, &Session::OnShutdown); if (mRecorder->mAudioNode) { mEncoder->ConnectAudioNode(mRecorder->mAudioNode, mRecorder->mAudioNodeOutput); } for (const auto& track : mMediaStreamTracks) { mEncoder->ConnectMediaStreamTrack(track); } // Set mRunningState to Running so that DoSessionEndTask will // take the responsibility to end the session. mRunningState = RunningState::Starting; } // This is the task that will stop recording per spec: // - If rv is NS_ERROR_ABORT or NS_ERROR_DOM_SECURITY_ERR, cancel the encoders // - Otherwise, stop the encoders gracefully, this still encodes buffered data // - Set state to "inactive" // - Fire an error event, if NS_FAILED(rv) // - Discard blob data if rv is NS_ERROR_DOM_SECURITY_ERR // - Fire a Blob event // - Fire an event named stop void DoSessionEndTask(nsresult rv) { MOZ_ASSERT(NS_IsMainThread()); if (mRunningState.isErr()) { // We have already ended with an error. return; } if (mRunningState.isOk() && mRunningState.inspect() == RunningState::Stopped) { // We have already ended gracefully. return; } bool needsStartEvent = false; if (mRunningState.isOk() && (mRunningState.inspect() == RunningState::Idling || mRunningState.inspect() == RunningState::Starting)) { needsStartEvent = true; } // Set a terminated running state. Future DoSessionEnd tasks will exit // early. if (rv == NS_OK) { mRunningState = RunningState::Stopped; } else { mRunningState = Err(rv); } RefPtr blobPromise; if (!mEncoder) { blobPromise = MediaEncoder::BlobPromise::CreateAndReject(NS_OK, __func__); } else { blobPromise = (rv == NS_ERROR_ABORT || rv == NS_ERROR_DOM_SECURITY_ERR ? mEncoder->Cancel() : mEncoder->Stop()) ->Then(mEncoderThread, __func__, [encoder = mEncoder]( const GenericNonExclusivePromise::ResolveOrRejectValue& aValue) { MOZ_DIAGNOSTIC_ASSERT(aValue.IsResolve()); return encoder->RequestData(); }); } blobPromise ->Then( mMainThread, __func__, [this, self = RefPtr(this), rv, needsStartEvent]( const MediaEncoder::BlobPromise::ResolveOrRejectValue& aRv) { if (mRecorder->mSessions.LastElement() == this) { // Set state to inactive, but only if the recorder is not // controlled by another session already. mRecorder->Inactivate(); } if (needsStartEvent) { mRecorder->DispatchSimpleEvent(u"start"_ns); } // If there was an error, Fire the appropriate one if (NS_FAILED(rv)) { mRecorder->NotifyError(rv); } // Fire a blob event named dataavailable RefPtr blobImpl; if (rv == NS_ERROR_DOM_SECURITY_ERR || aRv.IsReject()) { // In case of SecurityError, the blob data must be discarded. // We create a new empty one and throw the blob with its data // away. // In case we failed to gather blob data, we create an empty // memory blob instead. blobImpl = new EmptyBlobImpl(mMimeType); } else { blobImpl = aRv.ResolveValue(); } if (NS_FAILED(mRecorder->CreateAndDispatchBlobEvent(blobImpl))) { // Failed to dispatch blob event. That's unexpected. It's // probably all right to fire an error event if we haven't // already. if (NS_SUCCEEDED(rv)) { mRecorder->NotifyError(NS_ERROR_FAILURE); } } // Fire an event named stop mRecorder->DispatchSimpleEvent(u"stop"_ns); // And finally, Shutdown and destroy the Session return Shutdown(); }) ->Then(mMainThread, __func__, [this, self = RefPtr(this)] { // Guard against the case where we fail to add a blocker due to being // in XPCOM shutdown. If we're in this state we shouldn't try and get // a shutdown barrier as we'll fail. if (!mShutdownBlocker) { return; } MustGetShutdownBarrier()->RemoveBlocker(mShutdownBlocker); mShutdownBlocker = nullptr; }); } void OnStarted() { MOZ_ASSERT(NS_IsMainThread()); if (mRunningState.isErr()) { return; } RunningState state = mRunningState.inspect(); if (state == RunningState::Starting || state == RunningState::Stopping) { if (state == RunningState::Starting) { // We set it to Running in the runnable since we can only assign // mRunningState on main thread. We set it before running the start // event runnable since that dispatches synchronously (and may cause // js calls to methods depending on mRunningState). mRunningState = RunningState::Running; mRecorder->mMimeType = mEncoder->mMimeType; } mRecorder->DispatchSimpleEvent(u"start"_ns); } } void OnDataAvailable(const RefPtr& aBlob) { if (mRunningState.isErr() && mRunningState.unwrapErr() == NS_ERROR_DOM_SECURITY_ERR) { return; } if (NS_WARN_IF(NS_FAILED(mRecorder->CreateAndDispatchBlobEvent(aBlob)))) { LOG(LogLevel::Warning, ("MediaRecorder %p Creating or dispatching BlobEvent failed", this)); DoSessionEndTask(NS_OK); } } void OnError() { MOZ_ASSERT(NS_IsMainThread()); DoSessionEndTask(NS_ERROR_FAILURE); } void OnShutdown() { MOZ_ASSERT(NS_IsMainThread()); DoSessionEndTask(NS_OK); } RefPtr Shutdown() { MOZ_ASSERT(NS_IsMainThread()); LOG(LogLevel::Debug, ("Session Shutdown %p", this)); if (mShutdownPromise) { return mShutdownPromise; } // This is a coarse calculation and does not reflect the duration of the // final recording for reasons such as pauses. However it allows us an // idea of how long people are running their recorders for. TimeDuration timeDelta = TimeStamp::Now() - mStartTime; Telemetry::Accumulate(Telemetry::MEDIA_RECORDER_RECORDING_DURATION, timeDelta.ToSeconds()); mShutdownPromise = ShutdownPromise::CreateAndResolve(true, __func__); if (mEncoder) { mShutdownPromise = mShutdownPromise ->Then(mMainThread, __func__, [this, self = RefPtr(this)] { mStartedListener.DisconnectIfExists(); mDataAvailableListener.DisconnectIfExists(); mErrorListener.DisconnectIfExists(); mShutdownListener.DisconnectIfExists(); return mEncoder->Cancel(); }) ->Then(mEncoderThread, __func__, [] { // Meh, this is just to convert the promise type to match // mShutdownPromise. return ShutdownPromise::CreateAndResolve(true, __func__); }); } // Remove main thread state. This could be needed if Stop() wasn't called. if (mMediaStream) { mMediaStream->UnregisterTrackListener(this); mMediaStream = nullptr; } { auto tracks(std::move(mMediaStreamTracks)); for (RefPtr& track : tracks) { track->RemovePrincipalChangeObserver(this); } } // Break the cycle reference between Session and MediaRecorder. mShutdownPromise = mShutdownPromise->Then( mMainThread, __func__, [self = RefPtr(this)]() { self->mRecorder->RemoveSession(self); return ShutdownPromise::CreateAndResolve(true, __func__); }, []() { MOZ_ASSERT_UNREACHABLE("Unexpected reject"); return ShutdownPromise::CreateAndReject(false, __func__); }); if (mEncoderThread) { mShutdownPromise = mShutdownPromise->Then( mMainThread, __func__, [encoderThread = mEncoderThread]() { return encoderThread->BeginShutdown(); }, []() { MOZ_ASSERT_UNREACHABLE("Unexpected reject"); return ShutdownPromise::CreateAndReject(false, __func__); }); } return mShutdownPromise; } private: enum class RunningState { Idling, // Session has been created Starting, // MediaEncoder started, waiting for data Running, // MediaEncoder has received data Stopping, // Stop() has been called Stopped, // Session has stopped without any error }; // Our associated MediaRecorder. const RefPtr mRecorder; // Stream currently recorded. RefPtr mMediaStream; // Tracks currently recorded. This should be a subset of mMediaStream's track // set. nsTArray> mMediaStreamTracks; // Main thread used for MozPromise operations. const RefPtr mMainThread; // Runnable thread for reading data from MediaEncoder. RefPtr mEncoderThread; // MediaEncoder pipeline. RefPtr mEncoder; // Listener connected to mMediaEncoder::StartedEvent(). MediaEventListener mStartedListener; // Listener connected to mMediaEncoder::DataAvailableEvent(). MediaEventListener mDataAvailableListener; // Listener connected to mMediaEncoder::ErrorEvent(). MediaEventListener mErrorListener; // Listener connected to mMediaEncoder::ShutdownEvent(). MediaEventListener mShutdownListener; // Set in Shutdown() and resolved when shutdown is complete. RefPtr mShutdownPromise; // Session mimeType const nsString mMimeType; // The video bitrate the recorder was configured with. const uint32_t mVideoBitsPerSecond; // The audio bitrate the recorder was configured with. const uint32_t mAudioBitsPerSecond; // The time this session started, for telemetry. const TimeStamp mStartTime; // The session's current main thread state. The error type gets set when // ending a recording with an error. An NS_OK error is invalid. // Main thread only. Result mRunningState; // Shutdown blocker unique for this Session. Main thread only. RefPtr mShutdownBlocker; }; MediaRecorder::~MediaRecorder() { LOG(LogLevel::Debug, ("~MediaRecorder (%p)", this)); UnRegisterActivityObserver(); } MediaRecorder::MediaRecorder(nsPIDOMWindowInner* aOwnerWindow) : DOMEventTargetHelper(aOwnerWindow) { MOZ_ASSERT(aOwnerWindow); RegisterActivityObserver(); } void MediaRecorder::RegisterActivityObserver() { if (nsPIDOMWindowInner* window = GetOwner()) { mDocument = window->GetExtantDoc(); if (mDocument) { mDocument->RegisterActivityObserver( NS_ISUPPORTS_CAST(nsIDocumentActivity*, this)); } } } void MediaRecorder::UnRegisterActivityObserver() { if (mDocument) { mDocument->UnregisterActivityObserver( NS_ISUPPORTS_CAST(nsIDocumentActivity*, this)); } } void MediaRecorder::GetMimeType(nsString& aMimeType) { aMimeType = mMimeType; } void MediaRecorder::Start(const Optional& aTimeslice, ErrorResult& aResult) { LOG(LogLevel::Debug, ("MediaRecorder.Start %p", this)); InitializeDomExceptions(); // When a MediaRecorder object’s start() method is invoked, the UA MUST run // the following steps: // 1. Let recorder be the MediaRecorder object on which the method was // invoked. // 2. Let timeslice be the method’s first argument, if provided, or undefined. TimeDuration timeslice = aTimeslice.WasPassed() ? TimeDuration::FromMilliseconds(aTimeslice.Value()) : TimeDuration::Forever(); // 3. Let stream be the value of recorder’s stream attribute. // 4. Let tracks be the set of live tracks in stream’s track set. nsTArray> tracks; if (mStream) { mStream->GetTracks(tracks); } tracks.RemoveLastElements( tracks.end() - std::remove_if(tracks.begin(), tracks.end(), [](const auto& t) { return t->Ended(); })); // 5. If the value of recorder’s state attribute is not inactive, throw an // InvalidStateError DOMException and abort these steps. if (mState != RecordingState::Inactive) { aResult.ThrowInvalidStateError( "The MediaRecorder has already been started"); return; } // 6. If the isolation properties of stream disallow access from recorder, // throw a SecurityError DOMException and abort these steps. if (mStream) { RefPtr streamPrincipal = mStream->GetPrincipal(); if (!streamPrincipal) { // This is more or less part of the step 7, see below. aResult.ThrowNotSupportedError("The MediaStream is inactive"); return; } if (!PrincipalSubsumes(this, streamPrincipal)) { aResult.ThrowSecurityError( "The MediaStream's isolation properties disallow access from " "MediaRecorder"); return; } } if (mAudioNode && !AudioNodePrincipalSubsumes(this, mAudioNode)) { LOG(LogLevel::Warning, ("MediaRecorder %p Start AudioNode principal check failed", this)); aResult.ThrowSecurityError( "The AudioNode's isolation properties disallow access from " "MediaRecorder"); return; } // 7. If stream is inactive, throw a NotSupportedError DOMException and abort // these steps. if (mStream && !mStream->Active()) { aResult.ThrowNotSupportedError("The MediaStream is inactive"); return; } // 8. If the [[ConstrainedMimeType]] slot specifies a media type, container, // or codec, then run the following sub steps: // 1. Constrain the configuration of recorder to the media type, container, // and codec specified in the [[ConstrainedMimeType]] slot. // 2. For each track in tracks, if the User Agent cannot record the track // using the current configuration, then throw a NotSupportedError // DOMException and abort all steps. Maybe mime; if (mConstrainedMimeType.Length() > 0) { mime = MakeMediaContainerType(mConstrainedMimeType); MOZ_DIAGNOSTIC_ASSERT( mime, "Invalid media MIME type should have been caught by IsTypeSupported"); } for (const auto& track : tracks) { TypeSupport support = CanRecordWith(track, mime, mConstrainedMimeType); if (support != TypeSupport::Supported) { nsString id; track->GetId(id); aResult.ThrowNotSupportedError(nsPrintfCString( "%s track cannot be recorded: %s", track->AsAudioStreamTrack() ? "An audio" : "A video", TypeSupportToCString(support, mConstrainedMimeType).get())); return; } } if (mAudioNode) { TypeSupport support = CanRecordAudioTrackWith(mime, mConstrainedMimeType); if (support != TypeSupport::Supported) { aResult.ThrowNotSupportedError(nsPrintfCString( "An AudioNode cannot be recorded: %s", TypeSupportToCString(support, mConstrainedMimeType).get())); return; } } // 9. If recorder’s [[ConstrainedBitsPerSecond]] slot is not undefined, set // recorder’s videoBitsPerSecond and audioBitsPerSecond attributes to // values the User Agent deems reasonable for the respective media types, // for recording all tracks in tracks, such that the sum of // videoBitsPerSecond and audioBitsPerSecond is close to the value of // recorder’s // [[ConstrainedBitsPerSecond]] slot. uint8_t numVideoTracks = 0; uint8_t numAudioTracks = 0; for (const auto& t : tracks) { if (t->AsVideoStreamTrack() && numVideoTracks < UINT8_MAX) { ++numVideoTracks; } else if (t->AsAudioStreamTrack() && numAudioTracks < UINT8_MAX) { ++numAudioTracks; } } if (mAudioNode) { MOZ_DIAGNOSTIC_ASSERT(!mStream); ++numAudioTracks; } if (mConstrainedBitsPerSecond) { SelectBitrates(*mConstrainedBitsPerSecond, numVideoTracks, &mVideoBitsPerSecond, numAudioTracks, &mAudioBitsPerSecond); } // 10. Let videoBitrate be the value of recorder’s videoBitsPerSecond // attribute, and constrain the configuration of recorder to target an // aggregate bitrate of videoBitrate bits per second for all video tracks // recorder will be recording. videoBitrate is a hint for the encoder and // the value might be surpassed, not achieved, or only be achieved over a // long period of time. const uint32_t videoBitrate = mVideoBitsPerSecond; // 11. Let audioBitrate be the value of recorder’s audioBitsPerSecond // attribute, and constrain the configuration of recorder to target an // aggregate bitrate of audioBitrate bits per second for all audio tracks // recorder will be recording. audioBitrate is a hint for the encoder and // the value might be surpassed, not achieved, or only be achieved over a // long period of time. const uint32_t audioBitrate = mAudioBitsPerSecond; // 12. Constrain the configuration of recorder to encode using the BitrateMode // specified by the value of recorder’s audioBitrateMode attribute for all // audio tracks recorder will be recording. // -- NOT IMPLEMENTED // 13. For each track in tracks, if the User Agent cannot record the track // using the current configuration, then throw a NotSupportedError // DOMException and abort these steps. if (numVideoTracks > 1) { aResult.ThrowNotSupportedError( "MediaRecorder does not support recording more than one video track"_ns); return; } if (numAudioTracks > 1) { aResult.ThrowNotSupportedError( "MediaRecorder does not support recording more than one audio track"_ns); return; } // 14. Set recorder’s state to recording mState = RecordingState::Recording; MediaRecorderReporter::AddMediaRecorder(this); // Start a session. mSessions.AppendElement(); mSessions.LastElement() = new Session(this, std::move(tracks), videoBitrate, audioBitrate); mSessions.LastElement()->Start(timeslice); } void MediaRecorder::Stop(ErrorResult& aResult) { LOG(LogLevel::Debug, ("MediaRecorder.Stop %p", this)); MediaRecorderReporter::RemoveMediaRecorder(this); // When a MediaRecorder object’s stop() method is invoked, the UA MUST run the // following steps: // 1. Let recorder be the MediaRecorder object on which the method was // invoked. // 2. If recorder’s state attribute is inactive, abort these steps. if (mState == RecordingState::Inactive) { return; } // 3. Inactivate the recorder with recorder. Inactivate(); // 4. Queue a task, using the DOM manipulation task source, that runs the // following steps: // 1. Stop gathering data. // 2. Let blob be the Blob of collected data so far, then fire a blob event // named dataavailable at recorder with blob. // 3. Fire an event named stop at recorder. MOZ_ASSERT(mSessions.Length() > 0); mSessions.LastElement()->Stop(); // 5. return undefined. } void MediaRecorder::Pause(ErrorResult& aResult) { LOG(LogLevel::Debug, ("MediaRecorder.Pause %p", this)); // When a MediaRecorder object’s pause() method is invoked, the UA MUST run // the following steps: // 1. If state is inactive, throw an InvalidStateError DOMException and abort // these steps. if (mState == RecordingState::Inactive) { aResult.ThrowInvalidStateError("The MediaRecorder is inactive"); return; } // 2. If state is paused, abort these steps. if (mState == RecordingState::Paused) { return; } // 3. Set state to paused, and queue a task, using the DOM manipulation task // source, that runs the following steps: mState = RecordingState::Paused; // XXX - We pause synchronously pending spec issue // https://github.com/w3c/mediacapture-record/issues/131 // 1. Stop gathering data into blob (but keep it available so that // recording can be resumed in the future). MOZ_ASSERT(!mSessions.IsEmpty()); mSessions.LastElement()->Pause(); NS_DispatchToMainThread(NS_NewRunnableFunction( "MediaRecorder::Pause", [recorder = RefPtr(this)] { // 2. Let target be the MediaRecorder context object. Fire an event // named pause at target. recorder->DispatchSimpleEvent(u"pause"_ns); })); // 4. return undefined. } void MediaRecorder::Resume(ErrorResult& aResult) { LOG(LogLevel::Debug, ("MediaRecorder.Resume %p", this)); // When a MediaRecorder object’s resume() method is invoked, the UA MUST run // the following steps: // 1. If state is inactive, throw an InvalidStateError DOMException and abort // these steps. if (mState == RecordingState::Inactive) { aResult.ThrowInvalidStateError("The MediaRecorder is inactive"); return; } // 2. If state is recording, abort these steps. if (mState == RecordingState::Recording) { return; } // 3. Set state to recording, and queue a task, using the DOM manipulation // task source, that runs the following steps: mState = RecordingState::Recording; // XXX - We resume synchronously pending spec issue // https://github.com/w3c/mediacapture-record/issues/131 // 1. Resume (or continue) gathering data into the current blob. MOZ_ASSERT(!mSessions.IsEmpty()); mSessions.LastElement()->Resume(); NS_DispatchToMainThread(NS_NewRunnableFunction( "MediaRecorder::Resume", [recorder = RefPtr(this)] { // 2. Let target be the MediaRecorder context object. Fire an event // named resume at target. recorder->DispatchSimpleEvent(u"resume"_ns); })); // 4. return undefined. } void MediaRecorder::RequestData(ErrorResult& aResult) { LOG(LogLevel::Debug, ("MediaRecorder.RequestData %p", this)); // When a MediaRecorder object’s requestData() method is invoked, the UA MUST // run the following steps: // 1. If state is inactive throw an InvalidStateError DOMException and // terminate these steps. Otherwise the UA MUST queue a task, using the DOM // manipulation task source, that runs the following steps: // 1. Let blob be the Blob of collected data so far and let target be the // MediaRecorder context object, then fire a blob event named // dataavailable at target with blob. (Note that blob will be empty if no // data has been gathered yet.) // 2. Create a new Blob and gather subsequent data into it. if (mState == RecordingState::Inactive) { aResult.ThrowInvalidStateError("The MediaRecorder is inactive"); return; } MOZ_ASSERT(mSessions.Length() > 0); mSessions.LastElement()->RequestData(); // 2. return undefined. } JSObject* MediaRecorder::WrapObject(JSContext* aCx, JS::Handle aGivenProto) { return MediaRecorder_Binding::Wrap(aCx, this, aGivenProto); } /* static */ already_AddRefed MediaRecorder::Constructor( const GlobalObject& aGlobal, DOMMediaStream& aStream, const MediaRecorderOptions& aOptions, ErrorResult& aRv) { nsCOMPtr ownerWindow = do_QueryInterface(aGlobal.GetAsSupports()); if (!ownerWindow) { aRv.Throw(NS_ERROR_FAILURE); return nullptr; } // When the MediaRecorder() constructor is invoked, the User Agent MUST run // the following steps: // 1. Let stream be the constructor’s first argument. // 2. Let options be the constructor’s second argument. // 3. If invoking is type supported with options’ mimeType member as its // argument returns false, throw a NotSupportedError DOMException and abort // these steps. TypeSupport support = IsTypeSupportedImpl(aOptions.mMimeType); if (support != TypeSupport::Supported) { // This catches also the empty string mimeType when support for any encoders // has been disabled. aRv.ThrowNotSupportedError( TypeSupportToCString(support, aOptions.mMimeType)); return nullptr; } // 4. Let recorder be a newly constructed MediaRecorder object. RefPtr recorder = new MediaRecorder(ownerWindow); // 5. Let recorder have a [[ConstrainedMimeType]] internal slot, initialized // to the value of options' mimeType member. recorder->mConstrainedMimeType = aOptions.mMimeType; // 6. Let recorder have a [[ConstrainedBitsPerSecond]] internal slot, // initialized to the value of options’ bitsPerSecond member, if it is // present, otherwise undefined. recorder->mConstrainedBitsPerSecond = aOptions.mBitsPerSecond.WasPassed() ? Some(aOptions.mBitsPerSecond.Value()) : Nothing(); // 7. Initialize recorder’s stream attribute to stream. recorder->mStream = &aStream; // 8. Initialize recorder’s mimeType attribute to the value of recorder’s // [[ConstrainedMimeType]] slot. recorder->mMimeType = recorder->mConstrainedMimeType; // 9. Initialize recorder’s state attribute to inactive. recorder->mState = RecordingState::Inactive; // 10. Initialize recorder’s videoBitsPerSecond attribute to the value of // options’ videoBitsPerSecond member, if it is present. Otherwise, choose // a target value the User Agent deems reasonable for video. recorder->mVideoBitsPerSecond = aOptions.mVideoBitsPerSecond.WasPassed() ? aOptions.mVideoBitsPerSecond.Value() : DEFAULT_VIDEO_BITRATE_BPS; // 11. Initialize recorder’s audioBitsPerSecond attribute to the value of // options’ audioBitsPerSecond member, if it is present. Otherwise, choose // a target value the User Agent deems reasonable for audio. recorder->mAudioBitsPerSecond = aOptions.mAudioBitsPerSecond.WasPassed() ? aOptions.mAudioBitsPerSecond.Value() : DEFAULT_AUDIO_BITRATE_BPS; // 12. If recorder’s [[ConstrainedBitsPerSecond]] slot is not undefined, set // recorder’s videoBitsPerSecond and audioBitsPerSecond attributes to // values the User Agent deems reasonable for the respective media types, // such that the sum of videoBitsPerSecond and audioBitsPerSecond is close // to the value of recorder’s [[ConstrainedBitsPerSecond]] slot. if (recorder->mConstrainedBitsPerSecond) { SelectBitrates(*recorder->mConstrainedBitsPerSecond, 1, &recorder->mVideoBitsPerSecond, 1, &recorder->mAudioBitsPerSecond); } // 13. Return recorder. return recorder.forget(); } /* static */ already_AddRefed MediaRecorder::Constructor( const GlobalObject& aGlobal, AudioNode& aAudioNode, uint32_t aAudioNodeOutput, const MediaRecorderOptions& aOptions, ErrorResult& aRv) { // Allow recording from audio node only when pref is on. if (!Preferences::GetBool("media.recorder.audio_node.enabled", false)) { // Pretending that this constructor is not defined. aRv.ThrowTypeError("Argument 1", "MediaStream"); return nullptr; } nsCOMPtr ownerWindow = do_QueryInterface(aGlobal.GetAsSupports()); if (!ownerWindow) { aRv.Throw(NS_ERROR_FAILURE); return nullptr; } // aAudioNodeOutput doesn't matter to destination node because it has no // output. if (aAudioNode.NumberOfOutputs() > 0 && aAudioNodeOutput >= aAudioNode.NumberOfOutputs()) { aRv.ThrowIndexSizeError("Invalid AudioNode output index"); return nullptr; } // When the MediaRecorder() constructor is invoked, the User Agent MUST run // the following steps: // 1. Let stream be the constructor’s first argument. (we'll let audioNode be // the first arg, and audioNodeOutput the second) // 2. Let options be the constructor’s second argument. (we'll let options be // the third arg) // 3. If invoking is type supported with options’ mimeType member as its // argument returns false, throw a NotSupportedError DOMException and abort // these steps. TypeSupport support = IsTypeSupportedImpl(aOptions.mMimeType); if (support != TypeSupport::Supported) { // This catches also the empty string mimeType when support for any encoders // has been disabled. aRv.ThrowNotSupportedError( TypeSupportToCString(support, aOptions.mMimeType)); return nullptr; } // 4. Let recorder be a newly constructed MediaRecorder object. RefPtr recorder = new MediaRecorder(ownerWindow); // 5. Let recorder have a [[ConstrainedMimeType]] internal slot, initialized // to the value of options' mimeType member. recorder->mConstrainedMimeType = aOptions.mMimeType; // 6. Let recorder have a [[ConstrainedBitsPerSecond]] internal slot, // initialized to the value of options’ bitsPerSecond member, if it is // present, otherwise undefined. recorder->mConstrainedBitsPerSecond = aOptions.mBitsPerSecond.WasPassed() ? Some(aOptions.mBitsPerSecond.Value()) : Nothing(); // 7. Initialize recorder’s stream attribute to stream. (make that the // audioNode and audioNodeOutput equivalents) recorder->mAudioNode = &aAudioNode; recorder->mAudioNodeOutput = aAudioNodeOutput; // 8. Initialize recorder’s mimeType attribute to the value of recorder’s // [[ConstrainedMimeType]] slot. recorder->mMimeType = recorder->mConstrainedMimeType; // 9. Initialize recorder’s state attribute to inactive. recorder->mState = RecordingState::Inactive; // 10. Initialize recorder’s videoBitsPerSecond attribute to the value of // options’ videoBitsPerSecond member, if it is present. Otherwise, choose // a target value the User Agent deems reasonable for video. recorder->mVideoBitsPerSecond = aOptions.mVideoBitsPerSecond.WasPassed() ? aOptions.mVideoBitsPerSecond.Value() : DEFAULT_VIDEO_BITRATE_BPS; // 11. Initialize recorder’s audioBitsPerSecond attribute to the value of // options’ audioBitsPerSecond member, if it is present. Otherwise, choose // a target value the User Agent deems reasonable for audio. recorder->mAudioBitsPerSecond = aOptions.mAudioBitsPerSecond.WasPassed() ? aOptions.mAudioBitsPerSecond.Value() : DEFAULT_AUDIO_BITRATE_BPS; // 12. If recorder’s [[ConstrainedBitsPerSecond]] slot is not undefined, set // recorder’s videoBitsPerSecond and audioBitsPerSecond attributes to // values the User Agent deems reasonable for the respective media types, // such that the sum of videoBitsPerSecond and audioBitsPerSecond is close // to the value of recorder’s [[ConstrainedBitsPerSecond]] slot. if (recorder->mConstrainedBitsPerSecond) { SelectBitrates(*recorder->mConstrainedBitsPerSecond, 1, &recorder->mVideoBitsPerSecond, 1, &recorder->mAudioBitsPerSecond); } // 13. Return recorder. return recorder.forget(); } /* static */ bool MediaRecorder::IsTypeSupported(GlobalObject& aGlobal, const nsAString& aMIMEType) { return MediaRecorder::IsTypeSupported(aMIMEType); } /* static */ bool MediaRecorder::IsTypeSupported(const nsAString& aMIMEType) { return IsTypeSupportedImpl(aMIMEType) == TypeSupport::Supported; } nsresult MediaRecorder::CreateAndDispatchBlobEvent(BlobImpl* aBlobImpl) { MOZ_ASSERT(NS_IsMainThread(), "Not running on main thread"); if (!GetOwnerGlobal()) { // This MediaRecorder has been disconnected in the meantime. return NS_ERROR_FAILURE; } RefPtr blob = Blob::Create(GetOwnerGlobal(), aBlobImpl); if (NS_WARN_IF(!blob)) { return NS_ERROR_FAILURE; } BlobEventInit init; init.mBubbles = false; init.mCancelable = false; init.mData = blob; RefPtr event = BlobEvent::Constructor(this, u"dataavailable"_ns, init); event->SetTrusted(true); ErrorResult rv; DispatchEvent(*event, rv); return rv.StealNSResult(); } void MediaRecorder::DispatchSimpleEvent(const nsAString& aStr) { MOZ_ASSERT(NS_IsMainThread(), "Not running on main thread"); nsresult rv = CheckCurrentGlobalCorrectness(); if (NS_FAILED(rv)) { return; } rv = DOMEventTargetHelper::DispatchTrustedEvent(aStr); if (NS_FAILED(rv)) { LOG(LogLevel::Error, ("MediaRecorder.DispatchSimpleEvent: DispatchTrustedEvent failed %p", this)); NS_ERROR("Failed to dispatch the event!!!"); } } void MediaRecorder::NotifyError(nsresult aRv) { MOZ_ASSERT(NS_IsMainThread(), "Not running on main thread"); nsresult rv = CheckCurrentGlobalCorrectness(); if (NS_FAILED(rv)) { return; } MediaRecorderErrorEventInit init; init.mBubbles = false; init.mCancelable = false; // These DOMExceptions have been created earlier so they can contain stack // traces. We attach the appropriate one here to be fired. We should have // exceptions here, but defensively check. switch (aRv) { case NS_ERROR_DOM_SECURITY_ERR: if (!mSecurityDomException) { LOG(LogLevel::Debug, ("MediaRecorder.NotifyError: " "mSecurityDomException was not initialized")); mSecurityDomException = DOMException::Create(NS_ERROR_DOM_SECURITY_ERR); } init.mError = std::move(mSecurityDomException); break; default: if (mOtherDomException && aRv == mOtherDomException->GetResult()) { LOG(LogLevel::Debug, ("MediaRecorder.NotifyError: " "mOtherDomException being fired for aRv: %X", uint32_t(aRv))); init.mError = std::move(mOtherDomException); break; } if (!mUnknownDomException) { LOG(LogLevel::Debug, ("MediaRecorder.NotifyError: " "mUnknownDomException was not initialized")); mUnknownDomException = DOMException::Create(NS_ERROR_DOM_UNKNOWN_ERR); } LOG(LogLevel::Debug, ("MediaRecorder.NotifyError: " "mUnknownDomException being fired for aRv: %X", uint32_t(aRv))); init.mError = std::move(mUnknownDomException); break; } RefPtr event = MediaRecorderErrorEvent::Constructor(this, u"error"_ns, init); event->SetTrusted(true); IgnoredErrorResult res; DispatchEvent(*event, res); if (res.Failed()) { NS_ERROR("Failed to dispatch the error event!!!"); } } void MediaRecorder::RemoveSession(Session* aSession) { LOG(LogLevel::Debug, ("MediaRecorder.RemoveSession (%p)", aSession)); mSessions.RemoveElement(aSession); } void MediaRecorder::NotifyOwnerDocumentActivityChanged() { nsPIDOMWindowInner* window = GetOwner(); NS_ENSURE_TRUE_VOID(window); Document* doc = window->GetExtantDoc(); NS_ENSURE_TRUE_VOID(doc); LOG(LogLevel::Debug, ("MediaRecorder %p NotifyOwnerDocumentActivityChanged " "IsActive=%d, " "IsVisible=%d, ", this, doc->IsActive(), doc->IsVisible())); if (!doc->IsActive() || !doc->IsVisible()) { // Stop the session. ErrorResult result; Stop(result); result.SuppressException(); } } void MediaRecorder::Inactivate() { LOG(LogLevel::Debug, ("MediaRecorder.Inactivate %p", this)); // The Inactivate the recorder algorithm given a recorder, is as follows: // 1. Set recorder’s mimeType attribute to the value of the // [[ConstrainedMimeType]] slot. mMimeType = mConstrainedMimeType; // 2. Set recorder’s state attribute to inactive. mState = RecordingState::Inactive; // 3. If recorder’s [[ConstrainedBitsPerSecond]] slot is not undefined, set // recorder’s videoBitsPerSecond and audioBitsPerSecond attributes to // values the User Agent deems reasonable for the respective media types, // such that the sum of videoBitsPerSecond and audioBitsPerSecond is close // to the value of recorder’s [[ConstrainedBitsPerSecond]] slot. if (mConstrainedBitsPerSecond) { SelectBitrates(*mConstrainedBitsPerSecond, 1, &mVideoBitsPerSecond, 1, &mAudioBitsPerSecond); } } void MediaRecorder::InitializeDomExceptions() { mSecurityDomException = DOMException::Create(NS_ERROR_DOM_SECURITY_ERR); mUnknownDomException = DOMException::Create(NS_ERROR_DOM_UNKNOWN_ERR); } RefPtr MediaRecorder::SizeOfExcludingThis( mozilla::MallocSizeOf aMallocSizeOf) { MOZ_ASSERT(NS_IsMainThread()); // The return type of a chained MozPromise cannot be changed, so we create a // holder for our desired return type and resolve that from All()->Then(). auto holder = MakeRefPtr>>(); RefPtr promise = holder->Ensure(__func__); nsTArray> promises(mSessions.Length()); for (const RefPtr& session : mSessions) { promises.AppendElement(session->SizeOfExcludingThis(aMallocSizeOf)); } SizeOfPromise::All(GetCurrentSerialEventTarget(), promises) ->Then( GetCurrentSerialEventTarget(), __func__, [holder](const nsTArray& sizes) { size_t total = 0; for (const size_t& size : sizes) { total += size; } holder->Resolve(total, __func__); }, []() { MOZ_CRASH("Unexpected reject"); }); return promise; } StaticRefPtr MediaRecorderReporter::sUniqueInstance; } // namespace mozilla::dom #undef LOG