/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 2 -*- */ /* vim: set ts=8 sts=2 et sw=2 tw=80: */ /* This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ #include "mozilla/layers/TextureClient.h" #include // for uint8_t, uint32_t, etc #include "BufferTexture.h" #include "IPDLActor.h" #include "ImageContainer.h" // for PlanarYCbCrData, etc #include "MainThreadUtils.h" #include "gfx2DGlue.h" #include "gfxPlatform.h" // for gfxPlatform #include "gfxUtils.h" // for gfxUtils::GetAsLZ4Base64Str #include "mozilla/Atomics.h" #include "mozilla/Mutex.h" #include "mozilla/ProfilerLabels.h" #include "mozilla/SchedulerGroup.h" #include "mozilla/StaticPrefs_gfx.h" #include "mozilla/StaticPrefs_layers.h" #include "mozilla/gfx/2D.h" #include "mozilla/gfx/DataSurfaceHelpers.h" // for CreateDataSourceSurfaceByCloning #include "mozilla/gfx/Logging.h" // for gfxDebug #include "mozilla/gfx/gfxVars.h" #include "mozilla/ipc/CrossProcessSemaphore.h" #include "mozilla/ipc/SharedMemory.h" // for SharedMemory, etc #include "mozilla/layers/CompositableForwarder.h" #include "mozilla/layers/ISurfaceAllocator.h" #include "mozilla/layers/ImageBridgeChild.h" #include "mozilla/layers/ImageDataSerializer.h" #include "mozilla/layers/PTextureChild.h" #include "mozilla/layers/SynchronousTask.h" #include "mozilla/layers/TextureClientOGL.h" #include "mozilla/layers/TextureClientRecycleAllocator.h" #include "mozilla/layers/TextureRecorded.h" #include "nsDebug.h" // for NS_ASSERTION, NS_WARNING, etc #include "nsISerialEventTarget.h" #include "nsISupportsImpl.h" // for MOZ_COUNT_CTOR, etc #include "nsPrintfCString.h" // for nsPrintfCString #ifdef XP_WIN # include "gfx2DGlue.h" # include "gfxWindowsPlatform.h" # include "mozilla/gfx/DeviceManagerDx.h" # include "mozilla/layers/TextureD3D11.h" #endif #ifdef MOZ_WIDGET_GTK # include # include "gfxPlatformGtk.h" # include "mozilla/layers/DMABUFTextureClientOGL.h" # include "mozilla/widget/DMABufLibWrapper.h" #endif #ifdef MOZ_WAYLAND # include "mozilla/widget/nsWaylandDisplay.h" #endif #ifdef XP_MACOSX # include "mozilla/layers/MacIOSurfaceTextureClientOGL.h" #endif #if 0 # define RECYCLE_LOG(...) printf_stderr(__VA_ARGS__) #else # define RECYCLE_LOG(...) \ do { \ } while (0) #endif namespace mozilla::layers { using namespace mozilla::ipc; using namespace mozilla::gl; using namespace mozilla::gfx; struct TextureDeallocParams { TextureData* data = nullptr; RefPtr actor; RefPtr readLock; RefPtr allocator; bool clientDeallocation = false; bool syncDeallocation = false; TextureDeallocParams() = default; TextureDeallocParams(const TextureDeallocParams&) = delete; TextureDeallocParams& operator=(const TextureDeallocParams&) = delete; TextureDeallocParams(TextureDeallocParams&& aOther) : data(aOther.data), actor(std::move(aOther.actor)), readLock(std::move(aOther.readLock)), allocator(std::move(aOther.allocator)), clientDeallocation(aOther.clientDeallocation), syncDeallocation(aOther.syncDeallocation) { aOther.data = nullptr; } TextureDeallocParams& operator=(TextureDeallocParams&& aOther) { data = aOther.data; aOther.data = nullptr; actor = std::move(aOther.actor); readLock = std::move(aOther.readLock); allocator = std::move(aOther.allocator); clientDeallocation = aOther.clientDeallocation; syncDeallocation = aOther.syncDeallocation; return *this; } }; void DeallocateTextureClient(TextureDeallocParams& params); /** * TextureChild is the content-side incarnation of the PTexture IPDL actor. * * TextureChild is used to synchronize a texture client and its corresponding * TextureHost if needed (a TextureClient that is not shared with the compositor * does not have a TextureChild) * * During the deallocation phase, a TextureChild may hold its recently destroyed * TextureClient's data until the compositor side confirmed that it is safe to * deallocte or recycle the it. */ class TextureChild final : PTextureChild { ~TextureChild() { // We should have deallocated mTextureData in ActorDestroy MOZ_ASSERT(!mTextureData); MOZ_ASSERT_IF(!mOwnerCalledDestroy, !mTextureClient); } public: NS_INLINE_DECL_THREADSAFE_REFCOUNTING(TextureChild) TextureChild() : mCompositableForwarder(nullptr), mTextureForwarder(nullptr), mTextureClient(nullptr), mTextureData(nullptr), mDestroyed(false), mIPCOpen(false), mOwnsTextureData(false), mOwnerCalledDestroy(false), mUsesImageBridge(false) {} mozilla::ipc::IPCResult Recv__delete__() override { return IPC_OK(); } LayersIPCChannel* GetAllocator() { return mTextureForwarder; } void ActorDestroy(ActorDestroyReason why) override; bool IPCOpen() const { return mIPCOpen; } void Lock() const { if (mUsesImageBridge) { mLock.Enter(); } } void Unlock() const { if (mUsesImageBridge) { mLock.Leave(); } } private: // AddIPDLReference and ReleaseIPDLReference are only to be called by // CreateIPDLActor and DestroyIPDLActor, respectively. We intentionally make // them private to prevent misuse. The purpose of these methods is to be aware // of when the IPC system around this actor goes down: mIPCOpen is then set to // false. void AddIPDLReference() { MOZ_ASSERT(mIPCOpen == false); mIPCOpen = true; AddRef(); } void ReleaseIPDLReference() { MOZ_ASSERT(mIPCOpen == false); Release(); } /// The normal way to destroy the actor. /// /// This will asynchronously send a Destroy message to the parent actor, whom /// will send the delete message. void Destroy(const TextureDeallocParams& aParams); // This lock is used order to prevent several threads to access the // TextureClient's data concurrently. In particular, it prevents shutdown // code to destroy a texture while another thread is reading or writing into // it. // In most places, the lock is held in short and bounded scopes in which we // don't block on any other resource. There are few exceptions to this, which // are discussed below. // // The locking pattern of TextureClient may in some case upset deadlock // detection tools such as TSan. Typically our tile rendering code will lock // all of its tiles, render into them and unlock them all right after that, // which looks something like: // // Lock tile A // Lock tile B // Lock tile C // Apply drawing commands to tiles A, B and C // Unlock tile A // Unlock tile B // Unlock tile C // // And later, we may end up rendering a tile buffer that has the same tiles, // in a different order, for example: // // Lock tile B // Lock tile A // Lock tile D // Apply drawing commands to tiles A, B and D // Unlock tile B // Unlock tile A // Unlock tile D // // This is because textures being expensive to create, we recycle them as much // as possible and they may reappear in the tile buffer in a different order. // // Unfortunately this is not very friendly to TSan's analysis, which will see // that B was once locked while A was locked, and then A locked while B was // locked. TSan identifies this as a potential dead-lock which would be the // case if this kind of inconsistent and dependent locking order was happening // concurrently. // In the case of TextureClient, dependent locking only ever happens on the // thread that draws into the texture (let's call it the producer thread). // Other threads may call into a method that can lock the texture in a short // and bounded scope inside of which it is not allowed to do anything that // could cause the thread to block. A given texture can only have one producer // thread. // // Another example of TSan-unfriendly locking pattern is when copying a // texture into another, which also never happens outside of the producer // thread. Copying A into B looks like this: // // Lock texture B // Lock texture A // Copy A into B // Unlock A // Unlock B // // In a given frame we may need to copy A into B and in another frame copy // B into A. For example A and B can be the Front and Back buffers, // alternating roles and the copy is needed to avoid the cost of re-drawing // the valid region. // // The important rule is that all of the dependent locking must occur only // in the texture's producer thread to avoid deadlocks. mutable gfx::CriticalSection mLock; RefPtr mCompositableForwarder; RefPtr mTextureForwarder; TextureClient* mTextureClient; TextureData* mTextureData; Atomic mDestroyed; bool mIPCOpen; bool mOwnsTextureData; bool mOwnerCalledDestroy; bool mUsesImageBridge; friend class TextureClient; friend void DeallocateTextureClient(TextureDeallocParams& params); }; static inline gfx::BackendType BackendTypeForBackendSelector( LayersBackend aLayersBackend, BackendSelector aSelector) { switch (aSelector) { case BackendSelector::Canvas: return gfxPlatform::GetPlatform()->GetPreferredCanvasBackend(); case BackendSelector::Content: return gfxPlatform::GetPlatform()->GetContentBackendFor(aLayersBackend); default: MOZ_ASSERT_UNREACHABLE("Unknown backend selector"); return gfx::BackendType::NONE; } }; static TextureType ChooseTextureType(gfx::SurfaceFormat aFormat, gfx::IntSize aSize, KnowsCompositor* aKnowsCompositor, BackendSelector aSelector, TextureAllocationFlags aAllocFlags) { LayersBackend layersBackend = aKnowsCompositor->GetCompositorBackendType(); gfx::BackendType moz2DBackend = BackendTypeForBackendSelector(layersBackend, aSelector); Unused << moz2DBackend; #ifdef XP_WIN int32_t maxTextureSize = aKnowsCompositor->GetMaxTextureSize(); if ((layersBackend == LayersBackend::LAYERS_WR && !aKnowsCompositor->UsingSoftwareWebRender()) && (moz2DBackend == gfx::BackendType::DIRECT2D || moz2DBackend == gfx::BackendType::DIRECT2D1_1) && aSize.width <= maxTextureSize && aSize.height <= maxTextureSize && !(aAllocFlags & (ALLOC_UPDATE_FROM_SURFACE | ALLOC_DO_NOT_ACCELERATE))) { return TextureType::D3D11; } #endif #ifdef MOZ_WIDGET_GTK if ((layersBackend == LayersBackend::LAYERS_WR && !aKnowsCompositor->UsingSoftwareWebRender()) && widget::DMABufDevice::IsDMABufTexturesEnabled() && aFormat != SurfaceFormat::A8) { return TextureType::DMABUF; } #endif #ifdef XP_MACOSX if (StaticPrefs::gfx_use_iosurface_textures_AtStartup()) { return TextureType::MacIOSurface; } #endif #ifdef MOZ_WIDGET_ANDROID if (StaticPrefs::gfx_use_surfacetexture_textures_AtStartup()) { return TextureType::AndroidNativeWindow; } #endif return TextureType::Unknown; } TextureType PreferredCanvasTextureType(KnowsCompositor* aKnowsCompositor) { return ChooseTextureType(gfx::SurfaceFormat::R8G8B8A8, {1, 1}, aKnowsCompositor, BackendSelector::Canvas, TextureAllocationFlags::ALLOC_DEFAULT); } /* static */ TextureData* TextureData::Create(TextureType aTextureType, gfx::SurfaceFormat aFormat, const gfx::IntSize& aSize, TextureAllocationFlags aAllocFlags, gfx::BackendType aBackendType) { switch (aTextureType) { #ifdef XP_WIN case TextureType::D3D11: return D3D11TextureData::Create(aSize, aFormat, aAllocFlags); #endif #ifdef MOZ_WIDGET_GTK case TextureType::DMABUF: return DMABUFTextureData::Create(aSize, aFormat, aBackendType); #endif #ifdef XP_MACOSX case TextureType::MacIOSurface: return MacIOSurfaceTextureData::Create(aSize, aFormat, aBackendType); #endif #ifdef MOZ_WIDGET_ANDROID case TextureType::AndroidNativeWindow: return AndroidNativeWindowTextureData::Create(aSize, aFormat); #endif default: return nullptr; } } /* static */ TextureData* TextureData::Create(TextureForwarder* aAllocator, gfx::SurfaceFormat aFormat, gfx::IntSize aSize, KnowsCompositor* aKnowsCompositor, BackendSelector aSelector, TextureFlags aTextureFlags, TextureAllocationFlags aAllocFlags) { TextureType textureType = ChooseTextureType(aFormat, aSize, aKnowsCompositor, aSelector, aAllocFlags); if (aAllocFlags & ALLOC_FORCE_REMOTE) { RefPtr canvasChild = aAllocator->GetCanvasChild(); if (canvasChild) { return new RecordedTextureData(canvasChild.forget(), aSize, aFormat, textureType); } // If we must be remote, but there is no canvas child, then falling back // is not possible. return nullptr; } gfx::BackendType moz2DBackend = gfx::BackendType::NONE; #if defined(XP_MACOSX) || defined(MOZ_WIDGET_GTK) moz2DBackend = BackendTypeForBackendSelector( aKnowsCompositor->GetCompositorBackendType(), aSelector); #endif return TextureData::Create(textureType, aFormat, aSize, aAllocFlags, moz2DBackend); } /* static */ bool TextureData::IsRemote(KnowsCompositor* aKnowsCompositor, BackendSelector aSelector, gfx::SurfaceFormat aFormat, gfx::IntSize aSize) { if (aSelector != BackendSelector::Canvas || !gfxPlatform::UseRemoteCanvas()) { return false; } TextureType textureType = ChooseTextureType(aFormat, aSize, aKnowsCompositor, aSelector, TextureAllocationFlags::ALLOC_DEFAULT); switch (textureType) { case TextureType::D3D11: return true; default: return false; } } static void DestroyTextureData(TextureData* aTextureData, LayersIPCChannel* aAllocator, bool aDeallocate) { if (!aTextureData) { return; } if (aDeallocate) { aTextureData->Deallocate(aAllocator); } else { aTextureData->Forget(aAllocator); } delete aTextureData; } void TextureChild::ActorDestroy(ActorDestroyReason why) { AUTO_PROFILER_LABEL("TextureChild::ActorDestroy", GRAPHICS); MOZ_ASSERT(mIPCOpen); mIPCOpen = false; if (mTextureData) { DestroyTextureData(mTextureData, GetAllocator(), mOwnsTextureData); mTextureData = nullptr; } } void TextureChild::Destroy(const TextureDeallocParams& aParams) { MOZ_ASSERT(!mOwnerCalledDestroy); if (mOwnerCalledDestroy) { return; } mOwnerCalledDestroy = true; if (!IPCOpen()) { DestroyTextureData(aParams.data, aParams.allocator, aParams.clientDeallocation); return; } // DestroyTextureData will be called by TextureChild::ActorDestroy mTextureData = aParams.data; mOwnsTextureData = aParams.clientDeallocation; if (!mCompositableForwarder || !mCompositableForwarder->DestroyInTransaction(this)) { this->SendDestroy(); } } /* static */ Atomic TextureClient::sSerialCounter(0); /// The logic for synchronizing a TextureClient's deallocation goes here. /// /// This funciton takes care of dispatching work to the right thread using /// a synchronous proxy if needed, and handles client/host deallocation. void DeallocateTextureClient(TextureDeallocParams& params) { if (!params.actor && !params.readLock && !params.data) { // Nothing to do return; } TextureChild* actor = params.actor; nsCOMPtr ipdlThread; if (params.allocator) { ipdlThread = params.allocator->GetThread(); if (!ipdlThread) { // An allocator with no thread means we are too late in the shutdown // sequence. gfxCriticalError() << "Texture deallocated too late during shutdown"; return; } } // First make sure that the work is happening on the IPDL thread. if (ipdlThread && !ipdlThread->IsOnCurrentThread()) { if (params.syncDeallocation) { bool done = false; ReentrantMonitor barrier MOZ_UNANNOTATED("DeallocateTextureClient"); ReentrantMonitorAutoEnter autoMon(barrier); ipdlThread->Dispatch(NS_NewRunnableFunction( "DeallocateTextureClientSyncProxyRunnable", [&]() { DeallocateTextureClient(params); ReentrantMonitorAutoEnter autoMonInner(barrier); done = true; barrier.NotifyAll(); })); while (!done) { barrier.Wait(); } } else { ipdlThread->Dispatch( NS_NewRunnableFunction("DeallocateTextureClientRunnable", [params = std::move(params)]() mutable { DeallocateTextureClient(params); })); } // The work has been forwarded to the IPDL thread, we are done. return; } // Below this line, we are either in the IPDL thread or ther is no IPDL // thread anymore. if (!ipdlThread) { // If we don't have a thread we can't know for sure that we are in // the IPDL thread and use the LayersIPCChannel. // This should ideally not happen outside of gtest, but some shutdown // raciness could put us in this situation. params.allocator = nullptr; } if (params.readLock) { // This should be the last reference to the object, which will destroy it. params.readLock = nullptr; } if (!actor) { // We don't have an IPDL actor, probably because we destroyed the // TextureClient before sharing it with the compositor. It means the data // cannot be owned by the TextureHost since we never created the // TextureHost... DestroyTextureData(params.data, params.allocator, /* aDeallocate */ true); return; } actor->Destroy(params); } void TextureClient::Destroy() { // Async paints should have been flushed by now. MOZ_RELEASE_ASSERT(mPaintThreadRefs == 0); if (mActor && !mIsLocked) { mActor->Lock(); } mBorrowedDrawTarget = nullptr; mBorrowedSnapshot = false; RefPtr actor = std::move(mActor); RefPtr readLock; { MutexAutoLock lock(mMutex); readLock = std::move(mReadLock); } if (actor && !actor->mDestroyed.compareExchange(false, true)) { actor->Unlock(); actor = nullptr; } TextureData* data = mData; mData = nullptr; if (data || actor || readLock) { TextureDeallocParams params; params.actor = std::move(actor); params.readLock = std::move(readLock); params.allocator = mAllocator; params.clientDeallocation = !!(mFlags & TextureFlags::DEALLOCATE_CLIENT); params.data = data; // At the moment we always deallocate synchronously when deallocating on the // client side, but having asynchronous deallocate in some of the cases will // be a worthwhile optimization. params.syncDeallocation = !!(mFlags & TextureFlags::DEALLOCATE_CLIENT); // Release the lock before calling DeallocateTextureClient because the // latter may wait for the main thread which could create a dead-lock. if (params.actor) { params.actor->Unlock(); } DeallocateTextureClient(params); } } void TextureClient::LockActor() const { if (mActor) { mActor->Lock(); } } void TextureClient::UnlockActor() const { if (mActor) { mActor->Unlock(); } } void TextureClient::EnsureHasReadLock() { if (mFlags & TextureFlags::NON_BLOCKING_READ_LOCK) { MOZ_ASSERT(!(mFlags & TextureFlags::BLOCKING_READ_LOCK)); EnableReadLock(); } else if (mFlags & TextureFlags::BLOCKING_READ_LOCK) { MOZ_ASSERT(!(mFlags & TextureFlags::NON_BLOCKING_READ_LOCK)); EnableBlockingReadLock(); } } bool TextureClient::IsReadLocked() { if (!ShouldReadLock()) { return false; } nsCOMPtr thread; { MutexAutoLock lock(mMutex); if (mReadLock) { MOZ_ASSERT(mReadLock->AsNonBlockingLock(), "Can only check locked for non-blocking locks!"); return mReadLock->AsNonBlockingLock()->GetReadCount() > 1; } thread = mAllocator->GetThread(); if (!thread) { // We must be in the process of shutting down. return false; } if (thread->IsOnCurrentThread()) { EnsureHasReadLock(); if (NS_WARN_IF(!mReadLock)) { MOZ_ASSERT(!mAllocator->IPCOpen()); return false; } MOZ_ASSERT(mReadLock->AsNonBlockingLock(), "Can only check locked for non-blocking locks!"); return mReadLock->AsNonBlockingLock()->GetReadCount() > 1; } } MOZ_ASSERT(mAllocator->UsesImageBridge()); bool result = false; SynchronousTask task("TextureClient::IsReadLocked"); thread->Dispatch(NS_NewRunnableFunction("TextureClient::IsReadLocked", [&]() { AutoCompleteTask complete(&task); result = IsReadLocked(); })); task.Wait(); return result; } bool TextureClient::TryReadLock() { if (!ShouldReadLock()) { return true; } nsCOMPtr thread; { MutexAutoLock lock(mMutex); if (mIsReadLocked) { return true; } if (mReadLock) { if (mReadLock->AsNonBlockingLock() && mReadLock->AsNonBlockingLock()->GetReadCount() > 1) { return false; } if (!mReadLock->TryReadLock(TimeDuration::FromMilliseconds(500))) { return false; } mIsReadLocked = true; return true; } thread = mAllocator->GetThread(); if (!thread) { // We must be in the process of shutting down. return false; } if (thread->IsOnCurrentThread()) { EnsureHasReadLock(); if (NS_WARN_IF(!mReadLock)) { MOZ_ASSERT(!mAllocator->IPCOpen()); return false; } if (mReadLock->AsNonBlockingLock() && mReadLock->AsNonBlockingLock()->GetReadCount() > 1) { return false; } if (!mReadLock->TryReadLock(TimeDuration::FromMilliseconds(500))) { return false; } mIsReadLocked = true; return true; } } MOZ_ASSERT(mAllocator->UsesImageBridge()); bool result = false; SynchronousTask task("TextureClient::TryReadLock"); thread->Dispatch(NS_NewRunnableFunction("TextureClient::TryReadLock", [&]() { AutoCompleteTask complete(&task); result = TryReadLock(); })); task.Wait(); return result; } void TextureClient::ReadUnlock() { if (!ShouldReadLock()) { return; } MutexAutoLock lock(mMutex); if (!mIsReadLocked) { return; } MOZ_ASSERT(mReadLock); mReadLock->ReadUnlock(); mIsReadLocked = false; } bool TextureClient::Lock(OpenMode aMode) { MOZ_ASSERT(IsValid()); MOZ_ASSERT(!mIsLocked); if (!IsValid()) { return false; } if (mIsLocked) { return mOpenMode == aMode; } if ((aMode & OpenMode::OPEN_WRITE || !mInfo.canConcurrentlyReadLock) && !TryReadLock()) { // Only warn if attempting to write. Attempting to read is acceptable usage. if (aMode & OpenMode::OPEN_WRITE) { NS_WARNING( "Attempt to Lock a texture that is being read by the compositor!"); } return false; } LockActor(); mIsLocked = mData->Lock(aMode); mOpenMode = aMode; auto format = GetFormat(); if (mIsLocked && CanExposeDrawTarget() && (aMode & OpenMode::OPEN_READ_WRITE) == OpenMode::OPEN_READ_WRITE && NS_IsMainThread() && // the formats that we apparently expect, in the cairo backend. Any other // format will trigger an assertion in GfxFormatToCairoFormat. (format == SurfaceFormat::A8R8G8B8_UINT32 || format == SurfaceFormat::X8R8G8B8_UINT32 || format == SurfaceFormat::A8 || format == SurfaceFormat::R5G6B5_UINT16)) { if (!BorrowDrawTarget()) { // Failed to get a DrawTarget, means we won't be able to write into the // texture, might as well fail now. Unlock(); return false; } } if (!mIsLocked) { UnlockActor(); ReadUnlock(); } return mIsLocked; } void TextureClient::Unlock() { MOZ_ASSERT(IsValid()); MOZ_ASSERT(mIsLocked); if (!IsValid() || !mIsLocked) { return; } if (mBorrowedDrawTarget) { if (!(mOpenMode & OpenMode::OPEN_ASYNC)) { if (mOpenMode & OpenMode::OPEN_WRITE) { mBorrowedDrawTarget->Flush(); } mBorrowedDrawTarget->DetachAllSnapshots(); // If this assertion is hit, it means something is holding a strong // reference to our DrawTarget externally, which is not allowed. MOZ_ASSERT(mBorrowedDrawTarget->refCount() <= mExpectedDtRefs); } mBorrowedDrawTarget = nullptr; } mBorrowedSnapshot = false; if (mOpenMode & OpenMode::OPEN_WRITE) { mUpdated = true; } if (mData) { mData->Unlock(); } mIsLocked = false; mOpenMode = OpenMode::OPEN_NONE; UnlockActor(); ReadUnlock(); } void TextureClient::EnableReadLock() { MOZ_ASSERT(ShouldReadLock()); if (!mReadLock && mAllocator->GetTileLockAllocator()) { mReadLock = NonBlockingTextureReadLock::Create(mAllocator); } } void TextureClient::OnPrepareForwardToHost() { if (!ShouldReadLock()) { return; } MutexAutoLock lock(mMutex); if (NS_WARN_IF(!mReadLock)) { MOZ_ASSERT(!mAllocator->IPCOpen(), "Should have created readlock already!"); MOZ_ASSERT(!mIsPendingForwardReadLocked); return; } if (mIsPendingForwardReadLocked) { return; } mReadLock->ReadLock(); mIsPendingForwardReadLocked = true; } void TextureClient::OnAbandonForwardToHost() { if (!ShouldReadLock()) { return; } MutexAutoLock lock(mMutex); if (!mReadLock || !mIsPendingForwardReadLocked) { return; } mReadLock->ReadUnlock(); mIsPendingForwardReadLocked = false; } bool TextureClient::OnForwardedToHost() { if (mData) { mData->OnForwardedToHost(); } if (!ShouldReadLock()) { return false; } MutexAutoLock lock(mMutex); EnsureHasReadLock(); if (NS_WARN_IF(!mReadLock)) { MOZ_ASSERT(!mAllocator->IPCOpen()); return false; } if (!mUpdated) { if (mIsPendingForwardReadLocked) { mIsPendingForwardReadLocked = false; mReadLock->ReadUnlock(); } return false; } mUpdated = false; if (mIsPendingForwardReadLocked) { // We have successfully forwarded, just clear the flag and let the // TextureHost be responsible for unlocking. mIsPendingForwardReadLocked = false; } else { // Otherwise we did not need to readlock in advance, so do so now. We do // this on behalf of the TextureHost. mReadLock->ReadLock(); } return true; } TextureClient::~TextureClient() { // TextureClients should be kept alive while there are references on the // paint thread. MOZ_ASSERT(mPaintThreadRefs == 0); mReadLock = nullptr; Destroy(); } void TextureClient::UpdateFromSurface(gfx::SourceSurface* aSurface) { MOZ_ASSERT(IsValid()); MOZ_ASSERT(mIsLocked); MOZ_ASSERT(aSurface); // If you run into this assertion, make sure the texture was locked write-only // rather than read-write. MOZ_ASSERT(!mBorrowedDrawTarget); // XXX - It would be better to first try the DrawTarget approach and fallback // to the backend-specific implementation because the latter will usually do // an expensive read-back + cpu-side copy if the texture is on the gpu. // There is a bug with the DrawTarget approach, though specific to reading // back from WebGL (where R and B channel end up inverted) to figure out // first. if (mData->UpdateFromSurface(aSurface)) { return; } if (CanExposeDrawTarget() && NS_IsMainThread()) { RefPtr dt = BorrowDrawTarget(); MOZ_ASSERT(dt); if (dt) { dt->CopySurface(aSurface, gfx::IntRect(gfx::IntPoint(0, 0), aSurface->GetSize()), gfx::IntPoint(0, 0)); return; } } NS_WARNING("TextureClient::UpdateFromSurface failed"); } already_AddRefed TextureClient::CreateSimilar( LayersBackend aLayersBackend, TextureFlags aFlags, TextureAllocationFlags aAllocFlags) const { MOZ_ASSERT(IsValid()); MOZ_ASSERT(!mIsLocked); if (mIsLocked) { return nullptr; } LockActor(); TextureData* data = mData->CreateSimilar(mAllocator, aLayersBackend, aFlags, aAllocFlags); UnlockActor(); if (!data) { return nullptr; } return MakeAndAddRef(data, aFlags, mAllocator); } gfx::DrawTarget* TextureClient::BorrowDrawTarget() { MOZ_ASSERT(IsValid()); MOZ_ASSERT(mIsLocked); // TODO- We can't really assert that at the moment because there is code that // Borrows the DrawTarget, just to get a snapshot, which is legit in term of // OpenMode but we should have a way to get a SourceSurface directly instead. // MOZ_ASSERT(mOpenMode & OpenMode::OPEN_WRITE); if (!IsValid() || !mIsLocked) { return nullptr; } if (!mBorrowedDrawTarget) { mBorrowedDrawTarget = mData->BorrowDrawTarget(); #ifdef DEBUG mExpectedDtRefs = mBorrowedDrawTarget ? mBorrowedDrawTarget->refCount() : 0; #endif } return mBorrowedDrawTarget; } void TextureClient::EndDraw() { MOZ_ASSERT(mOpenMode & OpenMode::OPEN_READ_WRITE); // Because EndDraw is used when we are not unlocking this TextureClient at the // end of a transaction, we need to Flush and DetachAllSnapshots to ensure any // dependents are updated. mBorrowedDrawTarget->Flush(); mBorrowedDrawTarget->DetachAllSnapshots(); MOZ_ASSERT(mBorrowedDrawTarget->refCount() <= mExpectedDtRefs); mBorrowedDrawTarget = nullptr; mBorrowedSnapshot = false; mData->EndDraw(); } already_AddRefed TextureClient::BorrowSnapshot() { MOZ_ASSERT(mIsLocked); RefPtr surface = mData->BorrowSnapshot(); if (surface) { mBorrowedSnapshot = true; } else { RefPtr drawTarget = BorrowDrawTarget(); if (!drawTarget) { return nullptr; } surface = drawTarget->Snapshot(); } return surface.forget(); } void TextureClient::ReturnSnapshot( already_AddRefed aSnapshot) { RefPtr snapshot = aSnapshot; if (mBorrowedSnapshot) { mData->ReturnSnapshot(snapshot.forget()); mBorrowedSnapshot = false; } } bool TextureClient::BorrowMappedData(MappedTextureData& aMap) { MOZ_ASSERT(IsValid()); // TODO - SharedRGBImage just accesses the buffer without properly locking // the texture. It's bad. // MOZ_ASSERT(mIsLocked); // if (!mIsLocked) { // return nullptr; //} return mData ? mData->BorrowMappedData(aMap) : false; } bool TextureClient::BorrowMappedYCbCrData(MappedYCbCrTextureData& aMap) { MOZ_ASSERT(IsValid()); return mData ? mData->BorrowMappedYCbCrData(aMap) : false; } bool TextureClient::ToSurfaceDescriptor(SurfaceDescriptor& aOutDescriptor) { MOZ_ASSERT(IsValid()); return mData ? mData->Serialize(aOutDescriptor) : false; } // static PTextureChild* TextureClient::CreateIPDLActor() { TextureChild* c = new TextureChild(); c->AddIPDLReference(); return c; } // static bool TextureClient::DestroyIPDLActor(PTextureChild* actor) { static_cast(actor)->ReleaseIPDLReference(); return true; } // static already_AddRefed TextureClient::AsTextureClient( PTextureChild* actor) { if (!actor) { return nullptr; } TextureChild* tc = static_cast(actor); tc->Lock(); // Since TextureClient may be destroyed asynchronously with respect to its // IPDL actor, we must acquire a reference within a lock. The mDestroyed bit // tells us whether or not the main thread has disconnected the TextureClient // from its actor. if (tc->mDestroyed) { tc->Unlock(); return nullptr; } RefPtr texture = tc->mTextureClient; tc->Unlock(); return texture.forget(); } bool TextureClient::IsSharedWithCompositor() const { return mActor && mActor->IPCOpen(); } void TextureClient::AddFlags(TextureFlags aFlags) { MOZ_ASSERT( !IsSharedWithCompositor() || ((GetFlags() & TextureFlags::RECYCLE) && !IsAddedToCompositableClient())); mFlags |= aFlags; } void TextureClient::RemoveFlags(TextureFlags aFlags) { MOZ_ASSERT( !IsSharedWithCompositor() || ((GetFlags() & TextureFlags::RECYCLE) && !IsAddedToCompositableClient())); mFlags &= ~aFlags; } void TextureClient::RecycleTexture(TextureFlags aFlags) { MOZ_ASSERT(GetFlags() & TextureFlags::RECYCLE); MOZ_ASSERT(!mIsLocked); mAddedToCompositableClient = false; if (mFlags != aFlags) { mFlags = aFlags; } } void TextureClient::SetAddedToCompositableClient() { if (!mAddedToCompositableClient) { mAddedToCompositableClient = true; if (!(GetFlags() & TextureFlags::RECYCLE)) { return; } MOZ_ASSERT(!mIsLocked); LockActor(); if (IsValid() && mActor && !mActor->mDestroyed && mActor->IPCOpen()) { mActor->SendRecycleTexture(mFlags); } UnlockActor(); } } static void CancelTextureClientNotifyNotUsed(uint64_t aTextureId, LayersIPCChannel* aAllocator) { if (!aAllocator) { return; } nsCOMPtr thread = aAllocator->GetThread(); if (!thread) { return; } if (thread->IsOnCurrentThread()) { aAllocator->CancelWaitForNotifyNotUsed(aTextureId); } else { thread->Dispatch(NewRunnableFunction( "CancelTextureClientNotifyNotUsedRunnable", CancelTextureClientNotifyNotUsed, aTextureId, aAllocator)); } } void TextureClient::CancelWaitForNotifyNotUsed() { if (GetFlags() & TextureFlags::RECYCLE) { CancelTextureClientNotifyNotUsed(mSerial, GetAllocator()); return; } } /* static */ void TextureClient::TextureClientRecycleCallback(TextureClient* aClient, void* aClosure) { MOZ_ASSERT(aClient->GetRecycleAllocator()); aClient->GetRecycleAllocator()->RecycleTextureClient(aClient); } void TextureClient::SetRecycleAllocator( ITextureClientRecycleAllocator* aAllocator) { mRecycleAllocator = aAllocator; if (aAllocator) { SetRecycleCallback(TextureClientRecycleCallback, nullptr); } else { ClearRecycleCallback(); } } bool TextureClient::InitIPDLActor(CompositableForwarder* aForwarder) { MOZ_ASSERT(aForwarder && aForwarder->GetTextureForwarder()->GetThread() == mAllocator->GetThread()); if (mActor && !mActor->IPCOpen()) { return false; } if (mActor && !mActor->mDestroyed) { CompositableForwarder* currentFwd = mActor->mCompositableForwarder; TextureForwarder* currentTexFwd = mActor->mTextureForwarder; if (currentFwd != aForwarder) { // It's a bit iffy but right now ShadowLayerForwarder inherits // TextureForwarder even though it should not. // ShadowLayerForwarder::GetTextureForwarder actually returns a pointer to // the CompositorBridgeChild. It's Ok for a texture to move from a // ShadowLayerForwarder to another, but not form a CompositorBridgeChild // to another (they use different channels). if (currentTexFwd && currentTexFwd != aForwarder->GetTextureForwarder()) { gfxCriticalError() << "Attempt to move a texture to a different channel CF."; MOZ_ASSERT_UNREACHABLE("unexpected to be called"); return false; } if (currentFwd && currentFwd->GetCompositorBackendType() != aForwarder->GetCompositorBackendType()) { gfxCriticalError() << "Attempt to move a texture to different compositor backend."; MOZ_ASSERT_UNREACHABLE("unexpected to be called"); return false; } mActor->mCompositableForwarder = aForwarder; mActor->mUsesImageBridge = aForwarder->GetTextureForwarder()->UsesImageBridge(); } return true; } MOZ_ASSERT(!mActor || mActor->mDestroyed, "Cannot use a texture on several IPC channels."); SurfaceDescriptor desc; if (!ToSurfaceDescriptor(desc)) { return false; } // Try external image id allocation. mExternalImageId = aForwarder->GetTextureForwarder()->GetNextExternalImageId(); ReadLockDescriptor readLockDescriptor = null_t(); { MutexAutoLock lock(mMutex); EnsureHasReadLock(); if (mReadLock) { mReadLock->Serialize(readLockDescriptor, GetAllocator()->GetParentPid()); } } PTextureChild* actor = aForwarder->GetTextureForwarder()->CreateTexture( desc, std::move(readLockDescriptor), aForwarder->GetCompositorBackendType(), GetFlags(), dom::ContentParentId(), mSerial, mExternalImageId); if (!actor) { gfxCriticalNote << static_cast(desc.type()) << ", " << static_cast( aForwarder->GetCompositorBackendType()) << ", " << static_cast(GetFlags()) << ", " << mSerial; return false; } mActor = static_cast(actor); mActor->mCompositableForwarder = aForwarder; mActor->mTextureForwarder = aForwarder->GetTextureForwarder(); mActor->mTextureClient = this; // If the TextureClient is already locked, we have to lock TextureChild's // mutex since it will be unlocked in TextureClient::Unlock. if (mIsLocked) { LockActor(); } return mActor->IPCOpen(); } bool TextureClient::InitIPDLActor(KnowsCompositor* aKnowsCompositor, const dom::ContentParentId& aContentId) { MOZ_ASSERT(aKnowsCompositor && aKnowsCompositor->GetTextureForwarder()->GetThread() == mAllocator->GetThread()); TextureForwarder* fwd = aKnowsCompositor->GetTextureForwarder(); if (mActor && !mActor->mDestroyed) { CompositableForwarder* currentFwd = mActor->mCompositableForwarder; TextureForwarder* currentTexFwd = mActor->mTextureForwarder; if (currentFwd) { gfxCriticalError() << "Attempt to remove a texture from a CompositableForwarder."; return false; } if (currentTexFwd && currentTexFwd != fwd) { gfxCriticalError() << "Attempt to move a texture to a different channel TF."; return false; } mActor->mTextureForwarder = fwd; return true; } MOZ_ASSERT(!mActor || mActor->mDestroyed, "Cannot use a texture on several IPC channels."); SurfaceDescriptor desc; if (!ToSurfaceDescriptor(desc)) { return false; } // Try external image id allocation. mExternalImageId = aKnowsCompositor->GetTextureForwarder()->GetNextExternalImageId(); ReadLockDescriptor readLockDescriptor = null_t(); { MutexAutoLock lock(mMutex); EnsureHasReadLock(); if (mReadLock) { mReadLock->Serialize(readLockDescriptor, GetAllocator()->GetParentPid()); } } PTextureChild* actor = fwd->CreateTexture(desc, std::move(readLockDescriptor), aKnowsCompositor->GetCompositorBackendType(), GetFlags(), aContentId, mSerial, mExternalImageId); if (!actor) { gfxCriticalNote << static_cast(desc.type()) << ", " << static_cast( aKnowsCompositor->GetCompositorBackendType()) << ", " << static_cast(GetFlags()) << ", " << mSerial; return false; } mActor = static_cast(actor); mActor->mTextureForwarder = fwd; mActor->mTextureClient = this; // If the TextureClient is already locked, we have to lock TextureChild's // mutex since it will be unlocked in TextureClient::Unlock. if (mIsLocked) { LockActor(); } return mActor->IPCOpen(); } PTextureChild* TextureClient::GetIPDLActor() { return mActor; } // static already_AddRefed TextureClient::CreateForDrawing( KnowsCompositor* aAllocator, gfx::SurfaceFormat aFormat, gfx::IntSize aSize, BackendSelector aSelector, TextureFlags aTextureFlags, TextureAllocationFlags aAllocFlags) { return TextureClient::CreateForDrawing(aAllocator->GetTextureForwarder(), aFormat, aSize, aAllocator, aSelector, aTextureFlags, aAllocFlags); } // static already_AddRefed TextureClient::CreateForDrawing( TextureForwarder* aAllocator, gfx::SurfaceFormat aFormat, gfx::IntSize aSize, KnowsCompositor* aKnowsCompositor, BackendSelector aSelector, TextureFlags aTextureFlags, TextureAllocationFlags aAllocFlags) { LayersBackend layersBackend = aKnowsCompositor->GetCompositorBackendType(); gfx::BackendType moz2DBackend = BackendTypeForBackendSelector(layersBackend, aSelector); // also test the validity of aAllocator if (!aAllocator || !aAllocator->IPCOpen()) { return nullptr; } if (!gfx::Factory::AllowedSurfaceSize(aSize)) { return nullptr; } TextureData* data = TextureData::Create(aAllocator, aFormat, aSize, aKnowsCompositor, aSelector, aTextureFlags, aAllocFlags); if (data) { return MakeAndAddRef(data, aTextureFlags, aAllocator); } if (aAllocFlags & ALLOC_FORCE_REMOTE) { // If we must be remote, but allocation failed, then don't fall back. return nullptr; } // Can't do any better than a buffer texture client. return TextureClient::CreateForRawBufferAccess(aAllocator, aFormat, aSize, moz2DBackend, layersBackend, aTextureFlags, aAllocFlags); } // static already_AddRefed TextureClient::CreateFromSurface( KnowsCompositor* aAllocator, gfx::SourceSurface* aSurface, BackendSelector aSelector, TextureFlags aTextureFlags, TextureAllocationFlags aAllocFlags) { // also test the validity of aAllocator if (!aAllocator || !aAllocator->GetTextureForwarder()->IPCOpen()) { return nullptr; } gfx::IntSize size = aSurface->GetSize(); if (!gfx::Factory::AllowedSurfaceSize(size)) { return nullptr; } TextureData* data = nullptr; #if defined(XP_WIN) LayersBackend layersBackend = aAllocator->GetCompositorBackendType(); gfx::BackendType moz2DBackend = BackendTypeForBackendSelector(layersBackend, aSelector); int32_t maxTextureSize = aAllocator->GetMaxTextureSize(); if (layersBackend == LayersBackend::LAYERS_WR && (moz2DBackend == gfx::BackendType::DIRECT2D || moz2DBackend == gfx::BackendType::DIRECT2D1_1) && size.width <= maxTextureSize && size.height <= maxTextureSize) { data = D3D11TextureData::Create(aSurface, aAllocFlags); } #endif if (data) { return MakeAndAddRef(data, aTextureFlags, aAllocator->GetTextureForwarder()); } // Fall back to using UpdateFromSurface TextureAllocationFlags allocFlags = TextureAllocationFlags(aAllocFlags | ALLOC_UPDATE_FROM_SURFACE); RefPtr client = CreateForDrawing(aAllocator, aSurface->GetFormat(), size, aSelector, aTextureFlags, allocFlags); if (!client) { return nullptr; } TextureClientAutoLock autoLock(client, OpenMode::OPEN_WRITE_ONLY); if (!autoLock.Succeeded()) { return nullptr; } client->UpdateFromSurface(aSurface); return client.forget(); } // static already_AddRefed TextureClient::CreateForRawBufferAccess( KnowsCompositor* aAllocator, gfx::SurfaceFormat aFormat, gfx::IntSize aSize, gfx::BackendType aMoz2DBackend, TextureFlags aTextureFlags, TextureAllocationFlags aAllocFlags) { return CreateForRawBufferAccess( aAllocator->GetTextureForwarder(), aFormat, aSize, aMoz2DBackend, aAllocator->GetCompositorBackendType(), aTextureFlags, aAllocFlags); } // static already_AddRefed TextureClient::CreateForRawBufferAccess( LayersIPCChannel* aAllocator, gfx::SurfaceFormat aFormat, gfx::IntSize aSize, gfx::BackendType aMoz2DBackend, LayersBackend aLayersBackend, TextureFlags aTextureFlags, TextureAllocationFlags aAllocFlags) { // also test the validity of aAllocator if (!aAllocator || !aAllocator->IPCOpen()) { return nullptr; } if (!gfx::Factory::AllowedSurfaceSize(aSize)) { return nullptr; } if (aFormat == SurfaceFormat::B8G8R8X8) { // Skia doesn't support RGBX, so ensure we clear the buffer for the proper // alpha values. aAllocFlags = TextureAllocationFlags(aAllocFlags | ALLOC_CLEAR_BUFFER); } // Note that we ignore the backend type if we get here. It should only be D2D // or Skia, and D2D does not support data surfaces. Therefore it is safe to // force the buffer to be Skia. NS_WARNING_ASSERTION(aMoz2DBackend == gfx::BackendType::SKIA || aMoz2DBackend == gfx::BackendType::DIRECT2D || aMoz2DBackend == gfx::BackendType::DIRECT2D1_1, "Unsupported TextureClient backend type"); TextureData* texData = BufferTextureData::Create( aSize, aFormat, gfx::BackendType::SKIA, aLayersBackend, aTextureFlags, aAllocFlags, aAllocator); if (!texData) { return nullptr; } return MakeAndAddRef(texData, aTextureFlags, aAllocator); } // static already_AddRefed TextureClient::CreateForYCbCr( KnowsCompositor* aAllocator, const gfx::IntRect& aDisplay, const gfx::IntSize& aYSize, uint32_t aYStride, const gfx::IntSize& aCbCrSize, uint32_t aCbCrStride, StereoMode aStereoMode, gfx::ColorDepth aColorDepth, gfx::YUVColorSpace aYUVColorSpace, gfx::ColorRange aColorRange, gfx::ChromaSubsampling aSubsampling, TextureFlags aTextureFlags) { if (!aAllocator || !aAllocator->GetLayersIPCActor()->IPCOpen()) { return nullptr; } if (!gfx::Factory::AllowedSurfaceSize(aYSize)) { return nullptr; } TextureData* data = BufferTextureData::CreateForYCbCr( aAllocator, aDisplay, aYSize, aYStride, aCbCrSize, aCbCrStride, aStereoMode, aColorDepth, aYUVColorSpace, aColorRange, aSubsampling, aTextureFlags); if (!data) { return nullptr; } return MakeAndAddRef(data, aTextureFlags, aAllocator->GetTextureForwarder()); } TextureClient::TextureClient(TextureData* aData, TextureFlags aFlags, LayersIPCChannel* aAllocator) : AtomicRefCountedWithFinalize("TextureClient"), mMutex("TextureClient::mMutex"), mAllocator(aAllocator), mActor(nullptr), mData(aData), mFlags(aFlags), mOpenMode(OpenMode::OPEN_NONE) #ifdef DEBUG , mExpectedDtRefs(0) #endif , mIsLocked(false), mIsReadLocked(false), mUpdated(false), mAddedToCompositableClient(false), mFwdTransactionId(0), mSerial(++sSerialCounter) #ifdef GFX_DEBUG_TRACK_CLIENTS_IN_POOL , mPoolTracker(nullptr) #endif { mData->FillInfo(mInfo); mFlags |= mData->GetTextureFlags(); } bool TextureClient::CopyToTextureClient(TextureClient* aTarget, const gfx::IntRect* aRect, const gfx::IntPoint* aPoint) { MOZ_ASSERT(IsLocked()); MOZ_ASSERT(aTarget->IsLocked()); if (!aTarget->CanExposeDrawTarget() || !CanExposeDrawTarget()) { return false; } RefPtr destinationTarget = aTarget->BorrowDrawTarget(); if (!destinationTarget) { gfxWarning() << "TextureClient::CopyToTextureClient (dest) failed in " "BorrowDrawTarget"; return false; } RefPtr sourceTarget = BorrowDrawTarget(); if (!sourceTarget) { gfxWarning() << "TextureClient::CopyToTextureClient (src) failed in " "BorrowDrawTarget"; return false; } RefPtr source = sourceTarget->Snapshot(); destinationTarget->CopySurface( source, aRect ? *aRect : gfx::IntRect(gfx::IntPoint(0, 0), GetSize()), aPoint ? *aPoint : gfx::IntPoint(0, 0)); return true; } already_AddRefed TextureClient::GetAsSurface() { if (!Lock(OpenMode::OPEN_READ)) { return nullptr; } RefPtr data; { // scope so that the DrawTarget is destroyed before Unlock() RefPtr dt = BorrowDrawTarget(); if (dt) { RefPtr surf = dt->Snapshot(); if (surf) { data = surf->GetDataSurface(); } } } Unlock(); return data.forget(); } void TextureClient::GetSurfaceDescriptorRemoteDecoder( SurfaceDescriptorRemoteDecoder* const aOutDesc) { const auto handle = GetSerial(); RemoteDecoderVideoSubDescriptor subDesc = null_t(); MOZ_RELEASE_ASSERT(mData); mData->GetSubDescriptor(&subDesc); *aOutDesc = SurfaceDescriptorRemoteDecoder(handle, std::move(subDesc), Nothing()); } class MemoryTextureReadLock : public NonBlockingTextureReadLock { public: MemoryTextureReadLock(); virtual ~MemoryTextureReadLock(); bool ReadLock() override; int32_t ReadUnlock() override; int32_t GetReadCount() override; LockType GetType() override { return TYPE_NONBLOCKING_MEMORY; } bool IsValid() const override { return true; }; bool Serialize(ReadLockDescriptor& aOutput, base::ProcessId aOther) override; Atomic mReadCount; }; // The cross-prcess implementation of TextureReadLock. // // Since we don't use cross-process reference counting for the ReadLock objects, // we use the lock's internal counter as a way to know when to deallocate the // underlying shmem section: when the counter is equal to 1, it means that the // lock is not "held" (the texture is writable), when the counter is equal to 0 // it means that we can safely deallocate the shmem section without causing a // race condition with the other process. class ShmemTextureReadLock : public NonBlockingTextureReadLock { public: struct ShmReadLockInfo { int32_t readCount; }; explicit ShmemTextureReadLock(LayersIPCChannel* aAllocator); virtual ~ShmemTextureReadLock(); bool ReadLock() override; int32_t ReadUnlock() override; int32_t GetReadCount() override; bool IsValid() const override { return mAllocSuccess; }; LockType GetType() override { return TYPE_NONBLOCKING_SHMEM; } bool Serialize(ReadLockDescriptor& aOutput, base::ProcessId aOther) override; mozilla::layers::ShmemSection& GetShmemSection() { return mShmemSection; } explicit ShmemTextureReadLock( const mozilla::layers::ShmemSection& aShmemSection) : mShmemSection(aShmemSection), mAllocSuccess(true) { MOZ_COUNT_CTOR(ShmemTextureReadLock); } ShmReadLockInfo* GetShmReadLockInfoPtr() { return reinterpret_cast( mShmemSection.shmem().get() + mShmemSection.offset()); } RefPtr mClientAllocator; mozilla::layers::ShmemSection mShmemSection; bool mAllocSuccess; }; class CrossProcessSemaphoreReadLock : public TextureReadLock { public: CrossProcessSemaphoreReadLock() : mSemaphore(CrossProcessSemaphore::Create("TextureReadLock", 1)), mShared(false) {} explicit CrossProcessSemaphoreReadLock(CrossProcessSemaphoreHandle aHandle) : mSemaphore(CrossProcessSemaphore::Create(std::move(aHandle))), mShared(false) {} bool ReadLock() override { if (!IsValid()) { return false; } return mSemaphore->Wait(); } bool TryReadLock(TimeDuration aTimeout) override { if (!IsValid()) { return false; } return mSemaphore->Wait(Some(aTimeout)); } int32_t ReadUnlock() override { if (!IsValid()) { return 1; } mSemaphore->Signal(); return 1; } bool IsValid() const override { return !!mSemaphore; } bool Serialize(ReadLockDescriptor& aOutput, base::ProcessId aOther) override; LockType GetType() override { return TYPE_CROSS_PROCESS_SEMAPHORE; } UniquePtr mSemaphore; bool mShared; }; // static already_AddRefed TextureReadLock::Deserialize( ReadLockDescriptor&& aDescriptor, ISurfaceAllocator* aAllocator) { switch (aDescriptor.type()) { case ReadLockDescriptor::TShmemSection: { const ShmemSection& section = aDescriptor.get_ShmemSection(); MOZ_RELEASE_ASSERT(section.shmem().IsReadable()); return MakeAndAddRef(section); } case ReadLockDescriptor::Tuintptr_t: { if (!aAllocator->IsSameProcess()) { // Trying to use a memory based lock instead of a shmem based one in // the cross-process case is a bad security violation. NS_ERROR( "A client process may be trying to peek at the host's address " "space!"); return nullptr; } RefPtr lock = reinterpret_cast(aDescriptor.get_uintptr_t()); MOZ_ASSERT(lock); if (lock) { // The corresponding AddRef is in MemoryTextureReadLock::Serialize lock.get()->Release(); } return lock.forget(); } case ReadLockDescriptor::TCrossProcessSemaphoreDescriptor: { return MakeAndAddRef( std::move(aDescriptor.get_CrossProcessSemaphoreDescriptor().sem())); } case ReadLockDescriptor::Tnull_t: { return nullptr; } default: { // Invalid descriptor. MOZ_DIAGNOSTIC_ASSERT(false); } } return nullptr; } // static already_AddRefed NonBlockingTextureReadLock::Create( LayersIPCChannel* aAllocator) { if (aAllocator->IsSameProcess()) { // If our compositor is in the same process, we can save some cycles by not // using shared memory. return MakeAndAddRef(); } return MakeAndAddRef(aAllocator); } MemoryTextureReadLock::MemoryTextureReadLock() : mReadCount(1) { MOZ_COUNT_CTOR(MemoryTextureReadLock); } MemoryTextureReadLock::~MemoryTextureReadLock() { // One read count that is added in constructor. MOZ_ASSERT(mReadCount == 1); MOZ_COUNT_DTOR(MemoryTextureReadLock); } bool MemoryTextureReadLock::Serialize(ReadLockDescriptor& aOutput, base::ProcessId aOther) { // AddRef here and Release when receiving on the host side to make sure the // reference count doesn't go to zero before the host receives the message. // see TextureReadLock::Deserialize this->AddRef(); aOutput = ReadLockDescriptor(uintptr_t(this)); return true; } bool MemoryTextureReadLock::ReadLock() { ++mReadCount; return true; } int32_t MemoryTextureReadLock::ReadUnlock() { int32_t readCount = --mReadCount; MOZ_ASSERT(readCount >= 0); return readCount; } int32_t MemoryTextureReadLock::GetReadCount() { return mReadCount; } ShmemTextureReadLock::ShmemTextureReadLock(LayersIPCChannel* aAllocator) : mClientAllocator(aAllocator), mAllocSuccess(false) { MOZ_COUNT_CTOR(ShmemTextureReadLock); MOZ_ASSERT(mClientAllocator); MOZ_ASSERT(mClientAllocator->GetTileLockAllocator()); #define MOZ_ALIGN_WORD(x) (((x) + 3) & ~3) if (mClientAllocator->GetTileLockAllocator()->AllocShmemSection( MOZ_ALIGN_WORD(sizeof(ShmReadLockInfo)), &mShmemSection)) { ShmReadLockInfo* info = GetShmReadLockInfoPtr(); info->readCount = 1; mAllocSuccess = true; } } ShmemTextureReadLock::~ShmemTextureReadLock() { if (mClientAllocator) { // Release one read count that is added in constructor. // The count is kept for calling GetReadCount() by TextureClientPool. ReadUnlock(); } MOZ_COUNT_DTOR(ShmemTextureReadLock); } bool ShmemTextureReadLock::Serialize(ReadLockDescriptor& aOutput, base::ProcessId aOther) { aOutput = ReadLockDescriptor(GetShmemSection()); return true; } bool ShmemTextureReadLock::ReadLock() { if (!mAllocSuccess) { return false; } ShmReadLockInfo* info = GetShmReadLockInfoPtr(); PR_ATOMIC_INCREMENT(&info->readCount); return true; } int32_t ShmemTextureReadLock::ReadUnlock() { if (!mAllocSuccess) { return 0; } ShmReadLockInfo* info = GetShmReadLockInfoPtr(); int32_t readCount = PR_ATOMIC_DECREMENT(&info->readCount); MOZ_ASSERT(readCount >= 0); if (readCount > 0) { return readCount; } if (mClientAllocator) { if (nsCOMPtr thread = mClientAllocator->GetThread()) { if (thread->IsOnCurrentThread()) { if (auto* tileLockAllocator = mClientAllocator->GetTileLockAllocator()) { tileLockAllocator->DeallocShmemSection(mShmemSection); return readCount; } } else { thread->Dispatch(NS_NewRunnableFunction( __func__, [shmemSection = std::move(mShmemSection), clientAllocator = std::move(mClientAllocator)]() mutable { if (auto* tileLockAllocator = clientAllocator->GetTileLockAllocator()) { tileLockAllocator->DeallocShmemSection(shmemSection); } else { // we are on the compositor process, or IPC is down. FixedSizeSmallShmemSectionAllocator::FreeShmemSection( shmemSection); } })); return readCount; } } } // we are on the compositor process, or IPC is down. FixedSizeSmallShmemSectionAllocator::FreeShmemSection(mShmemSection); return readCount; } int32_t ShmemTextureReadLock::GetReadCount() { if (!mAllocSuccess) { return 0; } ShmReadLockInfo* info = GetShmReadLockInfoPtr(); return info->readCount; } bool CrossProcessSemaphoreReadLock::Serialize(ReadLockDescriptor& aOutput, base::ProcessId aOther) { if (!mShared && IsValid()) { aOutput = ReadLockDescriptor( CrossProcessSemaphoreDescriptor(mSemaphore->CloneHandle())); mSemaphore->CloseHandle(); mShared = true; return true; } else { return mShared; } } void TextureClient::EnableBlockingReadLock() { MOZ_ASSERT(ShouldReadLock()); if (!mReadLock) { mReadLock = new CrossProcessSemaphoreReadLock(); } } bool UpdateYCbCrTextureClient(TextureClient* aTexture, const PlanarYCbCrData& aData) { MOZ_ASSERT(aTexture); MOZ_ASSERT(aTexture->IsLocked()); MOZ_ASSERT(aTexture->GetFormat() == gfx::SurfaceFormat::YUV, "This textureClient can only use YCbCr data"); MOZ_ASSERT(!aTexture->IsImmutable()); MOZ_ASSERT(aTexture->IsValid()); MOZ_ASSERT(aData.mCbSkip == aData.mCrSkip); MappedYCbCrTextureData mapped; if (!aTexture->BorrowMappedYCbCrData(mapped)) { NS_WARNING("Failed to extract YCbCr info!"); return false; } uint32_t bytesPerPixel = BytesPerPixel(SurfaceFormatForColorDepth(aData.mColorDepth)); MappedYCbCrTextureData srcData; srcData.y.data = aData.mYChannel; srcData.y.size = aData.YDataSize(); srcData.y.stride = aData.mYStride; srcData.y.skip = aData.mYSkip; srcData.y.bytesPerPixel = bytesPerPixel; srcData.cb.data = aData.mCbChannel; srcData.cb.size = aData.CbCrDataSize(); srcData.cb.stride = aData.mCbCrStride; srcData.cb.skip = aData.mCbSkip; srcData.cb.bytesPerPixel = bytesPerPixel; srcData.cr.data = aData.mCrChannel; srcData.cr.size = aData.CbCrDataSize(); srcData.cr.stride = aData.mCbCrStride; srcData.cr.skip = aData.mCrSkip; srcData.cr.bytesPerPixel = bytesPerPixel; srcData.metadata = nullptr; if (!srcData.CopyInto(mapped)) { NS_WARNING("Failed to copy image data!"); return false; } if (TextureRequiresLocking(aTexture->GetFlags())) { // We don't have support for proper locking yet, so we'll // have to be immutable instead. aTexture->MarkImmutable(); } return true; } already_AddRefed TextureClient::CreateWithData( TextureData* aData, TextureFlags aFlags, LayersIPCChannel* aAllocator) { if (!aData) { return nullptr; } return MakeAndAddRef(aData, aFlags, aAllocator); } template static void copyData(PixelDataType* aDst, const MappedYCbCrChannelData& aChannelDst, PixelDataType* aSrc, const MappedYCbCrChannelData& aChannelSrc) { uint8_t* srcByte = reinterpret_cast(aSrc); const int32_t srcSkip = aChannelSrc.skip + 1; uint8_t* dstByte = reinterpret_cast(aDst); const int32_t dstSkip = aChannelDst.skip + 1; for (int32_t i = 0; i < aChannelSrc.size.height; ++i) { for (int32_t j = 0; j < aChannelSrc.size.width; ++j) { *aDst = *aSrc; aSrc += srcSkip; aDst += dstSkip; } srcByte += aChannelSrc.stride; aSrc = reinterpret_cast(srcByte); dstByte += aChannelDst.stride; aDst = reinterpret_cast(dstByte); } } bool MappedYCbCrChannelData::CopyInto(MappedYCbCrChannelData& aDst) { if (!data || !aDst.data || size != aDst.size) { return false; } if (stride == aDst.stride && skip == aDst.skip) { // fast path! // We assume that the padding in the destination is there for alignment // purposes and doesn't contain useful data. memcpy(aDst.data, data, stride * size.height); return true; } if (aDst.skip == 0 && skip == 0) { // fast-ish path for (int32_t i = 0; i < size.height; ++i) { memcpy(aDst.data + i * aDst.stride, data + i * stride, size.width * bytesPerPixel); } return true; } MOZ_ASSERT(bytesPerPixel == 1 || bytesPerPixel == 2); // slow path if (bytesPerPixel == 1) { copyData(aDst.data, aDst, data, *this); } else if (bytesPerPixel == 2) { copyData(reinterpret_cast(aDst.data), aDst, reinterpret_cast(data), *this); } return true; } } // namespace mozilla::layers