summaryrefslogtreecommitdiffstats
path: root/dom/webgpu/ipc/WebGPUParent.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'dom/webgpu/ipc/WebGPUParent.cpp')
-rw-r--r--dom/webgpu/ipc/WebGPUParent.cpp1557
1 files changed, 1557 insertions, 0 deletions
diff --git a/dom/webgpu/ipc/WebGPUParent.cpp b/dom/webgpu/ipc/WebGPUParent.cpp
new file mode 100644
index 0000000000..9b79988245
--- /dev/null
+++ b/dom/webgpu/ipc/WebGPUParent.cpp
@@ -0,0 +1,1557 @@
+/* -*- Mode: C++; tab-width: 20; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
+/* This Source Code Form is subject to the terms of the Mozilla Public
+ * License, v. 2.0. If a copy of the MPL was not distributed with this
+ * file, You can obtain one at http://mozilla.org/MPL/2.0/. */
+
+#include "WebGPUParent.h"
+
+#include <unordered_set>
+
+#include "mozilla/PodOperations.h"
+#include "mozilla/ScopeExit.h"
+#include "mozilla/dom/WebGPUBinding.h"
+#include "mozilla/gfx/FileHandleWrapper.h"
+#include "mozilla/layers/CompositorThread.h"
+#include "mozilla/layers/ImageDataSerializer.h"
+#include "mozilla/layers/RemoteTextureMap.h"
+#include "mozilla/layers/TextureHost.h"
+#include "mozilla/layers/WebRenderImageHost.h"
+#include "mozilla/layers/WebRenderTextureHost.h"
+#include "mozilla/webgpu/ExternalTexture.h"
+#include "mozilla/webgpu/ffi/wgpu.h"
+
+#if defined(XP_WIN)
+# include "mozilla/gfx/DeviceManagerDx.h"
+#endif
+
+namespace mozilla::webgpu {
+
+const uint64_t POLL_TIME_MS = 100;
+
+static mozilla::LazyLogModule sLogger("WebGPU");
+
+namespace ffi {
+
+extern bool wgpu_server_use_external_texture_for_swap_chain(
+ void* aParam, WGPUSwapChainId aSwapChainId) {
+ auto* parent = static_cast<WebGPUParent*>(aParam);
+
+ return parent->UseExternalTextureForSwapChain(aSwapChainId);
+}
+
+extern bool wgpu_server_ensure_external_texture_for_swap_chain(
+ void* aParam, WGPUSwapChainId aSwapChainId, WGPUDeviceId aDeviceId,
+ WGPUTextureId aTextureId, uint32_t aWidth, uint32_t aHeight,
+ struct WGPUTextureFormat aFormat, WGPUTextureUsages aUsage) {
+ auto* parent = static_cast<WebGPUParent*>(aParam);
+
+ return parent->EnsureExternalTextureForSwapChain(
+ aSwapChainId, aDeviceId, aTextureId, aWidth, aHeight, aFormat, aUsage);
+}
+
+extern void* wgpu_server_get_external_texture_handle(void* aParam,
+ WGPUTextureId aId) {
+ auto* parent = static_cast<WebGPUParent*>(aParam);
+
+ auto texture = parent->GetExternalTexture(aId);
+ if (!texture) {
+ MOZ_ASSERT_UNREACHABLE("unexpected to be called");
+ return nullptr;
+ }
+
+ void* sharedHandle = nullptr;
+#ifdef XP_WIN
+ sharedHandle = texture->GetExternalTextureHandle();
+ if (!sharedHandle) {
+ MOZ_ASSERT_UNREACHABLE("unexpected to be called");
+ gfxCriticalNoteOnce << "Failed to get shared handle";
+ return nullptr;
+ }
+#else
+ MOZ_ASSERT_UNREACHABLE("unexpected to be called");
+#endif
+ return sharedHandle;
+}
+
+} // namespace ffi
+
+// A fixed-capacity buffer for receiving textual error messages from
+// `wgpu_bindings`.
+//
+// The `ToFFI` method returns an `ffi::WGPUErrorBuffer` pointing to our
+// buffer, for you to pass to fallible FFI-visible `wgpu_bindings`
+// functions. These indicate failure by storing an error message in the
+// buffer, which you can retrieve by calling `GetError`.
+//
+// If you call `ToFFI` on this type, you must also call `GetError` to check for
+// an error. Otherwise, the destructor asserts.
+//
+// TODO: refactor this to avoid stack-allocating the buffer all the time.
+class ErrorBuffer {
+ // if the message doesn't fit, it will be truncated
+ static constexpr unsigned BUFFER_SIZE = 512;
+ ffi::WGPUErrorBufferType mType = ffi::WGPUErrorBufferType_None;
+ char mMessageUtf8[BUFFER_SIZE] = {};
+ bool mAwaitingGetError = false;
+
+ public:
+ ErrorBuffer() { mMessageUtf8[0] = 0; }
+ ErrorBuffer(const ErrorBuffer&) = delete;
+ ~ErrorBuffer() { MOZ_ASSERT(!mAwaitingGetError); }
+
+ ffi::WGPUErrorBuffer ToFFI() {
+ mAwaitingGetError = true;
+ ffi::WGPUErrorBuffer errorBuf = {&mType, mMessageUtf8, BUFFER_SIZE};
+ return errorBuf;
+ }
+
+ ffi::WGPUErrorBufferType GetType() { return mType; }
+
+ static Maybe<dom::GPUErrorFilter> ErrorTypeToFilterType(
+ ffi::WGPUErrorBufferType aType) {
+ switch (aType) {
+ case ffi::WGPUErrorBufferType_None:
+ case ffi::WGPUErrorBufferType_DeviceLost:
+ return {};
+ case ffi::WGPUErrorBufferType_Internal:
+ return Some(dom::GPUErrorFilter::Internal);
+ case ffi::WGPUErrorBufferType_Validation:
+ return Some(dom::GPUErrorFilter::Validation);
+ case ffi::WGPUErrorBufferType_OutOfMemory:
+ return Some(dom::GPUErrorFilter::Out_of_memory);
+ case ffi::WGPUErrorBufferType_Sentinel:
+ break;
+ }
+
+ MOZ_CRASH("invalid `ErrorBufferType`");
+ }
+
+ struct Error {
+ dom::GPUErrorFilter type;
+ bool isDeviceLost;
+ nsCString message;
+ };
+
+ // Retrieve the error message was stored in this buffer. Asserts that
+ // this instance actually contains an error (viz., that `GetType() !=
+ // ffi::WGPUErrorBufferType_None`).
+ //
+ // Mark this `ErrorBuffer` as having been handled, so its destructor
+ // won't assert.
+ Maybe<Error> GetError() {
+ mAwaitingGetError = false;
+ if (mType == ffi::WGPUErrorBufferType_DeviceLost) {
+ // This error is for a lost device, so we return an Error struct
+ // with the isDeviceLost bool set to true. It doesn't matter what
+ // GPUErrorFilter type we use, so we just use Validation. The error
+ // will not be reported.
+ return Some(Error{dom::GPUErrorFilter::Validation, true,
+ nsCString{mMessageUtf8}});
+ }
+ auto filterType = ErrorTypeToFilterType(mType);
+ if (!filterType) {
+ return {};
+ }
+ return Some(Error{*filterType, false, nsCString{mMessageUtf8}});
+ }
+};
+
+struct PendingSwapChainDrop {
+ layers::RemoteTextureTxnType mTxnType;
+ layers::RemoteTextureTxnId mTxnId;
+};
+
+class PresentationData {
+ NS_INLINE_DECL_REFCOUNTING(PresentationData);
+
+ public:
+ WeakPtr<WebGPUParent> mParent;
+ const bool mUseExternalTextureInSwapChain;
+ const RawId mDeviceId;
+ const RawId mQueueId;
+ const layers::RGBDescriptor mDesc;
+
+ uint64_t mSubmissionIndex = 0;
+
+ std::deque<std::shared_ptr<ExternalTexture>> mRecycledExternalTextures;
+
+ std::unordered_set<layers::RemoteTextureId, layers::RemoteTextureId::HashFn>
+ mWaitingReadbackTexturesForPresent;
+ Maybe<PendingSwapChainDrop> mPendingSwapChainDrop;
+
+ const uint32_t mSourcePitch;
+ std::vector<RawId> mUnassignedBufferIds MOZ_GUARDED_BY(mBuffersLock);
+ std::vector<RawId> mAvailableBufferIds MOZ_GUARDED_BY(mBuffersLock);
+ std::vector<RawId> mQueuedBufferIds MOZ_GUARDED_BY(mBuffersLock);
+ Mutex mBuffersLock;
+
+ PresentationData(WebGPUParent* aParent, bool aUseExternalTextureInSwapChain,
+ RawId aDeviceId, RawId aQueueId,
+ const layers::RGBDescriptor& aDesc, uint32_t aSourcePitch,
+ const nsTArray<RawId>& aBufferIds)
+ : mParent(aParent),
+ mUseExternalTextureInSwapChain(aUseExternalTextureInSwapChain),
+ mDeviceId(aDeviceId),
+ mQueueId(aQueueId),
+ mDesc(aDesc),
+ mSourcePitch(aSourcePitch),
+ mBuffersLock("WebGPU presentation buffers") {
+ MOZ_COUNT_CTOR(PresentationData);
+
+ for (const RawId id : aBufferIds) {
+ mUnassignedBufferIds.push_back(id);
+ }
+ }
+
+ private:
+ ~PresentationData() { MOZ_COUNT_DTOR(PresentationData); }
+};
+
+WebGPUParent::WebGPUParent() : mContext(ffi::wgpu_server_new(this)) {
+ mTimer.Start(base::TimeDelta::FromMilliseconds(POLL_TIME_MS), this,
+ &WebGPUParent::MaintainDevices);
+}
+
+WebGPUParent::~WebGPUParent() {
+ // All devices should have been dropped, but maybe they weren't. To
+ // ensure we don't leak memory, clear the mDeviceLostRequests.
+ mDeviceLostRequests.clear();
+}
+
+void WebGPUParent::MaintainDevices() {
+ ffi::wgpu_server_poll_all_devices(mContext.get(), false);
+}
+
+void WebGPUParent::LoseDevice(const RawId aDeviceId, Maybe<uint8_t> aReason,
+ const nsACString& aMessage) {
+ // Check to see if we've already sent a DeviceLost message to aDeviceId.
+ if (mLostDeviceIds.Contains(aDeviceId)) {
+ return;
+ }
+
+ // If the connection has been dropped, there is nobody to receive
+ // the DeviceLost message anyway.
+ if (CanSend()) {
+ if (!SendDeviceLost(aDeviceId, aReason, aMessage)) {
+ NS_ERROR("SendDeviceLost failed");
+ return;
+ }
+ }
+
+ mLostDeviceIds.Insert(aDeviceId);
+}
+
+bool WebGPUParent::ForwardError(const Maybe<RawId> aDeviceId,
+ ErrorBuffer& aError) {
+ if (auto error = aError.GetError()) {
+ // If this is a error has isDeviceLost true, then instead of reporting
+ // the error, we swallow it and call LoseDevice if we have an
+ // aDeviceID. This is to comply with the spec declaration in
+ // https://gpuweb.github.io/gpuweb/#lose-the-device
+ // "No errors are generated after device loss."
+ if (error->isDeviceLost) {
+ if (aDeviceId.isSome()) {
+ LoseDevice(*aDeviceId, Nothing(), error->message);
+ }
+ return false;
+ }
+ ReportError(aDeviceId, error->type, error->message);
+ return true;
+ }
+ return false;
+}
+
+// Generate an error on the Device timeline of aDeviceId.
+// aMessage is interpreted as UTF-8.
+void WebGPUParent::ReportError(const Maybe<RawId> aDeviceId,
+ const GPUErrorFilter aType,
+ const nsCString& aMessage) {
+ // find the appropriate error scope
+ if (aDeviceId) {
+ const auto& itr = mErrorScopeStackByDevice.find(*aDeviceId);
+ if (itr != mErrorScopeStackByDevice.end()) {
+ auto& stack = itr->second;
+ for (auto& scope : Reversed(stack)) {
+ if (scope.filter != aType) {
+ continue;
+ }
+ if (!scope.firstMessage) {
+ scope.firstMessage = Some(aMessage);
+ }
+ return;
+ }
+ }
+ }
+ // No error scope found, so fall back to the uncaptured error handler
+ if (!SendUncapturedError(aDeviceId, aMessage)) {
+ NS_ERROR("SendDeviceUncapturedError failed");
+ }
+}
+
+ipc::IPCResult WebGPUParent::RecvInstanceRequestAdapter(
+ const dom::GPURequestAdapterOptions& aOptions,
+ const nsTArray<RawId>& aTargetIds,
+ InstanceRequestAdapterResolver&& resolver) {
+ ffi::WGPURequestAdapterOptions options = {};
+ if (aOptions.mPowerPreference.WasPassed()) {
+ options.power_preference = static_cast<ffi::WGPUPowerPreference>(
+ aOptions.mPowerPreference.Value());
+ } else {
+ options.power_preference = ffi::WGPUPowerPreference_LowPower;
+ }
+ options.force_fallback_adapter = aOptions.mForceFallbackAdapter;
+
+ auto luid = GetCompositorDeviceLuid();
+
+ ErrorBuffer error;
+ int8_t index = ffi::wgpu_server_instance_request_adapter(
+ mContext.get(), &options, aTargetIds.Elements(), aTargetIds.Length(),
+ luid.ptrOr(nullptr), error.ToFFI());
+
+ ByteBuf infoByteBuf;
+ // Rust side expects an `Option`, so 0 maps to `None`.
+ uint64_t adapterId = 0;
+ if (index >= 0) {
+ adapterId = aTargetIds[index];
+ }
+ ffi::wgpu_server_adapter_pack_info(mContext.get(), adapterId,
+ ToFFI(&infoByteBuf));
+ resolver(std::move(infoByteBuf));
+ ForwardError(0, error);
+
+ // free the unused IDs
+ ipc::ByteBuf dropByteBuf;
+ for (size_t i = 0; i < aTargetIds.Length(); ++i) {
+ if (static_cast<int8_t>(i) != index) {
+ wgpu_server_adapter_free(aTargetIds[i], ToFFI(&dropByteBuf));
+ }
+ }
+ if (dropByteBuf.mData && !SendDropAction(std::move(dropByteBuf))) {
+ NS_ERROR("Unable to free free unused adapter IDs");
+ }
+ return IPC_OK();
+}
+
+/* static */ void WebGPUParent::DeviceLostCallback(uint8_t* aUserData,
+ uint8_t aReason,
+ const char* aMessage) {
+ DeviceLostRequest* req = reinterpret_cast<DeviceLostRequest*>(aUserData);
+ if (!req->mParent) {
+ // Parent is dead, never mind.
+ return;
+ }
+
+ RawId deviceId = req->mDeviceId;
+
+ // If aReason is 0, that corresponds to the "unknown" reason, which
+ // we treat as a Nothing() value. Any other value (which is positive)
+ // is mapped to the GPUDeviceLostReason values by subtracting 1.
+ Maybe<uint8_t> reason;
+ if (aReason > 0) {
+ uint8_t mappedReasonValue = (aReason - 1u);
+ reason = Some(mappedReasonValue);
+ }
+ nsAutoCString message(aMessage);
+ req->mParent->LoseDevice(deviceId, reason, message);
+
+ // We're no longer tracking the memory for this callback, so erase
+ // it to ensure we don't leak memory.
+ req->mParent->mDeviceLostRequests.erase(deviceId);
+}
+
+ipc::IPCResult WebGPUParent::RecvAdapterRequestDevice(
+ RawId aAdapterId, const ipc::ByteBuf& aByteBuf, RawId aDeviceId,
+ AdapterRequestDeviceResolver&& resolver) {
+ ErrorBuffer error;
+ ffi::wgpu_server_adapter_request_device(
+ mContext.get(), aAdapterId, ToFFI(&aByteBuf), aDeviceId, error.ToFFI());
+ if (ForwardError(0, error)) {
+ uint8_t reasonDestroyed = 0; // GPUDeviceLostReason::Destroyed
+ auto maybeError = error.GetError();
+ MOZ_ASSERT(maybeError.isSome());
+ LoseDevice(aDeviceId, Some(reasonDestroyed), maybeError->message);
+ resolver(false);
+ return IPC_OK();
+ }
+
+ mErrorScopeStackByDevice.insert({aDeviceId, {}});
+
+ // Setup the device lost callback.
+ std::unique_ptr<DeviceLostRequest> req(
+ new DeviceLostRequest{this, aDeviceId});
+ auto iter = mDeviceLostRequests.insert({aDeviceId, std::move(req)});
+ MOZ_ASSERT(iter.second, "Should be able to insert DeviceLostRequest.");
+ auto record = iter.first;
+ DeviceLostRequest* req_shadow = (record->second).get();
+ ffi::WGPUDeviceLostClosureC callback = {
+ &DeviceLostCallback, reinterpret_cast<uint8_t*>(req_shadow)};
+ ffi::wgpu_server_set_device_lost_callback(mContext.get(), aDeviceId,
+ callback);
+
+ resolver(true);
+
+#if defined(XP_WIN)
+ HANDLE handle =
+ wgpu_server_get_device_fence_handle(mContext.get(), aDeviceId);
+ if (handle) {
+ mFenceHandle = new gfx::FileHandleWrapper(UniqueFileHandle(handle));
+ }
+#endif
+
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvAdapterDrop(RawId aAdapterId) {
+ ffi::wgpu_server_adapter_drop(mContext.get(), aAdapterId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvDeviceDestroy(RawId aDeviceId) {
+ ffi::wgpu_server_device_destroy(mContext.get(), aDeviceId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvDeviceDrop(RawId aDeviceId) {
+ ffi::wgpu_server_device_drop(mContext.get(), aDeviceId);
+ MOZ_ASSERT(mDeviceLostRequests.find(aDeviceId) == mDeviceLostRequests.end(),
+ "DeviceLostRequest should have been invoked, then erased.");
+
+ mErrorScopeStackByDevice.erase(aDeviceId);
+ mLostDeviceIds.Remove(aDeviceId);
+ return IPC_OK();
+}
+
+WebGPUParent::BufferMapData* WebGPUParent::GetBufferMapData(RawId aBufferId) {
+ const auto iter = mSharedMemoryMap.find(aBufferId);
+ if (iter == mSharedMemoryMap.end()) {
+ return nullptr;
+ }
+
+ return &iter->second;
+}
+
+ipc::IPCResult WebGPUParent::RecvDeviceCreateBuffer(
+ RawId aDeviceId, RawId aBufferId, dom::GPUBufferDescriptor&& aDesc,
+ ipc::UnsafeSharedMemoryHandle&& aShmem) {
+ webgpu::StringHelper label(aDesc.mLabel);
+
+ auto shmem =
+ ipc::WritableSharedMemoryMapping::Open(std::move(aShmem)).value();
+
+ bool hasMapFlags = aDesc.mUsage & (dom::GPUBufferUsage_Binding::MAP_WRITE |
+ dom::GPUBufferUsage_Binding::MAP_READ);
+ bool shmAllocationFailed = false;
+ if (hasMapFlags || aDesc.mMappedAtCreation) {
+ if (shmem.Size() < aDesc.mSize) {
+ MOZ_RELEASE_ASSERT(shmem.Size() == 0);
+ // If we requested a non-zero mappable buffer and get a size of zero, it
+ // indicates that the shmem allocation failed on the client side.
+ shmAllocationFailed = true;
+ } else {
+ uint64_t offset = 0;
+ uint64_t size = 0;
+
+ if (aDesc.mMappedAtCreation) {
+ size = aDesc.mSize;
+ }
+
+ BufferMapData data = {std::move(shmem), hasMapFlags, offset, size,
+ aDeviceId};
+ mSharedMemoryMap.insert({aBufferId, std::move(data)});
+ }
+ }
+
+ ErrorBuffer error;
+ ffi::wgpu_server_device_create_buffer(mContext.get(), aDeviceId, aBufferId,
+ label.Get(), aDesc.mSize, aDesc.mUsage,
+ aDesc.mMappedAtCreation,
+ shmAllocationFailed, error.ToFFI());
+ ForwardError(aDeviceId, error);
+ return IPC_OK();
+}
+
+struct MapRequest {
+ RefPtr<WebGPUParent> mParent;
+ ffi::WGPUGlobal* mContext;
+ ffi::WGPUBufferId mBufferId;
+ ffi::WGPUHostMap mHostMap;
+ uint64_t mOffset;
+ uint64_t mSize;
+ WebGPUParent::BufferMapResolver mResolver;
+};
+
+static const char* MapStatusString(ffi::WGPUBufferMapAsyncStatus status) {
+ switch (status) {
+ case ffi::WGPUBufferMapAsyncStatus_Success:
+ return "Success";
+ case ffi::WGPUBufferMapAsyncStatus_AlreadyMapped:
+ return "Already mapped";
+ case ffi::WGPUBufferMapAsyncStatus_MapAlreadyPending:
+ return "Map is already pending";
+ case ffi::WGPUBufferMapAsyncStatus_Aborted:
+ return "Map aborted";
+ case ffi::WGPUBufferMapAsyncStatus_ContextLost:
+ return "Context lost";
+ case ffi::WGPUBufferMapAsyncStatus_Invalid:
+ return "Invalid buffer";
+ case ffi::WGPUBufferMapAsyncStatus_InvalidRange:
+ return "Invalid range";
+ case ffi::WGPUBufferMapAsyncStatus_InvalidAlignment:
+ return "Invalid alignment";
+ case ffi::WGPUBufferMapAsyncStatus_InvalidUsageFlags:
+ return "Invalid usage flags";
+ case ffi::WGPUBufferMapAsyncStatus_Error:
+ return "Map failed";
+ case ffi::WGPUBufferMapAsyncStatus_Sentinel: // For -Wswitch
+ break;
+ }
+
+ MOZ_CRASH("Bad ffi::WGPUBufferMapAsyncStatus");
+}
+
+void WebGPUParent::MapCallback(ffi::WGPUBufferMapAsyncStatus aStatus,
+ uint8_t* aUserData) {
+ auto* req = reinterpret_cast<MapRequest*>(aUserData);
+
+ if (!req->mParent->CanSend()) {
+ delete req;
+ return;
+ }
+
+ BufferMapResult result;
+
+ auto bufferId = req->mBufferId;
+ auto* mapData = req->mParent->GetBufferMapData(bufferId);
+ MOZ_RELEASE_ASSERT(mapData);
+
+ if (aStatus != ffi::WGPUBufferMapAsyncStatus_Success) {
+ // A buffer map operation that fails with a DeviceError gets
+ // mapped to the ContextLost status. If we have this status, we
+ // need to lose the device.
+ if (aStatus == ffi::WGPUBufferMapAsyncStatus_ContextLost) {
+ req->mParent->LoseDevice(
+ mapData->mDeviceId, Nothing(),
+ nsPrintfCString("Buffer %" PRIu64 " invalid", bufferId));
+ }
+
+ result = BufferMapError(nsPrintfCString("Mapping WebGPU buffer failed: %s",
+ MapStatusString(aStatus)));
+ } else {
+ auto size = req->mSize;
+ auto offset = req->mOffset;
+
+ if (req->mHostMap == ffi::WGPUHostMap_Read && size > 0) {
+ ErrorBuffer error;
+ const auto src = ffi::wgpu_server_buffer_get_mapped_range(
+ req->mContext, req->mBufferId, offset, size, error.ToFFI());
+
+ MOZ_RELEASE_ASSERT(!error.GetError());
+
+ MOZ_RELEASE_ASSERT(mapData->mShmem.Size() >= offset + size);
+ if (src.ptr != nullptr && src.length >= size) {
+ auto dst = mapData->mShmem.Bytes().Subspan(offset, size);
+ memcpy(dst.data(), src.ptr, size);
+ }
+ }
+
+ result =
+ BufferMapSuccess(offset, size, req->mHostMap == ffi::WGPUHostMap_Write);
+
+ mapData->mMappedOffset = offset;
+ mapData->mMappedSize = size;
+ }
+
+ req->mResolver(std::move(result));
+ delete req;
+}
+
+ipc::IPCResult WebGPUParent::RecvBufferMap(RawId aDeviceId, RawId aBufferId,
+ uint32_t aMode, uint64_t aOffset,
+ uint64_t aSize,
+ BufferMapResolver&& aResolver) {
+ MOZ_LOG(sLogger, LogLevel::Info,
+ ("RecvBufferMap %" PRIu64 " offset=%" PRIu64 " size=%" PRIu64 "\n",
+ aBufferId, aOffset, aSize));
+
+ ffi::WGPUHostMap mode;
+ switch (aMode) {
+ case dom::GPUMapMode_Binding::READ:
+ mode = ffi::WGPUHostMap_Read;
+ break;
+ case dom::GPUMapMode_Binding::WRITE:
+ mode = ffi::WGPUHostMap_Write;
+ break;
+ default: {
+ nsCString errorString(
+ "GPUBuffer.mapAsync 'mode' argument must be either GPUMapMode.READ "
+ "or GPUMapMode.WRITE");
+ aResolver(BufferMapError(errorString));
+ return IPC_OK();
+ }
+ }
+
+ auto* mapData = GetBufferMapData(aBufferId);
+
+ if (!mapData) {
+ nsCString errorString("Buffer is not mappable");
+ aResolver(BufferMapError(errorString));
+ return IPC_OK();
+ }
+
+ auto* request =
+ new MapRequest{this, mContext.get(), aBufferId, mode,
+ aOffset, aSize, std::move(aResolver)};
+
+ ffi::WGPUBufferMapCallbackC callback = {&MapCallback,
+ reinterpret_cast<uint8_t*>(request)};
+ ErrorBuffer mapError;
+ ffi::wgpu_server_buffer_map(mContext.get(), aBufferId, aOffset, aSize, mode,
+ callback, mapError.ToFFI());
+ ForwardError(aDeviceId, mapError);
+
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvBufferUnmap(RawId aDeviceId, RawId aBufferId,
+ bool aFlush) {
+ MOZ_LOG(sLogger, LogLevel::Info,
+ ("RecvBufferUnmap %" PRIu64 " flush=%d\n", aBufferId, aFlush));
+
+ auto* mapData = GetBufferMapData(aBufferId);
+
+ if (mapData && aFlush) {
+ uint64_t offset = mapData->mMappedOffset;
+ uint64_t size = mapData->mMappedSize;
+
+ ErrorBuffer getRangeError;
+ const auto mapped = ffi::wgpu_server_buffer_get_mapped_range(
+ mContext.get(), aBufferId, offset, size, getRangeError.ToFFI());
+ ForwardError(aDeviceId, getRangeError);
+
+ if (mapped.ptr != nullptr && mapped.length >= size) {
+ auto shmSize = mapData->mShmem.Size();
+ MOZ_RELEASE_ASSERT(offset <= shmSize);
+ MOZ_RELEASE_ASSERT(size <= shmSize - offset);
+
+ auto src = mapData->mShmem.Bytes().Subspan(offset, size);
+ memcpy(mapped.ptr, src.data(), size);
+ }
+
+ mapData->mMappedOffset = 0;
+ mapData->mMappedSize = 0;
+ }
+
+ ErrorBuffer unmapError;
+ ffi::wgpu_server_buffer_unmap(mContext.get(), aBufferId, unmapError.ToFFI());
+ ForwardError(aDeviceId, unmapError);
+
+ if (mapData && !mapData->mHasMapFlags) {
+ // We get here if the buffer was mapped at creation without map flags.
+ // We don't need the shared memory anymore.
+ DeallocBufferShmem(aBufferId);
+ }
+
+ return IPC_OK();
+}
+
+void WebGPUParent::DeallocBufferShmem(RawId aBufferId) {
+ const auto iter = mSharedMemoryMap.find(aBufferId);
+ if (iter != mSharedMemoryMap.end()) {
+ mSharedMemoryMap.erase(iter);
+ }
+}
+
+ipc::IPCResult WebGPUParent::RecvBufferDrop(RawId aBufferId) {
+ ffi::wgpu_server_buffer_drop(mContext.get(), aBufferId);
+ MOZ_LOG(sLogger, LogLevel::Info, ("RecvBufferDrop %" PRIu64 "\n", aBufferId));
+
+ DeallocBufferShmem(aBufferId);
+
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvBufferDestroy(RawId aBufferId) {
+ ffi::wgpu_server_buffer_destroy(mContext.get(), aBufferId);
+ MOZ_LOG(sLogger, LogLevel::Info,
+ ("RecvBufferDestroy %" PRIu64 "\n", aBufferId));
+
+ DeallocBufferShmem(aBufferId);
+
+ return IPC_OK();
+}
+
+void WebGPUParent::RemoveExternalTexture(RawId aTextureId) {
+ auto it = mExternalTextures.find(aTextureId);
+ if (it != mExternalTextures.end()) {
+ mExternalTextures.erase(it);
+ }
+}
+
+ipc::IPCResult WebGPUParent::RecvTextureDestroy(RawId aTextureId,
+ RawId aDeviceId) {
+ ffi::wgpu_server_texture_destroy(mContext.get(), aTextureId);
+ RemoveExternalTexture(aTextureId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvTextureDrop(RawId aTextureId) {
+ ffi::wgpu_server_texture_drop(mContext.get(), aTextureId);
+ RemoveExternalTexture(aTextureId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvTextureViewDrop(RawId aTextureViewId) {
+ ffi::wgpu_server_texture_view_drop(mContext.get(), aTextureViewId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvSamplerDrop(RawId aSamplerId) {
+ ffi::wgpu_server_sampler_drop(mContext.get(), aSamplerId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvCommandEncoderFinish(
+ RawId aEncoderId, RawId aDeviceId,
+ const dom::GPUCommandBufferDescriptor& aDesc) {
+ Unused << aDesc;
+ ffi::WGPUCommandBufferDescriptor desc = {};
+
+ webgpu::StringHelper label(aDesc.mLabel);
+ desc.label = label.Get();
+
+ ErrorBuffer error;
+ ffi::wgpu_server_encoder_finish(mContext.get(), aEncoderId, &desc,
+ error.ToFFI());
+
+ ForwardError(aDeviceId, error);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvCommandEncoderDrop(RawId aEncoderId) {
+ ffi::wgpu_server_encoder_drop(mContext.get(), aEncoderId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvRenderBundleDrop(RawId aBundleId) {
+ ffi::wgpu_server_render_bundle_drop(mContext.get(), aBundleId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvQueueSubmit(
+ RawId aQueueId, RawId aDeviceId, const nsTArray<RawId>& aCommandBuffers,
+ const nsTArray<RawId>& aTextureIds) {
+ ErrorBuffer error;
+ auto index = ffi::wgpu_server_queue_submit(
+ mContext.get(), aQueueId, aCommandBuffers.Elements(),
+ aCommandBuffers.Length(), error.ToFFI());
+ // Check if index is valid. 0 means error.
+ if (index != 0) {
+ for (const auto& textureId : aTextureIds) {
+ auto it = mExternalTextures.find(textureId);
+ if (it != mExternalTextures.end()) {
+ auto& externalTexture = it->second;
+
+ externalTexture->SetSubmissionIndex(index);
+ }
+ }
+ }
+ ForwardError(aDeviceId, error);
+ return IPC_OK();
+}
+
+struct OnSubmittedWorkDoneRequest {
+ RefPtr<WebGPUParent> mParent;
+ WebGPUParent::QueueOnSubmittedWorkDoneResolver mResolver;
+};
+
+void OnSubmittedWorkDoneCallback(uint8_t* userdata) {
+ auto req = std::unique_ptr<OnSubmittedWorkDoneRequest>(
+ reinterpret_cast<OnSubmittedWorkDoneRequest*>(userdata));
+ if (req->mParent->CanSend()) {
+ req->mResolver(void_t());
+ }
+}
+
+ipc::IPCResult WebGPUParent::RecvQueueOnSubmittedWorkDone(
+ RawId aQueueId, std::function<void(mozilla::void_t)>&& aResolver) {
+ std::unique_ptr<OnSubmittedWorkDoneRequest> request(
+ new OnSubmittedWorkDoneRequest{this, std::move(aResolver)});
+
+ ffi::WGPUSubmittedWorkDoneClosureC callback = {
+ &OnSubmittedWorkDoneCallback,
+ reinterpret_cast<uint8_t*>(request.release())};
+ ffi::wgpu_server_on_submitted_work_done(mContext.get(), aQueueId, callback);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvQueueWriteAction(
+ RawId aQueueId, RawId aDeviceId, const ipc::ByteBuf& aByteBuf,
+ ipc::UnsafeSharedMemoryHandle&& aShmem) {
+ auto mapping =
+ ipc::WritableSharedMemoryMapping::Open(std::move(aShmem)).value();
+
+ ErrorBuffer error;
+ ffi::wgpu_server_queue_write_action(mContext.get(), aQueueId,
+ ToFFI(&aByteBuf), mapping.Bytes().data(),
+ mapping.Size(), error.ToFFI());
+ ForwardError(aDeviceId, error);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvBindGroupLayoutDrop(RawId aBindGroupId) {
+ ffi::wgpu_server_bind_group_layout_drop(mContext.get(), aBindGroupId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvPipelineLayoutDrop(RawId aLayoutId) {
+ ffi::wgpu_server_pipeline_layout_drop(mContext.get(), aLayoutId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvBindGroupDrop(RawId aBindGroupId) {
+ ffi::wgpu_server_bind_group_drop(mContext.get(), aBindGroupId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvShaderModuleDrop(RawId aModuleId) {
+ ffi::wgpu_server_shader_module_drop(mContext.get(), aModuleId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvComputePipelineDrop(RawId aPipelineId) {
+ ffi::wgpu_server_compute_pipeline_drop(mContext.get(), aPipelineId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvRenderPipelineDrop(RawId aPipelineId) {
+ ffi::wgpu_server_render_pipeline_drop(mContext.get(), aPipelineId);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvImplicitLayoutDrop(
+ RawId aImplicitPlId, const nsTArray<RawId>& aImplicitBglIds) {
+ ffi::wgpu_server_pipeline_layout_drop(mContext.get(), aImplicitPlId);
+ for (const auto& id : aImplicitBglIds) {
+ ffi::wgpu_server_bind_group_layout_drop(mContext.get(), id);
+ }
+ return IPC_OK();
+}
+
+// TODO: proper destruction
+
+ipc::IPCResult WebGPUParent::RecvDeviceCreateSwapChain(
+ RawId aDeviceId, RawId aQueueId, const RGBDescriptor& aDesc,
+ const nsTArray<RawId>& aBufferIds,
+ const layers::RemoteTextureOwnerId& aOwnerId,
+ bool aUseExternalTextureInSwapChain) {
+ switch (aDesc.format()) {
+ case gfx::SurfaceFormat::R8G8B8A8:
+ case gfx::SurfaceFormat::B8G8R8A8:
+ break;
+ default:
+ MOZ_ASSERT_UNREACHABLE("Invalid surface format!");
+ return IPC_OK();
+ }
+
+ const auto bufferStrideWithMask =
+ Device::BufferStrideWithMask(aDesc.size(), aDesc.format());
+ if (!bufferStrideWithMask.isValid()) {
+ MOZ_ASSERT_UNREACHABLE("Invalid width / buffer stride!");
+ return IPC_OK();
+ }
+
+ constexpr uint32_t kBufferAlignmentMask = 0xff;
+ const uint32_t bufferStride =
+ bufferStrideWithMask.value() & ~kBufferAlignmentMask;
+
+ const auto rows = CheckedInt<uint32_t>(aDesc.size().height);
+ if (!rows.isValid()) {
+ MOZ_ASSERT_UNREACHABLE("Invalid height!");
+ return IPC_OK();
+ }
+
+ if (!mRemoteTextureOwner) {
+ mRemoteTextureOwner =
+ MakeRefPtr<layers::RemoteTextureOwnerClient>(OtherPid());
+ }
+ mRemoteTextureOwner->RegisterTextureOwner(aOwnerId);
+
+ auto data = MakeRefPtr<PresentationData>(this, aUseExternalTextureInSwapChain,
+ aDeviceId, aQueueId, aDesc,
+ bufferStride, aBufferIds);
+ if (!mPresentationDataMap.emplace(aOwnerId, data).second) {
+ NS_ERROR("External image is already registered as WebGPU canvas!");
+ }
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvDeviceCreateShaderModule(
+ RawId aDeviceId, RawId aModuleId, const nsString& aLabel,
+ const nsCString& aCode, DeviceCreateShaderModuleResolver&& aOutMessage) {
+ // TODO: this should probably be an optional label in the IPC message.
+ const nsACString* label = nullptr;
+ NS_ConvertUTF16toUTF8 utf8Label(aLabel);
+ if (!utf8Label.IsEmpty()) {
+ label = &utf8Label;
+ }
+
+ ffi::WGPUShaderModuleCompilationMessage message;
+ ErrorBuffer error;
+
+ bool ok = ffi::wgpu_server_device_create_shader_module(
+ mContext.get(), aDeviceId, aModuleId, label, &aCode, &message,
+ error.ToFFI());
+
+ ForwardError(aDeviceId, error);
+
+ nsTArray<WebGPUCompilationMessage> messages;
+
+ if (!ok) {
+ WebGPUCompilationMessage msg;
+ msg.lineNum = message.line_number;
+ msg.linePos = message.line_pos;
+ msg.offset = message.utf16_offset;
+ msg.length = message.utf16_length;
+ msg.message = message.message;
+ // wgpu currently only returns errors.
+ msg.messageType = WebGPUCompilationMessageType::Error;
+
+ messages.AppendElement(msg);
+ }
+
+ aOutMessage(messages);
+
+ return IPC_OK();
+}
+
+struct ReadbackPresentRequest {
+ ReadbackPresentRequest(
+ const ffi::WGPUGlobal* aContext, RefPtr<PresentationData>& aData,
+ RefPtr<layers::RemoteTextureOwnerClient>& aRemoteTextureOwner,
+ const layers::RemoteTextureId aTextureId,
+ const layers::RemoteTextureOwnerId aOwnerId)
+ : mContext(aContext),
+ mData(aData),
+ mRemoteTextureOwner(aRemoteTextureOwner),
+ mTextureId(aTextureId),
+ mOwnerId(aOwnerId) {}
+
+ const ffi::WGPUGlobal* mContext;
+ RefPtr<PresentationData> mData;
+ RefPtr<layers::RemoteTextureOwnerClient> mRemoteTextureOwner;
+ const layers::RemoteTextureId mTextureId;
+ const layers::RemoteTextureOwnerId mOwnerId;
+};
+
+static void ReadbackPresentCallback(ffi::WGPUBufferMapAsyncStatus status,
+ uint8_t* userdata) {
+ UniquePtr<ReadbackPresentRequest> req(
+ reinterpret_cast<ReadbackPresentRequest*>(userdata));
+
+ const auto onExit = mozilla::MakeScopeExit([&]() {
+ auto& waitingTextures = req->mData->mWaitingReadbackTexturesForPresent;
+ auto it = waitingTextures.find(req->mTextureId);
+ MOZ_ASSERT(it != waitingTextures.end());
+ if (it != waitingTextures.end()) {
+ waitingTextures.erase(it);
+ }
+ if (req->mData->mPendingSwapChainDrop.isSome() && waitingTextures.empty()) {
+ if (req->mData->mParent) {
+ auto& pendingDrop = req->mData->mPendingSwapChainDrop.ref();
+ req->mData->mParent->RecvSwapChainDrop(
+ req->mOwnerId, pendingDrop.mTxnType, pendingDrop.mTxnId);
+ req->mData->mPendingSwapChainDrop = Nothing();
+ }
+ }
+ });
+
+ if (!req->mRemoteTextureOwner->IsRegistered(req->mOwnerId)) {
+ // SwapChain is already Destroyed
+ return;
+ }
+
+ PresentationData* data = req->mData.get();
+ // get the buffer ID
+ RawId bufferId;
+ {
+ MutexAutoLock lock(data->mBuffersLock);
+ bufferId = data->mQueuedBufferIds.back();
+ data->mQueuedBufferIds.pop_back();
+ }
+
+ // Ensure we'll make the bufferId available for reuse
+ auto releaseBuffer = MakeScopeExit([data = RefPtr{data}, bufferId] {
+ MutexAutoLock lock(data->mBuffersLock);
+ data->mAvailableBufferIds.push_back(bufferId);
+ });
+
+ MOZ_LOG(sLogger, LogLevel::Info,
+ ("ReadbackPresentCallback for buffer %" PRIu64 " status=%d\n",
+ bufferId, status));
+ // copy the data
+ if (status == ffi::WGPUBufferMapAsyncStatus_Success) {
+ const auto bufferSize = data->mDesc.size().height * data->mSourcePitch;
+ ErrorBuffer getRangeError;
+ const auto mapped = ffi::wgpu_server_buffer_get_mapped_range(
+ req->mContext, bufferId, 0, bufferSize, getRangeError.ToFFI());
+ if (req->mData->mParent) {
+ req->mData->mParent->ForwardError(data->mDeviceId, getRangeError);
+ } else if (auto innerError = getRangeError.GetError()) {
+ // If an error occured in get_mapped_range, treat it as an internal error
+ // and crash. The error handling story for something unexpected happening
+ // during the present glue needs to befigured out in a more global way.
+ MOZ_LOG(sLogger, LogLevel::Info,
+ ("WebGPU present: buffer get_mapped_range failed: %s\n",
+ innerError->message.get()));
+ }
+
+ MOZ_RELEASE_ASSERT(mapped.length >= bufferSize);
+ auto textureData =
+ req->mRemoteTextureOwner->CreateOrRecycleBufferTextureData(
+ data->mDesc.size(), data->mDesc.format(), req->mOwnerId);
+ if (!textureData) {
+ gfxCriticalNoteOnce << "Failed to allocate BufferTextureData";
+ return;
+ }
+ layers::MappedTextureData mappedData;
+ if (textureData && textureData->BorrowMappedData(mappedData)) {
+ uint8_t* src = mapped.ptr;
+ uint8_t* dst = mappedData.data;
+ for (auto row = 0; row < data->mDesc.size().height; ++row) {
+ memcpy(dst, src, mappedData.stride);
+ dst += mappedData.stride;
+ src += data->mSourcePitch;
+ }
+ req->mRemoteTextureOwner->PushTexture(req->mTextureId, req->mOwnerId,
+ std::move(textureData));
+ } else {
+ NS_WARNING("WebGPU present skipped: the swapchain is resized!");
+ }
+ ErrorBuffer unmapError;
+ wgpu_server_buffer_unmap(req->mContext, bufferId, unmapError.ToFFI());
+ if (req->mData->mParent) {
+ req->mData->mParent->ForwardError(data->mDeviceId, unmapError);
+ } else if (auto innerError = unmapError.GetError()) {
+ MOZ_LOG(sLogger, LogLevel::Info,
+ ("WebGPU present: buffer unmap failed: %s\n",
+ innerError->message.get()));
+ }
+ } else {
+ // TODO: better handle errors
+ NS_WARNING("WebGPU frame mapping failed!");
+ }
+}
+
+ipc::IPCResult WebGPUParent::GetFrontBufferSnapshot(
+ IProtocol* aProtocol, const layers::RemoteTextureOwnerId& aOwnerId,
+ Maybe<Shmem>& aShmem, gfx::IntSize& aSize) {
+ const auto& lookup = mPresentationDataMap.find(aOwnerId);
+ if (lookup == mPresentationDataMap.end() || !mRemoteTextureOwner ||
+ !mRemoteTextureOwner->IsRegistered(aOwnerId)) {
+ return IPC_OK();
+ }
+
+ RefPtr<PresentationData> data = lookup->second.get();
+ aSize = data->mDesc.size();
+ uint32_t stride = layers::ImageDataSerializer::ComputeRGBStride(
+ data->mDesc.format(), aSize.width);
+ uint32_t len = data->mDesc.size().height * stride;
+ Shmem shmem;
+ if (!AllocShmem(len, &shmem)) {
+ return IPC_OK();
+ }
+
+ mRemoteTextureOwner->GetLatestBufferSnapshot(aOwnerId, shmem, aSize);
+ aShmem.emplace(std::move(shmem));
+
+ return IPC_OK();
+}
+
+void WebGPUParent::PostExternalTexture(
+ const std::shared_ptr<ExternalTexture>&& aExternalTexture,
+ const layers::RemoteTextureId aRemoteTextureId,
+ const layers::RemoteTextureOwnerId aOwnerId) {
+ const auto& lookup = mPresentationDataMap.find(aOwnerId);
+ if (lookup == mPresentationDataMap.end() || !mRemoteTextureOwner ||
+ !mRemoteTextureOwner->IsRegistered(aOwnerId)) {
+ NS_WARNING("WebGPU presenting on a destroyed swap chain!");
+ return;
+ }
+
+ const auto surfaceFormat = gfx::SurfaceFormat::B8G8R8A8;
+ const auto size = aExternalTexture->GetSize();
+ const auto index = aExternalTexture->GetSubmissionIndex();
+ MOZ_ASSERT(index != 0);
+
+ Maybe<gfx::FenceInfo> fenceInfo;
+ if (mFenceHandle) {
+ fenceInfo = Some(gfx::FenceInfo(mFenceHandle, index));
+ }
+
+ Maybe<layers::SurfaceDescriptor> desc =
+ aExternalTexture->ToSurfaceDescriptor(fenceInfo);
+ if (!desc) {
+ MOZ_ASSERT_UNREACHABLE("unexpected to be called");
+ return;
+ }
+
+ mRemoteTextureOwner->PushTexture(aRemoteTextureId, aOwnerId, aExternalTexture,
+ size, surfaceFormat, *desc);
+
+ RefPtr<PresentationData> data = lookup->second.get();
+
+ auto recycledTexture = mRemoteTextureOwner->GetRecycledExternalTexture(
+ size, surfaceFormat, desc->type(), aOwnerId);
+ if (recycledTexture) {
+ data->mRecycledExternalTextures.push_back(recycledTexture);
+ }
+}
+
+ipc::IPCResult WebGPUParent::RecvSwapChainPresent(
+ RawId aTextureId, RawId aCommandEncoderId,
+ const layers::RemoteTextureId& aRemoteTextureId,
+ const layers::RemoteTextureOwnerId& aOwnerId) {
+ // step 0: get the data associated with the swapchain
+ const auto& lookup = mPresentationDataMap.find(aOwnerId);
+ if (lookup == mPresentationDataMap.end() || !mRemoteTextureOwner ||
+ !mRemoteTextureOwner->IsRegistered(aOwnerId)) {
+ NS_WARNING("WebGPU presenting on a destroyed swap chain!");
+ return IPC_OK();
+ }
+
+ RefPtr<PresentationData> data = lookup->second.get();
+
+ if (data->mUseExternalTextureInSwapChain) {
+ auto it = mExternalTextures.find(aTextureId);
+ if (it == mExternalTextures.end()) {
+ MOZ_ASSERT_UNREACHABLE("unexpected to be called");
+ return IPC_OK();
+ }
+ std::shared_ptr<ExternalTexture> externalTexture = it->second;
+ mExternalTextures.erase(it);
+
+ PostExternalTexture(std::move(externalTexture), aRemoteTextureId, aOwnerId);
+ return IPC_OK();
+ }
+
+ RawId bufferId = 0;
+ const auto& size = data->mDesc.size();
+ const auto bufferSize = data->mDesc.size().height * data->mSourcePitch;
+
+ // step 1: find an available staging buffer, or create one
+ {
+ MutexAutoLock lock(data->mBuffersLock);
+ if (!data->mAvailableBufferIds.empty()) {
+ bufferId = data->mAvailableBufferIds.back();
+ data->mAvailableBufferIds.pop_back();
+ } else if (!data->mUnassignedBufferIds.empty()) {
+ bufferId = data->mUnassignedBufferIds.back();
+ data->mUnassignedBufferIds.pop_back();
+
+ ffi::WGPUBufferUsages usage =
+ WGPUBufferUsages_COPY_DST | WGPUBufferUsages_MAP_READ;
+
+ ErrorBuffer error;
+ ffi::wgpu_server_device_create_buffer(mContext.get(), data->mDeviceId,
+ bufferId, nullptr, bufferSize,
+ usage, false, false, error.ToFFI());
+ if (ForwardError(data->mDeviceId, error)) {
+ return IPC_OK();
+ }
+ } else {
+ bufferId = 0;
+ }
+
+ if (bufferId) {
+ data->mQueuedBufferIds.insert(data->mQueuedBufferIds.begin(), bufferId);
+ }
+ }
+
+ MOZ_LOG(sLogger, LogLevel::Info,
+ ("RecvSwapChainPresent with buffer %" PRIu64 "\n", bufferId));
+ if (!bufferId) {
+ // TODO: add a warning - no buffer are available!
+ return IPC_OK();
+ }
+
+ // step 3: submit a copy command for the frame
+ ffi::WGPUCommandEncoderDescriptor encoderDesc = {};
+ {
+ ErrorBuffer error;
+ ffi::wgpu_server_device_create_encoder(mContext.get(), data->mDeviceId,
+ &encoderDesc, aCommandEncoderId,
+ error.ToFFI());
+ if (ForwardError(data->mDeviceId, error)) {
+ return IPC_OK();
+ }
+ }
+
+ const ffi::WGPUImageCopyTexture texView = {
+ aTextureId,
+ };
+ const ffi::WGPUImageDataLayout bufLayout = {
+ 0,
+ &data->mSourcePitch,
+ nullptr,
+ };
+ const ffi::WGPUExtent3d extent = {
+ static_cast<uint32_t>(size.width),
+ static_cast<uint32_t>(size.height),
+ 1,
+ };
+
+ {
+ ErrorBuffer error;
+ ffi::wgpu_server_encoder_copy_texture_to_buffer(
+ mContext.get(), aCommandEncoderId, &texView, bufferId, &bufLayout,
+ &extent, error.ToFFI());
+ if (ForwardError(data->mDeviceId, error)) {
+ return IPC_OK();
+ }
+ }
+ ffi::WGPUCommandBufferDescriptor commandDesc = {};
+ {
+ ErrorBuffer error;
+ ffi::wgpu_server_encoder_finish(mContext.get(), aCommandEncoderId,
+ &commandDesc, error.ToFFI());
+ if (ForwardError(data->mDeviceId, error)) {
+ return IPC_OK();
+ }
+ }
+
+ {
+ ErrorBuffer error;
+ ffi::wgpu_server_queue_submit(mContext.get(), data->mQueueId,
+ &aCommandEncoderId, 1, error.ToFFI());
+ if (ForwardError(data->mDeviceId, error)) {
+ return IPC_OK();
+ }
+ }
+
+ auto& waitingTextures = data->mWaitingReadbackTexturesForPresent;
+ auto it = waitingTextures.find(aRemoteTextureId);
+ MOZ_ASSERT(it == waitingTextures.end());
+ if (it == waitingTextures.end()) {
+ waitingTextures.emplace(aRemoteTextureId);
+ }
+
+ // step 4: request the pixels to be copied into the external texture
+ // TODO: this isn't strictly necessary. When WR wants to Lock() the external
+ // texture,
+ // we can just give it the contents of the last mapped buffer instead of the
+ // copy.
+ auto presentRequest = MakeUnique<ReadbackPresentRequest>(
+ mContext.get(), data, mRemoteTextureOwner, aRemoteTextureId, aOwnerId);
+
+ ffi::WGPUBufferMapCallbackC callback = {
+ &ReadbackPresentCallback,
+ reinterpret_cast<uint8_t*>(presentRequest.release())};
+
+ ErrorBuffer error;
+ ffi::wgpu_server_buffer_map(mContext.get(), bufferId, 0, bufferSize,
+ ffi::WGPUHostMap_Read, callback, error.ToFFI());
+ if (ForwardError(data->mDeviceId, error)) {
+ return IPC_OK();
+ }
+
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvSwapChainDrop(
+ const layers::RemoteTextureOwnerId& aOwnerId,
+ layers::RemoteTextureTxnType aTxnType, layers::RemoteTextureTxnId aTxnId) {
+ const auto& lookup = mPresentationDataMap.find(aOwnerId);
+ MOZ_ASSERT(lookup != mPresentationDataMap.end());
+ if (lookup == mPresentationDataMap.end()) {
+ NS_WARNING("WebGPU presenting on a destroyed swap chain!");
+ return IPC_OK();
+ }
+
+ RefPtr<PresentationData> data = lookup->second.get();
+
+ auto waitingCount = data->mWaitingReadbackTexturesForPresent.size();
+ if (waitingCount > 0) {
+ // Defer SwapChainDrop until readback complete
+ data->mPendingSwapChainDrop = Some(PendingSwapChainDrop{aTxnType, aTxnId});
+ return IPC_OK();
+ }
+
+ if (mRemoteTextureOwner) {
+ if (aTxnType && aTxnId) {
+ mRemoteTextureOwner->WaitForTxn(aOwnerId, aTxnType, aTxnId);
+ }
+ mRemoteTextureOwner->UnregisterTextureOwner(aOwnerId);
+ }
+
+ mPresentationDataMap.erase(lookup);
+
+ MutexAutoLock lock(data->mBuffersLock);
+ ipc::ByteBuf dropByteBuf;
+ for (const auto bid : data->mUnassignedBufferIds) {
+ wgpu_server_buffer_free(bid, ToFFI(&dropByteBuf));
+ }
+ if (dropByteBuf.mData && !SendDropAction(std::move(dropByteBuf))) {
+ NS_WARNING("Unable to free an ID for non-assigned buffer");
+ }
+ for (const auto bid : data->mAvailableBufferIds) {
+ ffi::wgpu_server_buffer_drop(mContext.get(), bid);
+ }
+ for (const auto bid : data->mQueuedBufferIds) {
+ ffi::wgpu_server_buffer_drop(mContext.get(), bid);
+ }
+ return IPC_OK();
+}
+
+void WebGPUParent::ActorDestroy(ActorDestroyReason aWhy) {
+ mTimer.Stop();
+ mPresentationDataMap.clear();
+ if (mRemoteTextureOwner) {
+ mRemoteTextureOwner->UnregisterAllTextureOwners();
+ mRemoteTextureOwner = nullptr;
+ }
+ ffi::wgpu_server_poll_all_devices(mContext.get(), true);
+ mContext = nullptr;
+}
+
+ipc::IPCResult WebGPUParent::RecvDeviceAction(RawId aDeviceId,
+ const ipc::ByteBuf& aByteBuf) {
+ ErrorBuffer error;
+ ffi::wgpu_server_device_action(mContext.get(), aDeviceId, ToFFI(&aByteBuf),
+ error.ToFFI());
+
+ ForwardError(aDeviceId, error);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvDeviceActionWithAck(
+ RawId aDeviceId, const ipc::ByteBuf& aByteBuf,
+ DeviceActionWithAckResolver&& aResolver) {
+ auto result = RecvDeviceAction(aDeviceId, aByteBuf);
+ aResolver(true);
+ return result;
+}
+
+ipc::IPCResult WebGPUParent::RecvTextureAction(RawId aTextureId,
+ RawId aDeviceId,
+ const ipc::ByteBuf& aByteBuf) {
+ ErrorBuffer error;
+ ffi::wgpu_server_texture_action(mContext.get(), aTextureId, ToFFI(&aByteBuf),
+ error.ToFFI());
+
+ ForwardError(aDeviceId, error);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvCommandEncoderAction(
+ RawId aEncoderId, RawId aDeviceId, const ipc::ByteBuf& aByteBuf) {
+ ErrorBuffer error;
+ ffi::wgpu_server_command_encoder_action(mContext.get(), aEncoderId,
+ ToFFI(&aByteBuf), error.ToFFI());
+ ForwardError(aDeviceId, error);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvBumpImplicitBindGroupLayout(RawId aPipelineId,
+ bool aIsCompute,
+ uint32_t aIndex,
+ RawId aAssignId) {
+ ErrorBuffer error;
+ if (aIsCompute) {
+ ffi::wgpu_server_compute_pipeline_get_bind_group_layout(
+ mContext.get(), aPipelineId, aIndex, aAssignId, error.ToFFI());
+ } else {
+ ffi::wgpu_server_render_pipeline_get_bind_group_layout(
+ mContext.get(), aPipelineId, aIndex, aAssignId, error.ToFFI());
+ }
+
+ ForwardError(0, error);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvDevicePushErrorScope(
+ RawId aDeviceId, const dom::GPUErrorFilter aFilter) {
+ const auto& itr = mErrorScopeStackByDevice.find(aDeviceId);
+ if (itr == mErrorScopeStackByDevice.end()) {
+ // Content can cause this simply by destroying a device and then
+ // calling `pushErrorScope`.
+ return IPC_OK();
+ }
+ auto& stack = itr->second;
+
+ // Let's prevent `while (true) { pushErrorScope(); }`.
+ constexpr size_t MAX_ERROR_SCOPE_STACK_SIZE = 1'000'000;
+ if (stack.size() >= MAX_ERROR_SCOPE_STACK_SIZE) {
+ nsPrintfCString m("pushErrorScope: Hit MAX_ERROR_SCOPE_STACK_SIZE of %zu",
+ MAX_ERROR_SCOPE_STACK_SIZE);
+ ReportError(Some(aDeviceId), dom::GPUErrorFilter::Out_of_memory, m);
+ return IPC_OK();
+ }
+
+ const auto newScope = ErrorScope{aFilter};
+ stack.push_back(newScope);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvDevicePopErrorScope(
+ RawId aDeviceId, DevicePopErrorScopeResolver&& aResolver) {
+ const auto popResult = [&]() {
+ const auto& itr = mErrorScopeStackByDevice.find(aDeviceId);
+ if (itr == mErrorScopeStackByDevice.end()) {
+ // Content can cause this simply by destroying a device and then
+ // calling `popErrorScope`.
+ return PopErrorScopeResult{PopErrorScopeResultType::DeviceLost};
+ }
+
+ auto& stack = itr->second;
+ if (!stack.size()) {
+ // Content can cause this simply by calling `popErrorScope` when
+ // there is no error scope pushed.
+ return PopErrorScopeResult{PopErrorScopeResultType::ThrowOperationError,
+ "popErrorScope on empty stack"_ns};
+ }
+
+ const auto& scope = stack.back();
+ const auto popLater = MakeScopeExit([&]() { stack.pop_back(); });
+
+ auto ret = PopErrorScopeResult{PopErrorScopeResultType::NoError};
+ if (scope.firstMessage) {
+ ret.message = *scope.firstMessage;
+ switch (scope.filter) {
+ case dom::GPUErrorFilter::Validation:
+ ret.resultType = PopErrorScopeResultType::ValidationError;
+ break;
+ case dom::GPUErrorFilter::Out_of_memory:
+ ret.resultType = PopErrorScopeResultType::OutOfMemory;
+ break;
+ case dom::GPUErrorFilter::Internal:
+ ret.resultType = PopErrorScopeResultType::InternalError;
+ break;
+ case dom::GPUErrorFilter::EndGuard_:
+ MOZ_CRASH("Bad GPUErrorFilter");
+ }
+ }
+ return ret;
+ }();
+ aResolver(popResult);
+ return IPC_OK();
+}
+
+ipc::IPCResult WebGPUParent::RecvGenerateError(const Maybe<RawId> aDeviceId,
+ const dom::GPUErrorFilter aType,
+ const nsCString& aMessage) {
+ ReportError(aDeviceId, aType, aMessage);
+ return IPC_OK();
+}
+
+bool WebGPUParent::UseExternalTextureForSwapChain(
+ ffi::WGPUSwapChainId aSwapChainId) {
+ auto ownerId = layers::RemoteTextureOwnerId{aSwapChainId._0};
+ const auto& lookup = mPresentationDataMap.find(ownerId);
+ if (lookup == mPresentationDataMap.end()) {
+ MOZ_ASSERT_UNREACHABLE("unexpected to be called");
+ return IPC_OK();
+ }
+
+ RefPtr<PresentationData> data = lookup->second.get();
+
+ return data->mUseExternalTextureInSwapChain;
+}
+
+bool WebGPUParent::EnsureExternalTextureForSwapChain(
+ ffi::WGPUSwapChainId aSwapChainId, ffi::WGPUDeviceId aDeviceId,
+ ffi::WGPUTextureId aTextureId, uint32_t aWidth, uint32_t aHeight,
+ struct ffi::WGPUTextureFormat aFormat, ffi::WGPUTextureUsages aUsage) {
+ auto ownerId = layers::RemoteTextureOwnerId{aSwapChainId._0};
+ const auto& lookup = mPresentationDataMap.find(ownerId);
+ if (lookup == mPresentationDataMap.end()) {
+ MOZ_ASSERT_UNREACHABLE("unexpected to be called");
+ return false;
+ }
+
+ RefPtr<PresentationData> data = lookup->second.get();
+ if (!data->mUseExternalTextureInSwapChain) {
+ MOZ_ASSERT_UNREACHABLE("unexpected to be called");
+ return false;
+ }
+
+ // Recycled ExternalTexture if it exists.
+ if (!data->mRecycledExternalTextures.empty()) {
+ std::shared_ptr<ExternalTexture> texture =
+ data->mRecycledExternalTextures.front();
+ // Check if the texture is recyclable.
+ if (texture->mWidth == aWidth && texture->mHeight == aHeight &&
+ texture->mFormat.tag == aFormat.tag && texture->mUsage == aUsage) {
+ data->mRecycledExternalTextures.pop_front();
+ mExternalTextures.emplace(aTextureId, texture);
+ return true;
+ }
+ data->mRecycledExternalTextures.clear();
+ }
+
+ auto externalTexture = CreateExternalTexture(aDeviceId, aTextureId, aWidth,
+ aHeight, aFormat, aUsage);
+ if (!externalTexture) {
+ return false;
+ }
+ return true;
+}
+
+std::shared_ptr<ExternalTexture> WebGPUParent::CreateExternalTexture(
+ ffi::WGPUDeviceId aDeviceId, ffi::WGPUTextureId aTextureId, uint32_t aWidth,
+ uint32_t aHeight, const struct ffi::WGPUTextureFormat aFormat,
+ ffi::WGPUTextureUsages aUsage) {
+ MOZ_RELEASE_ASSERT(mExternalTextures.find(aTextureId) ==
+ mExternalTextures.end());
+
+ UniquePtr<ExternalTexture> texture =
+ ExternalTexture::Create(aWidth, aHeight, aFormat, aUsage);
+ if (!texture) {
+ MOZ_ASSERT_UNREACHABLE("unexpected to be called");
+ return nullptr;
+ }
+
+ std::shared_ptr<ExternalTexture> shared(texture.release());
+ mExternalTextures.emplace(aTextureId, shared);
+
+ return shared;
+}
+
+std::shared_ptr<ExternalTexture> WebGPUParent::GetExternalTexture(
+ ffi::WGPUTextureId aId) {
+ auto it = mExternalTextures.find(aId);
+ if (it == mExternalTextures.end()) {
+ return nullptr;
+ }
+ return it->second;
+}
+
+/* static */
+Maybe<ffi::WGPUFfiLUID> WebGPUParent::GetCompositorDeviceLuid() {
+#if defined(XP_WIN)
+ const RefPtr<ID3D11Device> d3d11Device =
+ gfx::DeviceManagerDx::Get()->GetCompositorDevice();
+ if (!d3d11Device) {
+ gfxCriticalNoteOnce << "CompositorDevice does not exist";
+ return Nothing();
+ }
+
+ RefPtr<IDXGIDevice> dxgiDevice;
+ d3d11Device->QueryInterface((IDXGIDevice**)getter_AddRefs(dxgiDevice));
+
+ RefPtr<IDXGIAdapter> dxgiAdapter;
+ dxgiDevice->GetAdapter(getter_AddRefs(dxgiAdapter));
+
+ DXGI_ADAPTER_DESC desc;
+ if (FAILED(dxgiAdapter->GetDesc(&desc))) {
+ gfxCriticalNoteOnce << "Failed to get DXGI_ADAPTER_DESC";
+ return Nothing();
+ }
+
+ return Some(
+ ffi::WGPUFfiLUID{desc.AdapterLuid.LowPart, desc.AdapterLuid.HighPart});
+#else
+ return Nothing();
+#endif
+}
+
+} // namespace mozilla::webgpu