WebGPU

W3C Working Draft,

More details about this document
This version:
https://www.w3.org/TR/2023/WD-webgpu-20230621/
Latest published version:
https://www.w3.org/TR/webgpu/
Editor's Draft:
https://gpuweb.github.io/gpuweb/
Previous Versions:
History:
https://www.w3.org/standards/history/webgpu
Feedback:
public-gpu@w3.org with subject line “[webgpu] … message topic …” (archives)
GitHub
Inline In Spec
Editors:
(Google)
(Google)
(Apple Inc.)
Former Editors:
(Mozilla)
(Apple Inc.)
Participate:
File an issue (open issues)

Abstract

WebGPU exposes an API for performing operations, such as rendering and computation, on a Graphics Processing Unit.

Status of this document

This section describes the status of this document at the time of its publication. A list of current W3C publications and the latest revision of this technical report can be found in the W3C technical reports index at https://www.w3.org/TR/.

Feedback and comments on this specification are welcome. GitHub Issues are preferred for discussion on this specification. Alternatively, you can send comments to the GPU for the Web Working Group’s mailing-list, public-gpu@w3.org (archives). This draft highlights some of the pending issues that are still to be discussed in the working group. No decision has been taken on the outcome of these issues including whether they are valid.

This document was published by the GPU for the Web Working Group as a Working Draft using the Recommendation track. This document is intended to become a W3C Recommendation.

Publication as a Working Draft does not imply endorsement by W3C and its Members.

This is a draft document and may be updated, replaced or obsoleted by other documents at any time. It is inappropriate to cite this document as other than work in progress.

This document was produced by a group operating under the W3C Patent Policy. W3C maintains a public list of any patent disclosures made in connection with the deliverables of the group; that page also includes instructions for disclosing a patent. An individual who has actual knowledge of a patent which the individual believes contains Essential Claim(s) must disclose the information in accordance with section 6 of the W3C Patent Policy.

This document is governed by the 2 November 2021 W3C Process Document.

1. Introduction

This section is non-normative.

Graphics Processing Units, or GPUs for short, have been essential in enabling rich rendering and computational applications in personal computing. WebGPU is an API that exposes the capabilities of GPU hardware for the Web. The API is designed from the ground up to efficiently map to (post-2014) native GPU APIs. WebGPU is not related to WebGL and does not explicitly target OpenGL ES.

WebGPU sees physical GPU hardware as GPUAdapters. It provides a connection to an adapter via GPUDevice, which manages resources, and the device’s GPUQueues, which execute commands. GPUDevice may have its own memory with high-speed access to the processing units. GPUBuffer and GPUTexture are the physical resources backed by GPU memory. GPUCommandBuffer and GPURenderBundle are containers for user-recorded commands. GPUShaderModule contains shader code. The other resources, such as GPUSampler or GPUBindGroup, configure the way physical resources are used by the GPU.

GPUs execute commands encoded in GPUCommandBuffers by feeding data through a pipeline, which is a mix of fixed-function and programmable stages. Programmable stages execute shaders, which are special programs designed to run on GPU hardware. Most of the state of a pipeline is defined by a GPURenderPipeline or a GPUComputePipeline object. The state not included in these pipeline objects is set during encoding with commands, such as beginRenderPass() or setBlendConstant().

2. Malicious use considerations

This section is non-normative. It describes the risks associated with exposing this API on the Web.

2.1. Security Considerations

The security requirements for WebGPU are the same as ever for the web, and are likewise non-negotiable. The general approach is strictly validating all the commands before they reach GPU, ensuring that a page can only work with its own data.

2.1.1. CPU-based undefined behavior

A WebGPU implementation translates the workloads issued by the user into API commands specific to the target platform. Native APIs specify the valid usage for the commands (for example, see vkCreateDescriptorSetLayout) and generally don’t guarantee any outcome if the valid usage rules are not followed. This is called "undefined behavior", and it can be exploited by an attacker to access memory they don’t own, or force the driver to execute arbitrary code.

In order to disallow insecure usage, the range of allowed WebGPU behaviors is defined for any input. An implementation has to validate all the input from the user and only reach the driver with the valid workloads. This document specifies all the error conditions and handling semantics. For example, specifying the same buffer with intersecting ranges in both "source" and "destination" of copyBufferToBuffer() results in GPUCommandEncoder generating an error, and no other operation occurring.

See § 22 Errors & Debugging for more information about error handling.

2.1.2. GPU-based undefined behavior

WebGPU shaders are executed by the compute units inside GPU hardware. In native APIs, some of the shader instructions may result in undefined behavior on the GPU. In order to address that, the shader instruction set and its defined behaviors are strictly defined by WebGPU. When a shader is provided to createShaderModule(), the WebGPU implementation has to validate it before doing any translation (to platform-specific shaders) or transformation passes.

2.1.3. Uninitialized data

Generally, allocating new memory may expose the leftover data of other applications running on the system. In order to address that, WebGPU conceptually initializes all the resources to zero, although in practice an implementation may skip this step if it sees the developer initializing the contents manually. This includes variables and shared workgroup memory inside shaders.

The precise mechanism of clearing the workgroup memory can differ between platforms. If the native API does not provide facilities to clear it, the WebGPU implementation transforms the compute shader to first do a clear across all invocations, synchronize them, and continue executing developer’s code.

Note: The initialization status of a resource used in a queue operation can only be known when the operation is enqueued (not when it is encoded into a command buffer, for example). Therefore, some implementations will require an unoptimized late-clear at enqueue time (e.g. clearing a texture, rather than changing GPULoadOp "load" to "clear").

As a result, all implementations should issue a developer console warning about this potential performance penalty, even if there is no penalty in that implementation.

2.1.4. Out-of-bounds access in shaders

Shaders can access physical resources either directly (for example, as a "uniform" GPUBufferBinding), or via texture units, which are fixed-function hardware blocks that handle texture coordinate conversions. Validation in the WebGPU API can only guarantee that all the inputs to the shader are provided and they have the correct usage and types. The WebGPU API can not guarantee that the data is accessed within bounds if the texture units are not involved.

In order to prevent the shaders from accessing GPU memory an application doesn’t own, the WebGPU implementation may enable a special mode (called "robust buffer access") in the driver that guarantees that the access is limited to buffer bounds.

Alternatively, an implementation may transform the shader code by inserting manual bounds checks. When this path is taken, the out-of-bound checks only apply to array indexing. They aren’t needed for plain field access of shader structures due to the minBindingSize validation on the host side.

If the shader attempts to load data outside of physical resource bounds, the implementation is allowed to:

  1. return a value at a different location within the resource bounds

  2. return a value vector of "(0, 0, 0, X)" with any "X"

  3. partially discard the draw or dispatch call

If the shader attempts to write data outside of physical resource bounds, the implementation is allowed to:

  1. write the value to a different location within the resource bounds

  2. discard the write operation

  3. partially discard the draw or dispatch call

2.1.5. Invalid data

When uploading floating-point data from CPU to GPU, or generating it on the GPU, we may end up with a binary representation that doesn’t correspond to a valid number, such as infinity or NaN (not-a-number). The GPU behavior in this case is subject to the accuracy of the GPU hardware implementation of the IEEE-754 standard. WebGPU guarantees that introducing invalid floating-point numbers would only affect the results of arithmetic computations and will not have other side effects.

2.1.6. Driver bugs

GPU drivers are subject to bugs like any other software. If a bug occurs, an attacker could possibly exploit the incorrect behavior of the driver to get access to unprivileged data. In order to reduce the risk, the WebGPU working group will coordinate with GPU vendors to integrate the WebGPU Conformance Test Suite (CTS) as part of their driver testing process, like it was done for WebGL. WebGPU implementations are expected to have workarounds for some of the discovered bugs, and disable WebGPU on drivers with known bugs that can’t be worked around.

2.1.7. Timing attacks

WebGPU is designed to later support multi-threaded use via Web Workers. As such, it is designed not to open the users to modern high-precision timing attacks. Some of the objects, like GPUBuffer or GPUQueue, have shared state which can be simultaneously accessed. This allows race conditions to occur, similar to those of accessing a SharedArrayBuffer from multiple Web Workers, which makes the thread scheduling observable.

WebGPU addresses this by limiting the ability to deserialize (or share) objects only to the agents inside the agent cluster, and only if the cross-origin isolated policies are in place. This restriction matches the mitigations against the malicious SharedArrayBuffer use. Similarly, the user agent may also serialize the agents sharing any handles to prevent any concurrency entirely.

In the end, the attack surface for races on shared state in WebGPU will be a small subset of the SharedArrayBuffer attacks.

WebGPU also specifies the "timestamp-query" feature, which provides high precision timing of GPU operations. The feature is optional, and a WebGPU implementation may limit its exposure only to those scenarios that are trusted. Alternatively, the timing query results could be processed by a compute shader and aligned to a lower precision.

2.1.8. Row hammer attacks

Row hammer is a class of attacks that exploit the leaking of states in DRAM cells. It could be used on GPU. WebGPU does not have any specific mitigations in place, and relies on platform-level solutions, such as reduced memory refresh intervals.

2.1.9. Denial of service

WebGPU applications have access to GPU memory and compute units. A WebGPU implementation may limit the available GPU memory to an application, in order to keep other applications responsive. For GPU processing time, a WebGPU implementation may set up "watchdog" timer that makes sure an application doesn’t cause GPU unresponsiveness for more than a few seconds. These measures are similar to those used in WebGL.

2.1.10. Workload identification

WebGPU provides access to constrained global resources shared between different programs (and web pages) running on the same machine. An application can try to indirectly probe how constrained these global resources are, in order to reason about workloads performed by other open web pages, based on the patterns of usage of these shared resources. These issues are generally analogous to issues with Javascript, such as system memory and CPU execution throughput. WebGPU does not provide any additional mitigations for this.

2.1.11. Memory resources

WebGPU exposes fallible allocations from machine-global memory heaps, such as VRAM. This allows for probing the size of the system’s remaining available memory (for a given heap type) by attempting to allocate and watching for allocation failures.

GPUs internally have one or more (typically only two) heaps of memory shared by all running applications. When a heap is depleted, WebGPU would fail to create a resource. This is observable, which may allow a malicious application to guess what heaps are used by other applications, and how much they allocate from them.

2.1.12. Computation resources

If one site uses WebGPU at the same time as another, it may observe the increase in time it takes to process some work. For example, if a site constantly submits compute workloads and tracks completion of work on the queue, it may observe that something else also started using the GPU.

A GPU has many parts that can be tested independently, such as the arithmetic units, texture sampling units, atomic units, etc. A malicious application may sense when some of these units are stressed, and attempt to guess the workload of another application by analyzing the stress patterns. This is analogous to the realities of CPU execution of Javascript.

2.1.13. Abuse of capabilities

Malicious sites could abuse the capabilities exposed by WebGPU to run computations that don’t benefit the user or their experience and instead only benefit the site. Examples would be hidden crypto-mining, password cracking or rainbow tables computations.

It is not possible to guard against these types of uses of the API because the browser is not able to distinguish between valid workloads and abusive workloads. This is a general problem with all general-purpose computation capabilities on the Web: JavaScript, WebAssembly or WebGL. WebGPU only makes some workloads easier to implement, or slightly more efficient to run than using WebGL.

To mitigate this form of abuse, browsers can throttle operations on background tabs, could warn that a tab is using a lot of resource, and restrict which contexts are allowed to use WebGPU.

User agents can heuristically issue warnings to users about high power use, especially due to potentially malicious usage. If a user agent implements such a warning, it should include WebGPU usage in its heuristics, in addition to JavaScript, WebAssembly, WebGL, and so on.

2.2. Privacy Considerations

There is a tracking vector here. The privacy considerations for WebGPU are similar to those of WebGL. GPU APIs are complex and must expose various aspects of a device’s capabilities out of necessity in order to enable developers to take advantage of those capabilities effectively. The general mitigation approach involves normalizing or binning potentially identifying information and enforcing uniform behavior where possible.

A user agent must not reveal more than 32 distinguishable configurations or buckets.

2.2.1. Machine-specific features and limits

WebGPU can expose a lot of detail on the underlying GPU architecture and the device geometry. This includes available physical adapters, many limits on the GPU and CPU resources that could be used (such as the maximum texture size), and any optional hardware-specific capabilities that are available.

User agents are not obligated to expose the real hardware limits, they are in full control of how much the machine specifics are exposed. One strategy to reduce fingerprinting is binning all the target platforms into a few number of bins. In general, the privacy impact of exposing the hardware limits matches the one of WebGL.

The default limits are also deliberately high enough to allow most applications to work without requesting higher limits. All the usage of the API is validated according to the requested limits, so the actual hardware capabilities are not exposed to the users by accident.

2.2.2. Machine-specific artifacts

There are some machine-specific rasterization/precision artifacts and performance differences that can be observed roughly in the same way as in WebGL. This applies to rasterization coverage and patterns, interpolation precision of the varyings between shader stages, compute unit scheduling, and more aspects of execution.

Generally, rasterization and precision fingerprints are identical across most or all of the devices of each vendor. Performance differences are relatively intractable, but also relatively low-signal (as with JS execution performance).

Privacy-critical applications and user agents should utilize software implementations to eliminate such artifacts.

2.2.3. Machine-specific performance

Another factor for differentiating users is measuring the performance of specific operations on the GPU. Even with low precision timing, repeated execution of an operation can show if the user’s machine is fast at specific workloads. This is a fairly common vector (present in both WebGL and Javascript), but it’s also low-signal and relatively intractable to truly normalize.

WebGPU compute pipelines expose access to GPU unobstructed by the fixed-function hardware. This poses an additional risk for unique device fingerprinting. User agents can take steps to dissociate logical GPU invocations with actual compute units to reduce this risk.

2.2.4. User Agent State

This specification doesn’t define any additional user-agent state for an origin. However it is expected that user agents will have compilation caches for the result of expensive compilation like GPUShaderModule, GPURenderPipeline and GPUComputePipeline. These caches are important to improve the loading time of WebGPU applications after the first visit.

For the specification, these caches are indifferentiable from incredibly fast compilation, but for applications it would be easy to measure how long createComputePipelineAsync() takes to resolve. This can leak information across origins (like "did the user access a site with this specific shader") so user agents should follow the best practices in storage partitioning.

The system’s GPU driver may also have its own cache of compiled shaders and pipelines. User agents may want to disable these when at all possible, or add per-partition data to shaders in ways that will make the GPU driver consider them different.

2.2.5. Driver bugs

In addition to the concerns outlined in Security Considerations, driver bugs may introduce differences in behavior that can be observed as a method of differentiating users. The mitigations mentioned in Security Considerations apply here as well, including coordinating with GPU vendors and implementing workarounds for known issues in the user agent.

2.2.6. Adapter Identifiers

Past experience with WebGL has demonstrated that developers have a legitimate need to be able to identify the GPU their code is running on in order to create and maintain robust GPU-based content. For example, to identify adapters with known driver bugs in order to work around them or to avoid features that perform more poorly than expected on a given class of hardware.

But exposing adapter identifiers also naturally expands the amount of fingerprinting information available, so there’s a desire to limit the precision with which we identify the adapter.

There are several mitigations that can be applied to strike a balance between enabling robust content and preserving privacy. First is that user agents can reduce the burden on developers by identifying and working around known driver issues, as they have since browsers began making use of GPUs.

When adapter identifiers are exposed by default they should be as broad as possible while still being useful. Possibly identifying, for example, the adapter’s vendor and general architecture without identifying the specific adapter in use. Similarly, in some cases identifiers for an adapter that is considered a reasonable proxy for the actual adapter may be reported.

In cases where full and detailed information about the adapter is useful (for example: when filing bug reports) the user can be asked for consent to reveal additional information about their hardware to the page.

Finally, the user agent will always have the discretion to not report adapter identifiers at all if it considers it appropriate, such as in enhanced privacy modes.

3. Fundamentals

3.1. Conventions

3.1.1. Syntactic Shorthands

In this specification, the following syntactic shorthands are used:

The . ("dot") syntax, common in programming languages.

The phrasing "Foo.Bar" means "the Bar member of the value (or interface) Foo." If Foo is an ordered map, asserts that the key Bar exists.

Editorial: Some phrasing in this spec may currently assume this resolves to undefined if Bar doesn’t exist.

The phrasing "Foo.Bar is provided" means "the Bar member exists in the map value Foo"

The ?. ("optional chaining") syntax, adopted from JavaScript.

The phrasing "Foo?.Bar" means "if Foo is null or undefined or Bar does not exist in Foo, undefined; otherwise, Foo.Bar".

For example, where buffer is a GPUBuffer, buffer?.\[[device]].\[[adapter]] means "if buffer is null or undefined, then undefined; otherwise, the \[[adapter]] internal slot of the \[[device]] internal slot of buffer.

The ?? ("nullish coalescing") syntax, adopted from JavaScript.

The phrasing "x ?? y" means "x, if x is not null/undefined, and y otherwise".

slot-backed attribute

A WebIDL attribute which is backed by an internal slot of the same name. It may or may not be mutable.

3.1.2. WebGPU Interfaces

A WebGPU interface defines a WebGPU object. It can be used:

The following special property types can be defined on WebGPU interfaces:

immutable property

A read-only slot set during initialization of the object. It can be accessed from any timeline.

Note: Since the slot is immutable, implementations may have a copy on multiple timelines, as needed. Immutable properties are defined in this way to avoid describing multiple copies in this spec.

If named [[with brackets]], it is an internal slot. If named withoutBrackets, it is a readonly slot-backed attribute.

content timeline property

A property which is only accessible from the content timeline where the object was created.

If named [[with brackets]], it is an internal slot. If named withoutBrackets, it is a slot-backed attribute.

Any interface which includes GPUObjectBase is a WebGPU interface.

interface mixin GPUObjectBase {
    attribute USVString label;
};
To create a new WebGPU object(GPUObjectBase parent, interface T, GPUObjectDescriptorBase descriptor) (where T extends GPUObjectBase):
  1. Let device be parent.[[device]].

  2. Let object be a new instance of T.

  3. Let internals be a new (uninitialized) instance of the type of T.\[[internals]] (which may override GPUObjectBase.[[internals]]) that is accessible only from the device timeline of device.

  4. Set object.[[device]] to device.

  5. Set object.[[internals]] to internals.

  6. Set object.label to descriptor.label.

  7. Return [object, internals].

GPUObjectBase has the following immutable properties:

[[internals]], of type internal object, readonly, overridable

The internal object.

Operations on the contents of this object assert they are running on the device timeline, and that the device is valid.

For each interface that subtypes GPUObjectBase, this may be overridden with a subtype of internal object. This slot is initially set to an uninitialized object of that type.

[[device]], of type device, readonly

The device that owns the internal object.

Operations on the contents of this object assert they are running on the device timeline, and that the device is valid.

GPUObjectBase has the following content timeline properties:

label, of type USVString

A developer-provided label which is used in an implementation-defined way. It can be used by the browser, OS, or other tools to help identify the underlying internal object to the developer. Examples include displaying the label in GPUError messages, console warnings, browser developer tools, and platform debugging utilities.

Implementations should use labels to enhance error messages by using them to identify WebGPU objects.

However, this need not be the only way of identifying objects: implementations should also use other available information, especially when no label is available. For example:

Note: The label is a property of the GPUObjectBase. Two GPUObjectBase "wrapper" objects have completely separate label states, even if they refer to the same underlying object (for example returned by getBindGroupLayout()). The label property will not change except by being set from JavaScript.

This means one underlying object could be associated with multiple labels. This specification does not define how the label is propagated to the device timeline. How labels are used is completely implementation-defined: error messages could show the most recently set label, all known labels, or no labels at all.

It is defined as a USVString because some user agents may supply it to the debug facilities of the underlying native APIs.

Note: Ideally WebGPU interfaces should not prevent their parent objects, such as the [[device]] that owns them, from being garbage collected. This cannot be guaranteed, however, as holding a strong reference to a parent object may be required in some implementations.

As a result, developers should assume that a WebGPU interface may not be garbage collected until all child objects of that interface have also been garbage collected. This may cause some resources to remain allocated longer than anticipated.

Calling the destroy method on a WebGPU interface (such as GPUDevice.destroy() or GPUBuffer.destroy()) should be favored over relying on garbage collection if predictable release of allocated resources is needed.

3.1.3. Internal Objects

An internal object tracks state of WebGPU objects that may only be used on the device timeline, in device timeline slots, which may be mutable.

device timeline slot

An internal slot which is only accessible from the device timeline.

All reads/writes to the mutable state of an internal object occur from steps executing on a single well-ordered device timeline. These steps may have been issued from a content timeline algorithm on any of multiple agents.

Note: An "agent" refers to a JavaScript "thread" (i.e. main thread, or Web Worker).

3.1.4. Object Descriptors

An object descriptor holds the information needed to create an object, which is typically done via one of the create* methods of GPUDevice.

dictionary GPUObjectDescriptorBase {
    USVString label = "";
};

GPUObjectDescriptorBase has the following members:

label, of type USVString, defaulting to ""

The initial value of GPUObjectBase.label.

3.2. Asynchrony

3.2.1. Invalid Internal Objects & Contagious Invalidity

Object creation operations in WebGPU don’t return promises, but nonetheless are internally asynchronous. Returned objects refer to internal objects which are manipulated on a device timeline. Rather than fail with exceptions or rejections, most errors that occur on a device timeline are communicated through GPUErrors generated on the associated device.

Internal objects are either valid or invalid. An invalid object will never become valid at a later time, but some valid objects may become invalid.

Objects are invalid from creation if it wasn’t possible to create them. This can happen, for example, if the object descriptor doesn’t describe a valid object, or if there is not enough memory to allocate a resource.

Internal objects of most types cannot become invalid after they are created, but still may become unusable, e.g. if the owning device is lost or destroyed, or the object has a special internal state, like buffer state "destroyed".

Internal objects of some types can become invalid after they are created; specifically, devices, adapters, GPUCommandBuffers, and command/pass/bundle encoders.

A given GPUObjectBase object is valid to use with a targetObject if and only if the following requirements are met:

3.2.2. Promise Ordering

Several operations in WebGPU return promises.

WebGPU does not make any guarantees about the order in which these promises settle (resolve or reject), except for the following:

Applications must not rely on any other promise settlement ordering.

3.3. Coordinate Systems

Rendering operations use the following coordinate systems:

Note: WebGPU’s coordinate systems match DirectX’s coordinate systems in a graphics pipeline.

3.4. Programming Model

3.4.1. Timelines

WebGPU’s behavior is described in terms of "timelines". Each operation (defined as algorithms) occurs on a timeline. Timelines clearly define both the order of operations, and which state is available to which operations.

Note: This "timeline" model describes the constraints of the multi-process models of browser engines (typically with a "content process" and "GPU process"), as well as the GPU itself as a separate execution unit in many implementations. Implementing WebGPU does not require timelines to execute in parallel, so does not require multiple processes, or even multiple threads.

Content timeline

Associated with the execution of the Web script. It includes calling all methods described by this specification.

To issue steps to the content timeline from an operation on GPUDevice device, queue a global task for GPUDevice device with those steps.

Device timeline

Associated with the GPU device operations that are issued by the user agent. It includes creation of adapters, devices, and GPU resources and state objects, which are typically synchronous operations from the point of view of the user agent part that controls the GPU, but can live in a separate OS process.

Queue timeline

Associated with the execution of operations on the compute units of the GPU. It includes actual draw, copy, and compute jobs that run on the GPU.

The following show the styling of steps and values associated with each timeline. This styling is non-normative; the specification text always describes the association.
Immutable value example definition

Can be used on any timeline.

Content-timeline example definition

Can only be used on the content timeline.

Device-timeline example definition

Can only be used on the device timeline.

Queue-timeline example definition

Can only be used on the queue timeline.

In this specification, asynchronous operations are used when the return value depends on work that happens on any timeline other than the Content timeline. They are represented by promises and events in API.

GPUComputePassEncoder.dispatchWorkgroups():
  1. User encodes a dispatchWorkgroups command by calling a method of the GPUComputePassEncoder which happens on the Content timeline.

  2. User issues GPUQueue.submit() that hands over the GPUCommandBuffer to the user agent, which processes it on the Device timeline by calling the OS driver to do a low-level submission.

  3. The submit gets dispatched by the GPU invocation scheduler onto the actual compute units for execution, which happens on the Queue timeline.

GPUDevice.createBuffer():
  1. User fills out a GPUBufferDescriptor and creates a GPUBuffer with it, which happens on the Content timeline.

  2. User agent creates a low-level buffer on the Device timeline.

GPUBuffer.mapAsync():
  1. User requests to map a GPUBuffer on the Content timeline and gets a promise in return.

  2. User agent checks if the buffer is currently used by the GPU and makes a reminder to itself to check back when this usage is over.

  3. After the GPU operating on Queue timeline is done using the buffer, the user agent maps it to memory and resolves the promise.

3.4.2. Memory Model

This section is non-normative.

Once a GPUDevice has been obtained during an application initialization routine, we can describe the WebGPU platform as consisting of the following layers:

  1. User agent implementing the specification.

  2. Operating system with low-level native API drivers for this device.

  3. Actual CPU and GPU hardware.

Each layer of the WebGPU platform may have different memory types that the user agent needs to consider when implementing the specification:

Most physical resources are allocated in the memory of type that is efficient for computation or rendering by the GPU. When the user needs to provide new data to the GPU, the data may first need to cross the process boundary in order to reach the user agent part that communicates with the GPU driver. Then it may need to be made visible to the driver, which sometimes requires a copy into driver-allocated staging memory. Finally, it may need to be transferred to the dedicated GPU memory, potentially changing the internal layout into one that is most efficient for GPUs to operate on.

All of these transitions are done by the WebGPU implementation of the user agent.

Note: This example describes the worst case, while in practice the implementation may not need to cross the process boundary, or may be able to expose the driver-managed memory directly to the user behind an ArrayBuffer, thus avoiding any data copies.

3.4.3. Resource Usages

A physical resource can be used on GPU with an internal usage:

input

Buffer with input data for draw or dispatch calls. Preserves the contents. Allowed by buffer INDEX, buffer VERTEX, or buffer INDIRECT.

constant

Resource bindings that are constant from the shader point of view. Preserves the contents. Allowed by buffer UNIFORM or texture TEXTURE_BINDING.

storage

Writable storage resource binding. Allowed by buffer STORAGE or texture STORAGE_BINDING.

storage-read

Read-only storage resource bindings. Preserves the contents. Allowed by buffer STORAGE.

attachment

Texture used as an output attachment in a render pass. Allowed by texture RENDER_ATTACHMENT.

attachment-read

Texture used as a read-only attachment in a render pass. Preserves the contents. Allowed by texture RENDER_ATTACHMENT.

We define subresource to be either a whole buffer, or a texture subresource.

Some internal usages are compatible with others. A subresource can be in a state that combines multiple usages together. We consider a list U to be a compatible usage list if (and only if) it satisfies any of the following rules:

Enforcing that the usages are only combined into a compatible usage list allows the API to limit when data races can occur in working with memory. That property makes applications written against WebGPU more likely to run without modification on different platforms.

Generally, when an implementation processes an operation that uses a subresource in a different way than its current usage allows, it schedules a transition of the resource into the new state. In some cases, like within an open GPURenderPassEncoder, such a transition is impossible due to the hardware limitations. We define these places as usage scopes.

The main usage rule is, for any one subresource, its list of internal usages within one usage scope must be a compatible usage list.

For example, binding the same buffer for storage as well as for input within the same GPURenderPassEncoder would put the encoder as well as the owning GPUCommandEncoder into the error state. This combination of usages does not make a compatible usage list.

Note: race condition of multiple writable storage buffer/texture usages in a single usage scope is allowed.

The subresources of textures included in the views provided to GPURenderPassColorAttachment.view and GPURenderPassColorAttachment.resolveTarget are considered to be used as attachment for the usage scope of this render pass.

3.4.4. Synchronization

For each subresource of a physical resource, its set of internal usage flags is tracked on the Queue timeline.

On the Queue timeline, there is an ordered sequence of usage scopes. For the duration of each scope, the set of internal usage flags of any given subresource is constant. A subresource may transition to new usages at the boundaries between usage scopes.

This specification defines the following usage scopes:

The above should probably talk about GPU commands. But we don’t have a way to reference specific GPU commands (like dispatch) yet.

Note: The above rules mean the following example resource usages are included in usage scope validation:

During command encoding, every usage of a subresource is recorded in one of the usage scopes in the command buffer. For each usage scope, the implementation performs usage scope validation by composing the list of all internal usage flags of each subresource used in the usage scope. If any of those lists is not a compatible usage list, GPUCommandEncoder.finish() will generate a validation error.

3.5. Core Internal Objects

3.5.1. Adapters

An adapter identifies an implementation of WebGPU on the system: both an instance of compute/rendering functionality on the platform underlying a browser, and an instance of a browser’s implementation of WebGPU on top of that functionality.

Adapters do not uniquely represent underlying implementations: calling requestAdapter() multiple times returns a different adapter object each time.

Each adapter object can only be used to create one device: upon a successful requestDevice(), the adapter becomes invalid. Additionally, adapter objects may expire at any time.

Note: This ensures applications use the latest system state for adapter selection when creating a device. It also encourages robustness to more scenarios by making them look similar: first initialization, reinitialization due to an unplugged adapter, reinitialization due to a test GPUDevice.destroy() call, etc.

An adapter may be considered a fallback adapter if it has significant performance caveats in exchange for some combination of wider compatibility, more predictable behavior, or improved privacy. It is not required that a fallback adapter is available on every system.

An adapter has the following internal slots:

[[features]], of type ordered set<GPUFeatureName>, readonly

The features which can be used to create devices on this adapter.

[[limits]], of type supported limits, readonly

The best limits which can be used to create devices on this adapter.

Each adapter limit must be the same or better than its default value in supported limits.

[[fallback]], of type boolean

If set to true indicates that the adapter is a fallback adapter.

[[unmaskedIdentifiers]], of type ordered set<DOMString>

A list of names of GPUAdapterInfo fields the user agent has chosen to report for this adapter. Initially populated with the names of any GPUAdapterInfo fields the user agent has chosen to report without user consent.

Adapters are exposed via GPUAdapter.

3.5.2. Devices

A device is the logical instantiation of an adapter, through which internal objects are created. It can be shared across multiple agents (e.g. dedicated workers).

A device is the exclusive owner of all internal objects created from it: when the device becomes invalid (is lost or destroyed), it and all objects created on it (directly, e.g. createTexture(), or indirectly, e.g. createView()) become implicitly unusable.

A device has the following internal slots:

[[adapter]], of type adapter, readonly

The adapter from which this device was created.

[[features]], of type ordered set<GPUFeatureName>, readonly

The features which can be used on this device. No additional features can be used, even if the underlying adapter can support them.

[[limits]], of type supported limits, readonly

The limits which can be used on this device. No better limits can be used, even if the underlying adapter can support them.

When a new device device is created from adapter adapter with GPUDeviceDescriptor descriptor:

Any time the user agent needs to revoke access to a device, it calls lose the device(device, "unknown") on the device’s device timeline, potentially ahead of other operations currently queued on that timeline.

If an operation fails with side effects that would observably change the state of objects on the device or potentially corrupt internal implementation/driver state, the device should be lost to prevent these changes from being observable.

Note: For all device losses not initiated by the application (via destroy(), user agents should consider issuing developer-visible warnings unconditionally, even if the lost promise is handled. These scenarios should be rare, and the signal is vital to developers because most of the WebGPU API tries to behave like nothing is wrong to avoid interrupting the runtime flow of the application: no validation errors are raised, most promises resolve normally, etc.

To lose the device(device, reason):
  1. Make device invalid.

  2. Let gpuDevice be the content timeline GPUDevice corresponding to device.

    Define this more rigorously.

  3. Issue the following steps on the content timeline of gpuDevice:

    1. Resolve device.lost with a new GPUDeviceLostInfo with reason set to reason and message set to an implementation-defined value.

      Note: message should not disclose unnecessary user/system information and should never be parsed by applications.

  4. Complete any outstanding mapAsync() steps.

  5. Complete any outstanding onSubmittedWorkDone() steps.

Note: No errors are generated after device loss. See § 22 Errors & Debugging.

Devices are exposed via GPUDevice.

3.6. Optional Capabilities

WebGPU adapters and devices have capabilities, which describe WebGPU functionality that differs between different implementations, typically due to hardware or system software constraints. A capability is either a feature or a limit.

A user agent must not reveal more than 32 distinguishable configurations or buckets.

The capabilities of an adapter must conform to § 4.2.1 Adapter Capability Guarantees.

Only supported capabilities may be requested in requestDevice(); requesting unsupported capabilities results in failure.

The capabilities of a device are exactly the ones which were requested in requestDevice(). These capabilities are enforced regardless of the capabilities of the adapter.

There is a tracking vector here. For privacy considerations, see § 2.2.1 Machine-specific features and limits.

3.6.1. Features

A feature is a set of optional WebGPU functionality that is not supported on all implementations, typically due to hardware or system software constraints.

Functionality that is part of a feature may only be used if the feature was requested at device creation (in requiredFeatures). Otherwise, using existing API surfaces in a new way typically results in a validation error, and using optional API surfaces results in the following:

A GPUFeatureName feature is enabled for a GPUObjectBase object if and only if object.[[device]].[[features]] contains feature.

See the Feature Index for a description of the functionality each feature enables.

3.6.2. Limits

Each limit is a numeric limit on the usage of WebGPU on a device.

Each limit has a default value. Every adapter is guaranteed to support the default value or better. The default is used if a value is not explicitly specified in requiredLimits.

One limit value may be better than another. A better limit value always relaxes validation, enabling strictly more programs to be valid. For each limit class, "better" is defined.

Different limits have different limit classes:

maximum

The limit enforces a maximum on some value passed into the API.

Higher values are better.

May only be set to values ≥ the default. Lower values are clamped to the default.

alignment

The limit enforces a minimum alignment on some value passed into the API; that is, the value must be a multiple of the limit.

Lower values are better.

May only be set to powers of 2 which are ≤ the default. Values which are not powers of 2 are invalid. Higher powers of 2 are clamped to the default.

Note: Setting "better" limits may not necessarily be desirable, as they may have a performance impact. Because of this, and to improve portability across devices and implementations, applications should generally request the "worst" limits that work for their content (ideally, the default values).

A supported limits object has a value for every limit defined by WebGPU:

Limit name Type Limit class Default
maxTextureDimension1D GPUSize32 maximum 8192
The maximum allowed value for the size.width of a texture created with dimension "1d".
maxTextureDimension2D GPUSize32 maximum 8192
The maximum allowed value for the size.width and size.height of a texture created with dimension "2d".
maxTextureDimension3D GPUSize32 maximum 2048
The maximum allowed value for the size.width, size.height and size.depthOrArrayLayers of a texture created with dimension "3d".
maxTextureArrayLayers GPUSize32 maximum 256
The maximum allowed value for the size.depthOrArrayLayers of a texture created with dimension "2d".
maxBindGroups GPUSize32 maximum 4
The maximum number of GPUBindGroupLayouts allowed in bindGroupLayouts when creating a GPUPipelineLayout.
maxBindGroupsPlusVertexBuffers GPUSize32 maximum 24
The maximum number of bind group and vertex buffer slots used simultaneously, counting any empty slots below the highest index. Validated in createRenderPipeline() and in draw calls.
maxBindingsPerBindGroup GPUSize32 maximum 1000
The number of binding indices available when creating a GPUBindGroupLayout.

Note: This limit is normative, but arbitrary. With the default binding slot limits, it is impossible to use 1000 bindings in one bind group, but this allows GPUBindGroupLayoutEntry.binding values up to 999. This limit allows implementations to treat binding space as an array, within reasonable memory space, rather than a sparse map structure.

maxDynamicUniformBuffersPerPipelineLayout GPUSize32 maximum 8
The maximum number of GPUBindGroupLayoutEntry entries across a GPUPipelineLayout which are uniform buffers with dynamic offsets. See Exceeds the binding slot limits.
maxDynamicStorageBuffersPerPipelineLayout GPUSize32 maximum 4
The maximum number of GPUBindGroupLayoutEntry entries across a GPUPipelineLayout which are storage buffers with dynamic offsets. See Exceeds the binding slot limits.
maxSampledTexturesPerShaderStage GPUSize32 maximum 16
For each possible GPUShaderStage stage, the maximum number of GPUBindGroupLayoutEntry entries across a GPUPipelineLayout which are sampled textures. See Exceeds the binding slot limits.
maxSamplersPerShaderStage GPUSize32 maximum 16
For each possible GPUShaderStage stage, the maximum number of GPUBindGroupLayoutEntry entries across a GPUPipelineLayout which are samplers. See Exceeds the binding slot limits.
maxStorageBuffersPerShaderStage GPUSize32 maximum 8
For each possible GPUShaderStage stage, the maximum number of GPUBindGroupLayoutEntry entries across a GPUPipelineLayout which are storage buffers. See Exceeds the binding slot limits.
maxStorageTexturesPerShaderStage GPUSize32 maximum 4
For each possible GPUShaderStage stage, the maximum number of GPUBindGroupLayoutEntry entries across a GPUPipelineLayout which are storage textures. See Exceeds the binding slot limits.
maxUniformBuffersPerShaderStage GPUSize32 maximum 12
For each possible GPUShaderStage stage, the maximum number of GPUBindGroupLayoutEntry entries across a GPUPipelineLayout which are uniform buffers. See Exceeds the binding slot limits.
maxUniformBufferBindingSize GPUSize64 maximum 65536 bytes
The maximum GPUBufferBinding.size for bindings with a GPUBindGroupLayoutEntry entry for which entry.buffer?.type is "uniform".
maxStorageBufferBindingSize GPUSize64 maximum 134217728 bytes (128 MiB)
The maximum GPUBufferBinding.size for bindings with a GPUBindGroupLayoutEntry entry for which entry.buffer?.type is "storage" or "read-only-storage".
minUniformBufferOffsetAlignment GPUSize32 alignment 256 bytes
The required alignment for GPUBufferBinding.offset and the dynamic offsets provided in setBindGroup(), for bindings with a GPUBindGroupLayoutEntry entry for which entry.buffer?.type is "uniform".
minStorageBufferOffsetAlignment GPUSize32 alignment 256 bytes
The required alignment for GPUBufferBinding.offset and the dynamic offsets provided in setBindGroup(), for bindings with a GPUBindGroupLayoutEntry entry for which entry.buffer?.type is "storage" or "read-only-storage".
maxVertexBuffers GPUSize32 maximum 8
The maximum number of buffers when creating a GPURenderPipeline.
maxBufferSize GPUSize64 maximum 268435456 bytes (256 MiB)
The maximum size of size when creating a GPUBuffer.
maxVertexAttributes GPUSize32 maximum 16
The maximum number of attributes in total across buffers when creating a GPURenderPipeline.
maxVertexBufferArrayStride GPUSize32 maximum 2048 bytes
The maximum allowed arrayStride when creating a GPURenderPipeline.
maxInterStageShaderComponents GPUSize32 maximum 60
The maximum allowed number of components of input or output variables for inter-stage communication (like vertex outputs or fragment inputs).
maxInterStageShaderVariables GPUSize32 maximum 16
The maximum allowed number of input or output variables for inter-stage communication (like vertex outputs or fragment inputs).
maxColorAttachments GPUSize32 maximum 8
The maximum allowed number of color attachments in GPURenderPipelineDescriptor.fragment.targets, GPURenderPassDescriptor.colorAttachments, and GPURenderPassLayout.colorFormats.
maxColorAttachmentBytesPerSample GPUSize32 maximum 32
The maximum number of bytes necessary to hold one sample (pixel or subpixel) of render pipeline output data, across all color attachments.
maxComputeWorkgroupStorageSize GPUSize32 maximum 16384 bytes
The maximum number of bytes of workgroup storage used for a compute stage GPUShaderModule entry-point.
maxComputeInvocationsPerWorkgroup GPUSize32 maximum 256
The maximum value of the product of the workgroup_size dimensions for a compute stage GPUShaderModule entry-point.
maxComputeWorkgroupSizeX GPUSize32 maximum 256
The maximum value of the workgroup_size X dimension for a compute stage GPUShaderModule entry-point.
maxComputeWorkgroupSizeY GPUSize32 maximum 256
The maximum value of the workgroup_size Y dimensions for a compute stage GPUShaderModule entry-point.
maxComputeWorkgroupSizeZ GPUSize32 maximum 64
The maximum value of the workgroup_size Z dimensions for a compute stage GPUShaderModule entry-point.
maxComputeWorkgroupsPerDimension GPUSize32 maximum 65535
The maximum value for the arguments of dispatchWorkgroups(workgroupCountX, workgroupCountY, workgroupCountZ).
3.6.2.1. GPUSupportedLimits

GPUSupportedLimits exposes the limits supported by an adapter or device. See GPUAdapter.limits and GPUDevice.limits.

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUSupportedLimits {
    readonly attribute unsigned long maxTextureDimension1D;
    readonly attribute unsigned long maxTextureDimension2D;
    readonly attribute unsigned long maxTextureDimension3D;
    readonly attribute unsigned long maxTextureArrayLayers;
    readonly attribute unsigned long maxBindGroups;
    readonly attribute unsigned long maxBindGroupsPlusVertexBuffers;
    readonly attribute unsigned long maxBindingsPerBindGroup;
    readonly attribute unsigned long maxDynamicUniformBuffersPerPipelineLayout;
    readonly attribute unsigned long maxDynamicStorageBuffersPerPipelineLayout;
    readonly attribute unsigned long maxSampledTexturesPerShaderStage;
    readonly attribute unsigned long maxSamplersPerShaderStage;
    readonly attribute unsigned long maxStorageBuffersPerShaderStage;
    readonly attribute unsigned long maxStorageTexturesPerShaderStage;
    readonly attribute unsigned long maxUniformBuffersPerShaderStage;
    readonly attribute unsigned long long maxUniformBufferBindingSize;
    readonly attribute unsigned long long maxStorageBufferBindingSize;
    readonly attribute unsigned long minUniformBufferOffsetAlignment;
    readonly attribute unsigned long minStorageBufferOffsetAlignment;
    readonly attribute unsigned long maxVertexBuffers;
    readonly attribute unsigned long long maxBufferSize;
    readonly attribute unsigned long maxVertexAttributes;
    readonly attribute unsigned long maxVertexBufferArrayStride;
    readonly attribute unsigned long maxInterStageShaderComponents;
    readonly attribute unsigned long maxInterStageShaderVariables;
    readonly attribute unsigned long maxColorAttachments;
    readonly attribute unsigned long maxColorAttachmentBytesPerSample;
    readonly attribute unsigned long maxComputeWorkgroupStorageSize;
    readonly attribute unsigned long maxComputeInvocationsPerWorkgroup;
    readonly attribute unsigned long maxComputeWorkgroupSizeX;
    readonly attribute unsigned long maxComputeWorkgroupSizeY;
    readonly attribute unsigned long maxComputeWorkgroupSizeZ;
    readonly attribute unsigned long maxComputeWorkgroupsPerDimension;
};
3.6.2.2. GPUSupportedFeatures

GPUSupportedFeatures is a setlike interface. Its set entries are the GPUFeatureName values of the features supported by an adapter or device. It must only contain strings from the GPUFeatureName enum.

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUSupportedFeatures {
    readonly setlike<DOMString>;
};
Note: The type of the GPUSupportedFeatures set entries is DOMString to allow user agents to gracefully handle valid GPUFeatureNames which are added in later revisions of the spec but which the user agent has not been updated to recognize yet. If the set entries type was GPUFeatureName the following code would throw an TypeError rather than reporting false:
Check for support of an unrecognized feature:
if (adapter.features.has('unknown-feature')) {
    // Use unknown-feature
} else {
    console.warn('unknown-feature is not supported by this adapter.');
}
3.6.2.3. WGSLLanguageFeatures

WGSLLanguageFeatures is a setlike interface. Its set entries are the string names of the WGSL language extensions supported all adapters.

[Exposed=(Window, DedicatedWorker), SecureContext]
interface WGSLLanguageFeatures {
    readonly setlike<DOMString>;
};
3.6.2.4. GPUAdapterInfo

GPUAdapterInfo exposes various identifying information about an adapter.

None of the members in GPUAdapterInfo are guaranteed to be populated. It is at the user agent’s discretion which values to reveal, and it is likely that on some devices none of the values will be populated. As such, applications must be able to handle any possible GPUAdapterInfo values, including the absence of those values.

There is a tracking vector here. For privacy considerations, see § 2.2.6 Adapter Identifiers.

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUAdapterInfo {
    readonly attribute DOMString vendor;
    readonly attribute DOMString architecture;
    readonly attribute DOMString device;
    readonly attribute DOMString description;
};

GPUAdapterInfo has the following attributes:

vendor, of type DOMString, readonly

The name of the vendor of the adapter, if available. Empty string otherwise.

architecture, of type DOMString, readonly

The name of the family or class of GPUs the adapter belongs to, if available. Empty string otherwise.

device, of type DOMString, readonly

A vendor-specific identifier for the adapter, if available. Empty string otherwise.

Note: This is a value that represents the type of adapter. For example, it may be a PCI device ID. It does not uniquely identify a given piece of hardware like a serial number.

description, of type DOMString, readonly

A human readable string describing the adapter as reported by the driver, if available. Empty string otherwise.

Note: Because no formatting is applied to description attempting to parse this value is not recommended. Applications which change their behavior based on the GPUAdapterInfo, such as applying workarounds for known driver issues, should rely on the other fields when possible.

To create a new adapter info for a given adapter adapter, run the following steps:
  1. Let adapterInfo be a new GPUAdapterInfo.

  2. Let unmaskedValues be adapter.[[unmaskedIdentifiers]]

  3. If unmaskedValues contains "vendor" and the vendor is known:

    1. Set adapterInfo.vendor to the name of adapter’s vendor as a normalized identifier string.

    Otherwise:

    1. Set adapterInfo.vendor to the empty string or a reasonable approximation of the vendor as a normalized identifier string.

  4. If unmaskedValues contains "architecture" and the architecture is known:

    1. Set adapterInfo.architecture to a normalized identifier string representing the family or class of adapters to which adapter belongs.

    Otherwise:

    1. Set adapterInfo.architecture to the empty string or a reasonable approximation of the architecture as a normalized identifier string.

  5. If unmaskedValues contains "device" and the device is known:

    1. Set adapterInfo.device to a normalized identifier string representing a vendor-specific identifier for adapter.

    Otherwise:

    1. Set adapterInfo.device to to the empty string or a reasonable approximation of a vendor-specific identifier as a normalized identifier string.

  6. If unmaskedValues contains "description" and a description is known:

    1. Set adapterInfo.description to a description of the adapter as reported by the driver.

    Otherwise:

    1. Set adapterInfo.description to the empty string or a reasonable approximation of a description.

  7. Return adapterInfo.

A normalized identifier string is one that follows the following pattern:

[a-z0-9]+(-[a-z0-9]+)*

Examples of valid normalized identifier strings include:
  • gpu

  • 3d

  • 0x3b2f

  • next-gen

  • series-x20-ultra

3.7. Extension Documents

"Extension Documents" are additional documents which describe new functionality which is non-normative and not part of the WebGPU/WGSL specifications. They describe functionality that builds upon these specifications, often including one or more new API feature flags and/or WGSL enable directives, or interactions with other draft web specifications.

WebGPU implementations must not expose extension functionality; doing so is a spec violation. New functionality does not become part of the WebGPU standard until it is integrated into the WebGPU specification (this document) and/or WGSL specification.

3.8. Origin Restrictions

WebGPU allows accessing image data stored in images, videos, and canvases. Restrictions are imposed on the use of cross-domain media, because shaders can be used to indirectly deduce the contents of textures which have been uploaded to the GPU.

WebGPU disallows uploading an image source if it is not origin-clean.

This also implies that the origin-clean flag for a canvas rendered using WebGPU will never be set to false.

For more information on issuing CORS requests for image and video elements, consult:

3.9. Task Sources

3.9.1. WebGPU Task Source

WebGPU defines a new task source called the WebGPU task source. It is used for the uncapturederror event and GPUDevice.lost.

To queue a global task for GPUDevice device, with a series of steps steps:
  1. Queue a global task on the WebGPU task source, with the global object that was used to create device, and the steps steps.

3.9.2. Automatic Expiry Task Source

WebGPU defines a new task source called the automatic expiry task source. It is used for the automatic, timed expiry (destruction) of certain objects:

To queue an automatic expiry task with GPUDevice device and a series of steps steps:
  1. Queue a global task on the automatic expiry task source, with the global object that was used to create device, and the steps steps.

Tasks from the automatic expiry task source should be processed with high priority; in particular, once queued, they should run before user-defined (JavaScript) tasks.

Note: This behavior is more predictable, and the strictness helps developers write more portable applications by eagerly detecting incorrect assumptions about implicit lifetimes that may be hard to detect. Developers are still strongly encouraged to test in multiple implementations.

Implementation note: It is valid to implement a high-priority expiry "task" by instead inserting additional steps at a fixed point inside the event loop processing model rather than running an actual task.

3.10. Color Spaces and Encoding

WebGPU does not provide color management. All values within WebGPU (such as texture elements) are raw numeric values, not color-managed color values.

WebGPU does interface with color-managed outputs (via GPUCanvasConfiguration) and inputs (via copyExternalImageToTexture() and importExternalTexture()). Thus, color conversion must be performed between the WebGPU numeric values and the external color values. Each such interface point locally defines an encoding (color space, transfer function, and alpha premultiplication) in which the WebGPU numeric values are to be interpreted.

WebGPU allows all of the color spaces in the PredefinedColorSpace enum. Note, each color space is defined over an extended range, as defined by the referenced CSS definitions, to represent color values outside of its space (in both chrominance and luminance).

An out-of-gamut premultiplied RGBA value is one where any of the R/G/B channel values exceeds the alpha channel value. For example, the premultiplied sRGB RGBA value [1.0, 0, 0, 0.5] represents the (unpremultiplied) color [2, 0, 0] with 50% alpha, written rgb(srgb 2 0 0 / 50%) in CSS. Just like any color value outside the sRGB color gamut, this is a well defined point in the extended color space (except when alpha is 0, in which case there is no color). However, when such values are output to a visible canvas, the result is undefined (see GPUCanvasAlphaMode "premultiplied").

3.10.1. Color Space Conversions

A color is converted between spaces by translating its representation in one space to a representation in another according to the definitions above.

If the source value has fewer than 4 RGBA channels, the missing green/blue/alpha channels are set to 0, 0, 1, respectively, before converting for color space/encoding and alpha premultiplication. After conversion, if the destination needs fewer than 4 channels, the additional channels are ignored.

Note: Grayscale images generally represent RGB values (V, V, V), or RGBA values (V, V, V, A) in their color space.

Colors are not lossily clamped during conversion: converting from one color space to another will result in values outside the range [0, 1] if the source color values were outside the range of the destination color space’s gamut. For an sRGB destination, for example, this can occur if the source is rgba16float, in a wider color space like Display-P3, or is premultiplied and contains out-of-gamut values.

Similarly, if the source value has a high bit depth (e.g. PNG with 16 bits per component) or extended range (e.g. canvas with float16 storage), these colors are preserved through color space conversion, with intermediate computations having at least the precision of the source.

3.10.2. Color Space Conversion Elision

If the source and destination of a color space/encoding conversion are the same, then conversion is not necessary. In general, if any given step of the conversion is an identity function (no-op), implementations should elide it, for performance.

For optimal performance, applications should set their color space and encoding options so that the number of necessary conversions is minimized throughout the process. For various image sources of GPUImageCopyExternalImage:

Note: Check browser implementation support for these features before relying on them.

3.11. Numeric conversions from JavaScript to WGSL

Several parts of the WebGPU API (pipeline-overridable constants and render pass clear values) take numeric values from WebIDL (double or float) and convert them to WGSL values (bool, i32, u32, f32, f16).

To convert an IDL value idlValue of type double or float to WGSL type T, possibly throwing a TypeError:

Note: This TypeError is generated in the device timeline and never surfaced to JavaScript.

  1. Assert idlValue is a finite value, since it is not unrestricted double or unrestricted float.

  2. Let v be the ECMAScript Number resulting from ! converting idlValue to an ECMAScript value.

  3. If T is bool

    Return the WGSL bool value corresponding to the result of ! converting v to an IDL value of type boolean.

    Note: This algorithm is called after the conversion from an ECMAScript value to an IDL double or float value. If the original ECMAScript value was a non-numeric, non-boolean value like [] or {}, then the WGSL bool result may be different than if the ECMAScript value had been converted to IDL boolean directly.

    If T is i32

    Return the WGSL i32 value corresponding to the result of ? converting v to an IDL value of type [EnforceRange] long.

    If T is u32

    Return the WGSL u32 value corresponding to the result of ? converting v to an IDL value of type [EnforceRange] unsigned long.

    If T is f32

    Return the WGSL f32 value corresponding to the result of ? converting v to an IDL value of type float.

    If T is f16
    1. Let wgslF32 be the WGSL f32 value corresponding to the result of ? converting v to an IDL value of type float.

    2. Return f16(wgslF32), the result of ! converting the WGSL f32 value to f16 as defined in WGSL floating point conversion.

    Note: As long as the value is in-range of f32, no error is thrown, even if the value is out-of-range of f16.

To convert a GPUColor color to a texel value of texture format format, possibly throwing a TypeError:

Note: This TypeError is generated in the device timeline and never surfaced to JavaScript.

  1. If the components of format (assert they all have the same type) are:

    floating-point types or normalized types

    Let T be f32.

    signed integer types

    Let T be i32.

    unsigned integer types

    Let T be u32.

  2. Let wgslColor be a WGSL value of type vec4<T>, where the 4 components are the RGBA channels of color, each ? converted to WGSL type T.

  3. Convert wgslColor to format using the same conversion rules as the § 23.3.7 Output Merging step, and return the result.

    Note: For non-integer types, the exact choice of value is implementation-defined. For normalized types, the value is clamped to the range of the type.

Note: In other words, the value written will be as if it was written by a WGSL shader that outputs the value represented as a vec4 of f32, i32, or u32.

4. Initialization

A GPU object is available in the Window and DedicatedWorkerGlobalScope contexts through the Navigator and WorkerNavigator interfaces respectively and is exposed via navigator.gpu:

interface mixin NavigatorGPU {
    [SameObject, SecureContext] readonly attribute GPU gpu;
};
Navigator includes NavigatorGPU;
WorkerNavigator includes NavigatorGPU;

NavigatorGPU has the following attributes:

gpu, of type GPU, readonly

A global singleton providing top-level entry points like requestAdapter().

4.2. GPU

GPU is the entry point to WebGPU.

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPU {
    Promise<GPUAdapter?> requestAdapter(optional GPURequestAdapterOptions options = {});
    GPUTextureFormat getPreferredCanvasFormat();
    [SameObject] readonly attribute WGSLLanguageFeatures wgslLanguageFeatures;
};

GPU has the following methods and attributes:

requestAdapter(options)

Requests an adapter from the user agent. The user agent chooses whether to return an adapter, and, if so, chooses according to the provided options.

Called on: GPU this.

Arguments:

Arguments for the GPU.requestAdapter(options) method.
Parameter Type Nullable Optional Description
options GPURequestAdapterOptions Criteria used to select the adapter.

Returns: Promise<GPUAdapter?>

Content timeline steps:

  1. Let contentTimeline be the current Content timeline.

  2. Let promise be a new promise.

  3. Issue the initialization steps on the Device timeline of this.

  4. Return promise.

Device timeline initialization steps:
  1. Let adapter be null.

  2. If the user agent chooses to return an adapter, it should:

    1. Set adapter to a valid adapter, chosen according to the rules in § 4.2.2 Adapter Selection and the criteria in options, adhering to § 4.2.1 Adapter Capability Guarantees.

      The supported limits of the adapter must adhere to the requirements defined in § 3.6.2 Limits.

    2. If adapter meets the criteria of a fallback adapter set adapter.[[fallback]] to true.

  3. Issue the subsequent steps on contentTimeline.

Content timeline steps:
  1. If adapter is not null:

    1. Resolve promise with a new GPUAdapter encapsulating adapter.

  2. Otherwise, Resolve promise with null.

getPreferredCanvasFormat()

Returns an optimal GPUTextureFormat for displaying 8-bit depth, standard dynamic range content on this system. Must only return "rgba8unorm" or "bgra8unorm".

The returned value can be passed as the format to configure() calls on a GPUCanvasContext to ensure the associated canvas is able to display its contents efficiently.

Note: Canvases which are not displayed to the screen may or may not benefit from using this format.

Called on: GPU this.

Returns: GPUTextureFormat

Content timeline steps:

  1. Return either "rgba8unorm" or "bgra8unorm", depending on which format is optimal for displaying WebGPU canvases on this system.

wgslLanguageFeatures, of type WGSLLanguageFeatures, readonly

The names of supported WGSL language extensions. Supported language extensions are automatically enabled.

Adapters may become invalid ("expire") at any time. Upon any change in the system’s state that could affect the result of any requestAdapter() call, the user agent should expire all previously-returned adapters. For example:

Note: User agents may choose to expire adapters often, even when there has been no system state change (e.g. seconds or minutes after the adapter was created). This can help obfuscate real system state changes, and make developers more aware that calling requestAdapter() again is always necessary before calling requestDevice(). If an application does encounter this situation, standard device-loss recovery handling should allow it to recover.

Requesting a GPUAdapter with no hints:
const gpuAdapter = await navigator.gpu.requestAdapter();

4.2.1. Adapter Capability Guarantees

Any GPUAdapter returned by requestAdapter() must provide the following guarantees:

4.2.2. Adapter Selection

GPURequestAdapterOptions provides hints to the user agent indicating what configuration is suitable for the application.

dictionary GPURequestAdapterOptions {
    GPUPowerPreference powerPreference;
    boolean forceFallbackAdapter = false;
};
enum GPUPowerPreference {
    "low-power",
    "high-performance",
};

GPURequestAdapterOptions has the following members:

powerPreference, of type GPUPowerPreference

Optionally provides a hint indicating what class of adapter should be selected from the system’s available adapters.

The value of this hint may influence which adapter is chosen, but it must not influence whether an adapter is returned or not.

Note: The primary utility of this hint is to influence which GPU is used in a multi-GPU system. For instance, some laptops have a low-power integrated GPU and a high-performance discrete GPU. This hint may also affect the power configuration of the selected GPU to match the requested power preference.

Note: Depending on the exact hardware configuration, such as battery status and attached displays or removable GPUs, the user agent may select different adapters given the same power preference. Typically, given the same hardware configuration and state and powerPreference, the user agent is likely to select the same adapter.

It must be one of the following values:

undefined (or not present)

Provides no hint to the user agent.

"low-power"

Indicates a request to prioritize power savings over performance.

Note: Generally, content should use this if it is unlikely to be constrained by drawing performance; for example, if it renders only one frame per second, draws only relatively simple geometry with simple shaders, or uses a small HTML canvas element. Developers are encouraged to use this value if their content allows, since it may significantly improve battery life on portable devices.

"high-performance"

Indicates a request to prioritize performance over power consumption.

Note: By choosing this value, developers should be aware that, for devices created on the resulting adapter, user agents are more likely to force device loss, in order to save power by switching to a lower-power adapter. Developers are encouraged to only specify this value if they believe it is absolutely necessary, since it may significantly decrease battery life on portable devices.

forceFallbackAdapter, of type boolean, defaulting to false

When set to true indicates that only a fallback adapter may be returned. If the user agent does not support a fallback adapter, will cause requestAdapter() to resolve to null.

Note: requestAdapter() may still return a fallback adapter if forceFallbackAdapter is set to false and either no other appropriate adapter is available or the user agent chooses to return a fallback adapter. Developers that wish to prevent their applications from running on fallback adapters should check the GPUAdapter.isFallbackAdapter attribute prior to requesting a GPUDevice.

Requesting a "high-performance" GPUAdapter:
const gpuAdapter = await navigator.gpu.requestAdapter({
    powerPreference: 'high-performance'
});

4.3. GPUAdapter

A GPUAdapter encapsulates an adapter, and describes its capabilities (features and limits).

To get a GPUAdapter, use requestAdapter().

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUAdapter {
    [SameObject] readonly attribute GPUSupportedFeatures features;
    [SameObject] readonly attribute GPUSupportedLimits limits;
    readonly attribute boolean isFallbackAdapter;

    Promise<GPUDevice> requestDevice(optional GPUDeviceDescriptor descriptor = {});
    Promise<GPUAdapterInfo> requestAdapterInfo(optional sequence<DOMString> unmaskHints = []);
};

GPUAdapter has the following attributes:

features, of type GPUSupportedFeatures, readonly

The set of values in this.[[adapter]].[[features]].

limits, of type GPUSupportedLimits, readonly

The limits in this.[[adapter]].[[limits]].

isFallbackAdapter, of type boolean, readonly

Returns the value of [[adapter]].[[fallback]].

GPUAdapter has the following internal slots:

[[adapter]], of type adapter, readonly

The adapter to which this GPUAdapter refers.

GPUAdapter has the following methods:

requestDevice(descriptor)

Requests a device from the adapter.

This is a one-time action: if a device is returned successfully, the adapter becomes invalid.

Called on: GPUAdapter this.

Arguments:

Arguments for the GPUAdapter.requestDevice(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPUDeviceDescriptor Description of the GPUDevice to request.

Returns: Promise<GPUDevice>

Content timeline steps:

  1. Let contentTimeline be the current Content timeline.

  2. Let promise be a new promise.

  3. Let adapter be this.[[adapter]].

  4. Issue the initialization steps to the Device timeline of this.

  5. Return promise.

Device timeline initialization steps:
  1. If any of the following requirements are unmet:

    Then issue the following steps on contentTimeline and return:

    Content timeline steps:
    1. Reject promise with a TypeError.

    Note: This is the same error that is produced if a feature name isn’t known by the browser at all (in its GPUFeatureName definition). This converges the behavior when the browser doesn’t support a feature with the behavior when a particular adapter doesn’t support a feature.

  2. If any of the following requirements are unmet:

    Then issue the following steps on contentTimeline and return:

    Content timeline steps:
    1. Reject promise with an OperationError.

  3. If adapter is invalid, or the user agent otherwise cannot fulfill the request:

    1. Let device be a new device.

    2. Lose the device(device, "unknown").

      Note: This makes adapter invalid, if it wasn’t already.

      Note: User agents should consider issuing developer-visible warnings in most or all cases when this occurs. Applications should perform reinitialization logic starting with requestAdapter().

    Otherwise:

    1. Let device be a new device with the capabilities described by descriptor.

    2. Make adapter.[[adapter]] invalid.

  4. Issue the subsequent steps on contentTimeline.

Content timeline steps:
  1. Resolve promise with a new GPUDevice object device.

    Note: If the device is already lost because the adapter could not fulfill the request, device.lost has already resolved before promise resolves.

requestAdapterInfo()

Requests the GPUAdapterInfo for this GPUAdapter.

Note: Adapter info values are returned with a Promise to give user agents an opportunity to perform potentially long-running checks when requesting unmasked values, such as asking for user consent before returning. If no unmaskHints are specified, however, no dialogs should be displayed to the user.

Called on: GPUAdapter this.

Arguments:

Arguments for the GPUAdapter.requestAdapterInfo() method.
Parameter Type Nullable Optional Description
unmaskHints sequence<DOMString> A list of GPUAdapterInfo attribute names for which unmasked values are desired if available.

Returns: Promise<GPUAdapterInfo>

Content timeline steps:

  1. Let promise be a new promise.

  2. Let adapter be this.[[adapter]].

  3. Let hasActivation be true if the relevant global object for this has transient activation, and false otherwise.

  4. Run the following steps in parallel:

    1. If unmaskHints.length > 0:

      1. If hasActivation is false reject promise with a NotAllowedError and abort these steps.

      2. Let unmaskedKeys be a list of the fields specified in unmaskHints which the user agent decides to unmask, if any.

        Note: The user agent is free to use any method it deems appropriate to decide which fields to unmask.

      3. Append unmaskedKeys to adapter.[[unmaskedIdentifiers]].

    2. Resolve promise with a new adapter info for adapter.

  5. Return promise.

Requesting a GPUDevice with default features and limits:
const gpuAdapter = await navigator.gpu.requestAdapter();
const gpuDevice = await gpuAdapter.requestDevice();

4.3.1. GPUDeviceDescriptor

GPUDeviceDescriptor describes a device request.

dictionary GPUDeviceDescriptor
         : GPUObjectDescriptorBase {
    sequence<GPUFeatureName> requiredFeatures = [];
    record<DOMString, GPUSize64> requiredLimits = {};
    GPUQueueDescriptor defaultQueue = {};
};

GPUDeviceDescriptor has the following members:

requiredFeatures, of type sequence<GPUFeatureName>, defaulting to []

Specifies the features that are required by the device request. The request will fail if the adapter cannot provide these features.

Exactly the specified set of features, and no more or less, will be allowed in validation of API calls on the resulting device.

requiredLimits, of type record<DOMString, GPUSize64>, defaulting to {}

Specifies the limits that are required by the device request. The request will fail if the adapter cannot provide these limits.

Each key must be the name of a member of supported limits. Exactly the specified limits, and no better or worse, will be allowed in validation of API calls on the resulting device.

defaultQueue, of type GPUQueueDescriptor, defaulting to {}

The descriptor for the default GPUQueue.

Requesting a GPUDevice with the "texture-compression-astc" feature if supported:
const gpuAdapter = await navigator.gpu.requestAdapter();

const requiredFeatures = [];
if (gpuAdapter.features.has('texture-compression-astc')) {
    requiredFeatures.push('texture-compression-astc')
}

const gpuDevice = await gpuAdapter.requestDevice({
    requiredFeatures
});
4.3.1.1. GPUFeatureName

Each GPUFeatureName identifies a set of functionality which, if available, allows additional usages of WebGPU that would have otherwise been invalid.

enum GPUFeatureName {
    "depth-clip-control",
    "depth32float-stencil8",
    "texture-compression-bc",
    "texture-compression-etc2",
    "texture-compression-astc",
    "timestamp-query",
    "indirect-first-instance",
    "shader-f16",
    "rg11b10ufloat-renderable",
    "bgra8unorm-storage",
    "float32-filterable",
};

4.4. GPUDevice

A GPUDevice encapsulates a device and exposes the functionality of that device.

GPUDevice is the top-level interface through which WebGPU interfaces are created.

To get a GPUDevice, use requestDevice().

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUDevice : EventTarget {
    [SameObject] readonly attribute GPUSupportedFeatures features;
    [SameObject] readonly attribute GPUSupportedLimits limits;

    [SameObject] readonly attribute GPUQueue queue;

    undefined destroy();

    GPUBuffer createBuffer(GPUBufferDescriptor descriptor);
    GPUTexture createTexture(GPUTextureDescriptor descriptor);
    GPUSampler createSampler(optional GPUSamplerDescriptor descriptor = {});
    GPUExternalTexture importExternalTexture(GPUExternalTextureDescriptor descriptor);

    GPUBindGroupLayout createBindGroupLayout(GPUBindGroupLayoutDescriptor descriptor);
    GPUPipelineLayout createPipelineLayout(GPUPipelineLayoutDescriptor descriptor);
    GPUBindGroup createBindGroup(GPUBindGroupDescriptor descriptor);

    GPUShaderModule createShaderModule(GPUShaderModuleDescriptor descriptor);
    GPUComputePipeline createComputePipeline(GPUComputePipelineDescriptor descriptor);
    GPURenderPipeline createRenderPipeline(GPURenderPipelineDescriptor descriptor);
    Promise<GPUComputePipeline> createComputePipelineAsync(GPUComputePipelineDescriptor descriptor);
    Promise<GPURenderPipeline> createRenderPipelineAsync(GPURenderPipelineDescriptor descriptor);

    GPUCommandEncoder createCommandEncoder(optional GPUCommandEncoderDescriptor descriptor = {});
    GPURenderBundleEncoder createRenderBundleEncoder(GPURenderBundleEncoderDescriptor descriptor);

    GPUQuerySet createQuerySet(GPUQuerySetDescriptor descriptor);
};
GPUDevice includes GPUObjectBase;

GPUDevice has the following attributes:

features, of type GPUSupportedFeatures, readonly

A set containing the GPUFeatureName values of the features supported by the device (i.e. the ones with which it was created).

limits, of type GPUSupportedLimits, readonly

Exposes the limits supported by the device (which are exactly the ones with which it was created).

queue, of type GPUQueue, readonly

The primary GPUQueue for this device.

The [[device]] for a GPUDevice is the device that the GPUDevice refers to.

GPUDevice has the methods listed in its WebIDL definition above. Those not defined here are defined elsewhere in this document.

destroy()

Destroys the device, preventing further operations on it. Outstanding asynchronous operations will fail.

Note: It is valid to destroy a device multiple times.

Called on: GPUDevice this.

Content timeline steps:

  1. unmap() all GPUBuffers from this device.

  2. Issue the subsequent steps on the Device timeline of this.

Device timeline steps:
  1. Once all currently-enqueued operations on any queue on this device are completed, issue the subsequent steps on the current timeline.

  1. Lose the device(this.[[device]], "destroyed").

Note: Since no further operations can be enqueued on this device, implementations can abort outstanding asynchronous operations immediately and free resource allocations, including mapped memory that was just unmapped.

A GPUDevice's allowed buffer usages are:
A GPUDevice's allowed texture usages are:

4.5. Example

A more robust example of requesting a GPUAdapter and GPUDevice with error handling:
let gpuDevice = null;

async function initializeWebGPU() {
    // Check to ensure the user agent supports WebGPU.
    if (!('gpu' in navigator)) {
        console.error("User agent doesn’t support WebGPU.");
        return false;
    }

    // Request an adapter.
    const gpuAdapter = await navigator.gpu.requestAdapter();

    // requestAdapter may resolve with null if no suitable adapters are found.
    if (!gpuAdapter) {
        console.error('No WebGPU adapters found.');
        return false;
    }

    // Request a device.
    // Note that the promise will reject if invalid options are passed to the optional
    // dictionary. To avoid the promise rejecting always check any features and limits
    // against the adapters features and limits prior to calling requestDevice().
    gpuDevice = await gpuAdapter.requestDevice();

    // requestDevice will never return null, but if a valid device request can’t be
    // fulfilled for some reason it may resolve to a device which has already been lost.
    // Additionally, devices can be lost at any time after creation for a variety of reasons
    // (ie: browser resource management, driver updates), so it’s a good idea to always
    // handle lost devices gracefully.
    gpuDevice.lost.then((info) => {
        console.error(`WebGPU device was lost: ${info.message}`);

        gpuDevice = null;

        // Many causes for lost devices are transient, so applications should try getting a
        // new device once a previous one has been lost unless the loss was caused by the
        // application intentionally destroying the device. Note that any WebGPU resources
        // created with the previous device (buffers, textures, etc) will need to be
        // re-created with the new one.
        if (info.reason != 'destroyed') {
            initializeWebGPU();
        }
    });

    onWebGPUInitialized();

    return true;
}

function onWebGPUInitialized() {
    // Begin creating WebGPU resources here...
}

initializeWebGPU();

5. Buffers

5.1. GPUBuffer

A GPUBuffer represents a block of memory that can be used in GPU operations. Data is stored in linear layout, meaning that each byte of the allocation can be addressed by its offset from the start of the GPUBuffer, subject to alignment restrictions depending on the operation. Some GPUBuffers can be mapped which makes the block of memory accessible via an ArrayBuffer called its mapping.

GPUBuffers are created via createBuffer(). Buffers may be mappedAtCreation.

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUBuffer {
    readonly attribute GPUSize64Out size;
    readonly attribute GPUFlagsConstant usage;

    readonly attribute GPUBufferMapState mapState;

    Promise<undefined> mapAsync(GPUMapModeFlags mode, optional GPUSize64 offset = 0, optional GPUSize64 size);
    ArrayBuffer getMappedRange(optional GPUSize64 offset = 0, optional GPUSize64 size);
    undefined unmap();

    undefined destroy();
};
GPUBuffer includes GPUObjectBase;

enum GPUBufferMapState {
    "unmapped",
    "pending",
    "mapped",
};

GPUBuffer has the following immutable properties:

size, of type GPUSize64Out, readonly

The length of the GPUBuffer allocation in bytes.

usage, of type GPUFlagsConstant, readonly

The allowed usages for this GPUBuffer.

[[internals]], of type buffer internals, readonly, override

GPUBuffer has the following content timeline properties:

mapState, of type GPUBufferMapState, readonly

The current GPUBufferMapState of the buffer:

"unmapped"

The buffer is not mapped for use by this.getMappedRange().

"pending"

A mapping of the buffer has been requested, but is pending. It may succeed, or fail validation in mapAsync().

"mapped"

The buffer is mapped and this.getMappedRange() may be used.

The getter steps are:

Content timeline steps:
  1. If this.[[mapping]] is not null, return "mapped".

  2. If this.[[pending_map]] is not null, return "pending".

  3. Return "unmapped".

[[pending_map]], of type Promise<void> or null, initially null

The Promise returned by the currently-pending mapAsync() call.

There is never more than one pending map, because mapAsync() will refuse immediately if a request is already in flight.

[[mapping]], of type active buffer mapping or null, initially null

Set if and only if the buffer is currently mapped for use by getMappedRange(). Null otherwise (even if there is a [[pending_map]]).

An active buffer mapping is a structure with the following fields:

data, of type Data Block

The mapping for this GPUBuffer. This data is accessed through ArrayBuffers which are views onto this data, returned by getMappedRange() and stored in views.

mode, of type GPUMapModeFlags

The GPUMapModeFlags of the map, as specified in the corresponding call to mapAsync() or createBuffer().

range, of type tuple [unsigned long long, unsigned long long]

The range of this GPUBuffer that is mapped.

views, of type list<ArrayBuffer>

The ArrayBuffers returned via getMappedRange() to the application. They are tracked so they can be detached when unmap() is called.

To initialize an active buffer mapping with mode mode and range range:
  1. Let size be range[1] - range[0].

  2. Let data be ? CreateByteDataBlock(size).

    Note: This may result in a RangeError being thrown. For consistency and predictability:
    • For any size at which new ArrayBuffer() would succeed at a given moment, this allocation should succeed at that moment.

    • For any size at which new ArrayBuffer() deterministically throws a RangeError, this allocation should as well.

  3. Return an active buffer mapping with:

GPUBuffer's internal object is buffer internals, which extends internal object with the following device timeline slots:

state

The current internal state of the buffer:

"available"

The buffer may be used in queue operations (unless it is invalid).

"unavailable"

The buffer may not be used in queue operations due to being mapped.

"destroyed"

The buffer may not be used in any operations due to being destroy()ed.

Mapping and unmapping a buffer.
Failing to map a buffer.

5.1.1. GPUBufferDescriptor

dictionary GPUBufferDescriptor
         : GPUObjectDescriptorBase {
    required GPUSize64 size;
    required GPUBufferUsageFlags usage;
    boolean mappedAtCreation = false;
};

GPUBufferDescriptor has the following members:

size, of type GPUSize64

The size of the buffer in bytes.

usage, of type GPUBufferUsageFlags

The allowed usages for the buffer.

mappedAtCreation, of type boolean, defaulting to false

If true creates the buffer in an already mapped state, allowing getMappedRange() to be called immediately. It is valid to set mappedAtCreation to true even if usage does not contain MAP_READ or MAP_WRITE. This can be used to set the buffer’s initial data.

Guarantees that even if the buffer creation eventually fails, it will still appear as if the mapped range can be written/read to until it is unmapped.

5.1.2. Buffer Usages

typedef [EnforceRange] unsigned long GPUBufferUsageFlags;
[Exposed=(Window, DedicatedWorker), SecureContext]
namespace GPUBufferUsage {
    const GPUFlagsConstant MAP_READ      = 0x0001;
    const GPUFlagsConstant MAP_WRITE     = 0x0002;
    const GPUFlagsConstant COPY_SRC      = 0x0004;
    const GPUFlagsConstant COPY_DST      = 0x0008;
    const GPUFlagsConstant INDEX         = 0x0010;
    const GPUFlagsConstant VERTEX        = 0x0020;
    const GPUFlagsConstant UNIFORM       = 0x0040;
    const GPUFlagsConstant STORAGE       = 0x0080;
    const GPUFlagsConstant INDIRECT      = 0x0100;
    const GPUFlagsConstant QUERY_RESOLVE = 0x0200;
};

The GPUBufferUsage flags determine how a GPUBuffer may be used after its creation:

MAP_READ

The buffer can be mapped for reading. (Example: calling mapAsync() with GPUMapMode.READ)

May only be combined with COPY_DST.

MAP_WRITE

The buffer can be mapped for writing. (Example: calling mapAsync() with GPUMapMode.WRITE)

May only be combined with COPY_SRC.

COPY_SRC

The buffer can be used as the source of a copy operation. (Examples: as the source argument of a copyBufferToBuffer() or copyBufferToTexture() call.)

COPY_DST

The buffer can be used as the destination of a copy or write operation. (Examples: as the destination argument of a copyBufferToBuffer() or copyTextureToBuffer() call, or as the target of a writeBuffer() call.)

INDEX

The buffer can be used as an index buffer. (Example: passed to setIndexBuffer().)

VERTEX

The buffer can be used as a vertex buffer. (Example: passed to setVertexBuffer().)

UNIFORM

The buffer can be used as a uniform buffer. (Example: as a bind group entry for a GPUBufferBindingLayout with a buffer.type of "uniform".)

STORAGE

The buffer can be used as a storage buffer. (Example: as a bind group entry for a GPUBufferBindingLayout with a buffer.type of "storage" or "read-only-storage".)

INDIRECT

The buffer can be used as to store indirect command arguments. (Examples: as the indirectBuffer argument of a drawIndirect() or dispatchWorkgroupsIndirect() call.)

QUERY_RESOLVE

The buffer can be used to capture query results. (Example: as the destination argument of a resolveQuerySet() call.)

5.1.3. Buffer Creation

createBuffer(descriptor)

Creates a GPUBuffer.

Called on: GPUDevice this.

Arguments:

Arguments for the GPUDevice.createBuffer(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPUBufferDescriptor Description of the GPUBuffer to create.

Returns: GPUBuffer

Content timeline steps:

  1. Let [b, bi] be ! create a new WebGPU object(this, GPUBuffer, descriptor).

  2. Set b.size to descriptor.size.

  3. Set b.usage to descriptor.usage.

  4. If descriptor.mappedAtCreation is true:

    1. Set b.[[mapping]] to ? initialize an active buffer mapping with mode WRITE and range [0, descriptor.size].

  5. Issue the initialization steps on the Device timeline of this.

  6. Return b.

Device timeline initialization steps:
  1. If any of the following requirements are unmet, generate a validation error, make bi invalid, and stop.

Note: If buffer creation fails, and descriptor.mappedAtCreation is false, any calls to mapAsync() will reject, so any resources allocated to enable mapping can and may be discarded or recycled.

  1. If descriptor.mappedAtCreation is true:

    1. Set bi.state to "unavailable".

    Else:

    1. Set bi.state to "available".

  2. Create a device allocation for bi where each byte is zero.

    If the allocation fails without side-effects, generate an out-of-memory error, make bi invalid, and return.

Creating a 128 byte uniform buffer that can be written into:
const buffer = gpuDevice.createBuffer({
    size: 128,
    usage: GPUBufferUsage.UNIFORM | GPUBufferUsage.COPY_DST
});

5.1.4. Buffer Destruction

An application that no longer requires a GPUBuffer can choose to lose access to it before garbage collection by calling destroy(). Destroying a buffer also unmaps it, freeing any memory allocated for the mapping.

Note: This allows the user agent to reclaim the GPU memory associated with the GPUBuffer once all previously submitted operations using it are complete.

destroy()

Destroys the GPUBuffer.

Note: It is valid to destroy a buffer multiple times.

Called on: GPUBuffer this.

Returns: undefined

Content timeline steps:

  1. Call this.unmap().

  2. Issue the subsequent steps on the Device timeline of this.[[device]].

Device timeline steps:
  1. Set this.[[internals]].state to "destroyed".

Note: Since no further operations can be enqueued using this buffer, implementations can free resource allocations, including mapped memory that was just unmapped.

5.2. Buffer Mapping

An application can request to map a GPUBuffer so that they can access its content via ArrayBuffers that represent part of the GPUBuffer's allocations. Mapping a GPUBuffer is requested asynchronously with mapAsync() so that the user agent can ensure the GPU finished using the GPUBuffer before the application can access its content. A mapped GPUBuffer cannot be used by the GPU and must be unmapped using unmap() before work using it can be submitted to the Queue timeline.

Once the GPUBuffer is mapped, the application can synchronously ask for access to ranges of its content with getMappedRange(). The returned ArrayBuffer can only be detached by unmap() (directly, or via GPUBuffer.destroy() or GPUDevice.destroy()), and cannot be transferred. A TypeError is thrown by any other operation that attempts to do so.

typedef [EnforceRange] unsigned long GPUMapModeFlags;
[Exposed=(Window, DedicatedWorker), SecureContext]
namespace GPUMapMode {
    const GPUFlagsConstant READ  = 0x0001;
    const GPUFlagsConstant WRITE = 0x0002;
};

The GPUMapMode flags determine how a GPUBuffer is mapped when calling mapAsync():

READ

Only valid with buffers created with the MAP_READ usage.

Once the buffer is mapped, calls to getMappedRange() will return an ArrayBuffer containing the buffer’s current values. Changes to the returned ArrayBuffer will be discarded after unmap() is called.

WRITE

Only valid with buffers created with the MAP_WRITE usage.

Once the buffer is mapped, calls to getMappedRange() will return an ArrayBuffer containing the buffer’s current values. Changes to the returned ArrayBuffer will be stored in the GPUBuffer after unmap() is called.

Note: Since the MAP_WRITE buffer usage may only be combined with the COPY_SRC buffer usage, mapping for writing can never return values produced by the GPU, and the returned ArrayBuffer will only ever contain the default initialized data (zeros) or data written by the webpage during a previous mapping.

mapAsync(mode, offset, size)

Maps the given range of the GPUBuffer and resolves the returned Promise when the GPUBuffer's content is ready to be accessed with getMappedRange().

The resolution of the returned Promise only indicates that the buffer has been mapped. It does not guarantee the completion of any other operations visible to the content timeline, and in particular does not imply that any other Promise returned from onSubmittedWorkDone() or mapAsync() on other GPUBuffers have resolved.

The resolution of the Promise returned from onSubmittedWorkDone() does imply the completion of mapAsync() calls made prior to that call, on GPUBuffers last used exclusively on that queue.

Called on: GPUBuffer this.

Arguments:

Arguments for the GPUBuffer.mapAsync(mode, offset, size) method.
Parameter Type Nullable Optional Description
mode GPUMapModeFlags Whether the buffer should be mapped for reading or writing.
offset GPUSize64 Offset in bytes into the buffer to the start of the range to map.
size GPUSize64 Size in bytes of the range to map.

Returns: Promise<undefined>

Content timeline steps:

  1. Let contentTimeline be the current Content timeline.

  2. If this.[[pending_map]] is not null:

    1. Return a promise rejected with OperationError.

  3. Let p be a new Promise.

  4. Set this.[[pending_map]] to p.

  5. Issue the validation steps on the Device timeline of this.[[device]].

  6. Return p.

Device timeline validation steps:
  1. If size is undefined:

    1. Let rangeSize be max(0, this.size - offset).

    Otherwise:

    1. Let rangeSize be size.

  2. If any of the following conditions are unsatisfied:

    Then:

    1. Issue the map failure steps on contentTimeline.

    2. Generate a validation error.

    3. Return.

  3. Set this.[[internals]].state to "unavailable".

    Note: Since the buffer is mapped, its contents cannot change between this completion and unmap().

  4. If this.[[device]] is lost, or when it becomes lost:

    1. Issue the map failure steps on contentTimeline.

    Otherwise, at an unspecified point:

    • after the completion of currently-enqueued operations that use this,

    • and no later than the next device timeline operation after the device timeline becomes informed of the completion of all currently-enqueued operations (regardless of whether they use this),

    run the following steps:

    1. Let internalStateAtCompletion be this.[[internals]].state.

      Note: If, and only if, at this point the buffer has become "available" again due to an unmap() call, then [[pending_map]] != p below, so mapping will not succeed in the steps below.

    2. Let dataForMappedRegion be the contents of this starting at offset offset, for rangeSize bytes.

    3. Issue the map success steps on the contentTimeline.

Content timeline map success steps:
  1. If this.[[pending_map]] != p:

    Note: The map has been cancelled by unmap().

    1. Assert p is rejected.

    2. Return.

  2. Assert p is pending.

  3. Assert internalStateAtCompletion is "unavailable".

  4. Let mapping be initialize an active buffer mapping with mode mode and range [offset, offset + rangeSize].

    If this allocation fails:

    1. Set this.[[pending_map]] to null, and reject p with a RangeError.

    2. Return.

  5. Set the content of mapping.data to dataForMappedRegion.

  6. Set this.[[mapping]] to mapping.

  7. Set this.[[pending_map]] to null, and resolve p.

Content timeline map failure steps:
  1. If this.[[pending_map]] != p:

    Note: The map has been cancelled by unmap().

    1. Assert p is already rejected.

    2. Return.

  2. Assert p is still pending.

  3. Set this.[[pending_map]] to null, and reject p with an OperationError.

getMappedRange(offset, size)

Returns an ArrayBuffer with the contents of the GPUBuffer in the given mapped range.

Called on: GPUBuffer this.

Arguments:

Arguments for the GPUBuffer.getMappedRange(offset, size) method.
Parameter Type Nullable Optional Description
offset GPUSize64 Offset in bytes into the buffer to return buffer contents from.
size GPUSize64 Size in bytes of the ArrayBuffer to return.

Returns: ArrayBuffer

Content timeline steps:

  1. If size is missing:

    1. Let rangeSize be max(0, this.size - offset).

    Otherwise, let rangeSize be size.

  2. If any of the following conditions are unsatisfied, throw an OperationError and stop.

    Note: It is always valid to get mapped ranges of a GPUBuffer that is mappedAtCreation, even if it is invalid, because the Content timeline might not know it is invalid.

  3. Let data be this.[[mapping]].data.

  4. Let view be ! create an ArrayBuffer of size rangeSize, but with its pointer mutably referencing the content of data at offset (offset - [[mapping]].range[0]).

    Note: A RangeError may not be thrown here, because the data has already been allocated during mapAsync() or createBuffer().

  5. Set view.[[ArrayBufferDetachKey]] to "WebGPUBufferMapping".

    Note: This causes a TypeError to be thrown if an attempt is made to DetachArrayBuffer, except by unmap().

  6. Append view to this.[[mapping]].views.

  7. Return view.

Note: User agents should consider issuing a developer-visible warning if getMappedRange() succeeds without having checked the status of the map, by waiting for mapAsync() to succeed, querying a mapState of "mapped", or waiting for a later onSubmittedWorkDone() call to succeed.

unmap()

Unmaps the mapped range of the GPUBuffer and makes it’s contents available for use by the GPU again.

Called on: GPUBuffer this.

Returns: undefined

Content timeline steps:

  1. If this.[[pending_map]] is not null:

    1. Reject this.[[pending_map]] with an AbortError.

    2. Set this.[[pending_map]] to null.

  2. If this.[[mapping]] is null:

    1. Return.

  3. For each ArrayBuffer ab in this.[[mapping]].views:

    1. Perform DetachArrayBuffer(ab, "WebGPUBufferMapping").

  4. Let bufferUpdate be null.

  5. If this.[[mapping]].mode contains WRITE:

    1. Set bufferUpdate to { data: this.[[mapping]].data, offset: this.[[mapping]].range[0] }.

    Note: When a buffer is mapped without the WRITE mode, then unmapped, any local modifications done by the application to the mapped ranges ArrayBuffer are discarded and will not affect the content of later mappings.

  6. Set this.[[mapping]] to null.

  7. Issue the subsequent steps on the Device timeline of this.[[device]].

Device timeline steps:
  1. If this.[[device]] is invalid, return.

  2. If bufferUpdate is not null:

    1. Issue the following steps on the Queue timeline of this.[[device]].queue:

      Queue timeline steps:
      1. Update the contents of this at offset bufferUpdate.offset with the data bufferUpdate.data.

  3. Set this.[[internals]].state to "available".

6. Textures and Texture Views

6.1. GPUTexture

Remove this definition: texture

One texture consists of one or more texture subresources, each uniquely identified by a mipmap level and, for 2d textures only, array layer and aspect.

A texture subresource is a subresource: each can be used in different internal usages within a single usage scope.

Each subresource in a mipmap level is approximately half the size, in each spatial dimension, of the corresponding resource in the lesser level (see logical miplevel-specific texture extent). The subresource in level 0 has the dimensions of the texture itself. These are typically used to represent levels of detail of a texture. GPUSampler and WGSL provide facilities for selecting and interpolating between levels of detail, explicitly or automatically.

A "2d" texture may be an array of array layers. Each subresource in a layer is the same size as the corresponding resources in other layers. For non-2d textures, all subresources have an array layer index of 0.

Each subresource has an aspect. Color textures have just one aspect: color. Depth-or-stencil format textures may have multiple aspects: a depth aspect, a stencil aspect, or both, and may be used in special ways, such as in depthStencilAttachment and in "depth" bindings.

A "3d" texture may have multiple slices, each being the two-dimensional image at a particular z value in the texture. Slices are not separate subresources.

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUTexture {
    GPUTextureView createView(optional GPUTextureViewDescriptor descriptor = {});

    undefined destroy();

    readonly attribute GPUIntegerCoordinateOut width;
    readonly attribute GPUIntegerCoordinateOut height;
    readonly attribute GPUIntegerCoordinateOut depthOrArrayLayers;
    readonly attribute GPUIntegerCoordinateOut mipLevelCount;
    readonly attribute GPUSize32Out sampleCount;
    readonly attribute GPUTextureDimension dimension;
    readonly attribute GPUTextureFormat format;
    readonly attribute GPUFlagsConstant usage;
};
GPUTexture includes GPUObjectBase;

GPUTexture has the following attributes:

width, of type GPUIntegerCoordinateOut, readonly

The width of this GPUTexture.

height, of type GPUIntegerCoordinateOut, readonly

The height of this GPUTexture.

depthOrArrayLayers, of type GPUIntegerCoordinateOut, readonly

The depth or layer count of this GPUTexture.

mipLevelCount, of type GPUIntegerCoordinateOut, readonly

The number of mip levels of this GPUTexture.

sampleCount, of type GPUSize32Out, readonly

The number of sample count of this GPUTexture.

dimension, of type GPUTextureDimension, readonly

The dimension of the set of texel for each of this GPUTexture's subresources.

format, of type GPUTextureFormat, readonly

The format of this GPUTexture.

usage, of type GPUFlagsConstant, readonly

The allowed usages for this GPUTexture.

GPUTexture has the following internal slots:

[[size]], of type GPUExtent3D

The size of the texture (same as the width, height, and depthOrArrayLayers attributes).

[[viewFormats]], of type sequence<GPUTextureFormat>

The set of GPUTextureFormats that can be used GPUTextureViewDescriptor.format when creating views on this GPUTexture.

[[destroyed]], of type boolean, initially false

If the texture is destroyed, it can no longer be used in any operation, and its underlying memory can be freed.

compute render extent(baseSize, mipLevel)

Arguments:

Returns: GPUExtent3DDict

  1. Let extent be a new GPUExtent3DDict object.

  2. Set extent.width to max(1, baseSize.widthmipLevel).

  3. Set extent.height to max(1, baseSize.heightmipLevel).

  4. Set extent.depthOrArrayLayers to 1.

  5. Return extent.

The logical miplevel-specific texture extent of a texture is the size of the texture in texels at a specific miplevel. It is calculated by this procedure:

Logical miplevel-specific texture extent(descriptor, mipLevel)

Arguments:

Returns: GPUExtent3DDict

  1. Let extent be a new GPUExtent3DDict object.

  2. If descriptor.dimension is:

    "1d"
    "2d"
    "3d"
  3. Return extent.

The physical miplevel-specific texture extent of a texture is the size of the texture in texels at a specific miplevel that includes the possible extra padding to form complete texel blocks in the texture. It is calculated by this procedure:

Physical miplevel-specific texture extent(descriptor, mipLevel)

Arguments:

Returns: GPUExtent3DDict

  1. Let extent be a new GPUExtent3DDict object.

  2. Let logicalExtent be logical miplevel-specific texture extent(descriptor, mipLevel).

  3. If descriptor.dimension is:

    "1d"
    "2d"
    "3d"
  4. Return extent.

6.1.1. GPUTextureDescriptor

dictionary GPUTextureDescriptor
         : GPUObjectDescriptorBase {
    required GPUExtent3D size;
    GPUIntegerCoordinate mipLevelCount = 1;
    GPUSize32 sampleCount = 1;
    GPUTextureDimension dimension = "2d";
    required GPUTextureFormat format;
    required GPUTextureUsageFlags usage;
    sequence<GPUTextureFormat> viewFormats = [];
};

GPUTextureDescriptor has the following members:

size, of type GPUExtent3D

The width, height, and depth or layer count of the texture.

mipLevelCount, of type GPUIntegerCoordinate, defaulting to 1

The number of mip levels the texture will contain.

sampleCount, of type GPUSize32, defaulting to 1

The sample count of the texture. A sampleCount > 1 indicates a multisampled texture.

dimension, of type GPUTextureDimension, defaulting to "2d"

Whether the texture is one-dimensional, an array of two-dimensional layers, or three-dimensional.

format, of type GPUTextureFormat

The format of the texture.

usage, of type GPUTextureUsageFlags

The allowed usages for the texture.

viewFormats, of type sequence<GPUTextureFormat>, defaulting to []

Specifies what view format values will be allowed when calling createView() on this texture (in addition to the texture’s actual format).

Note: Adding a format to this list may have a significant performance impact, so it is best to avoid adding formats unnecessarily.

The actual performance impact is highly dependent on the target system; developers must test various systems to find out the impact on their particular application. For example, on some systems any texture with a format or viewFormats entry including "rgba8unorm-srgb" will perform less optimally than a "rgba8unorm" texture which does not. Similar caveats exist for other formats and pairs of formats on other systems.

Formats in this list must be texture view format compatible with the texture format.

Two GPUTextureFormats format and viewFormat are texture view format compatible if:
  • format equals viewFormat, or

  • format and viewFormat differ only in whether they are srgb formats (have the -srgb suffix).

enum GPUTextureDimension {
    "1d",
    "2d",
    "3d",
};
"1d"

Specifies a texture that has one dimension, width.

"2d"

Specifies a texture that has a width and height, and may have layers. Only "2d" textures may have mipmaps, be multisampled, use a compressed or depth/stencil format, and be used as a render attachment.

"3d"

Specifies a texture that has a width, height, and depth.

6.1.2. Texture Usages

typedef [EnforceRange] unsigned long GPUTextureUsageFlags;
[Exposed=(Window, DedicatedWorker), SecureContext]
namespace GPUTextureUsage {
    const GPUFlagsConstant COPY_SRC          = 0x01;
    const GPUFlagsConstant COPY_DST          = 0x02;
    const GPUFlagsConstant TEXTURE_BINDING   = 0x04;
    const GPUFlagsConstant STORAGE_BINDING   = 0x08;
    const GPUFlagsConstant RENDER_ATTACHMENT = 0x10;
};

The GPUTextureUsage flags determine how a GPUTexture may be used after its creation:

COPY_SRC

The texture can be used as the source of a copy operation. (Examples: as the source argument of a copyTextureToTexture() or copyTextureToBuffer() call.)

COPY_DST

The texture can be used as the destination of a copy or write operation. (Examples: as the destination argument of a copyTextureToTexture() or copyBufferToTexture() call, or as the target of a writeTexture() call.)

TEXTURE_BINDING

The texture can be bound for use as a sampled texture in a shader (Example: as a bind group entry for a GPUTextureBindingLayout.)

STORAGE_BINDING

The texture can be bound for use as a storage texture in a shader (Example: as a bind group entry for a GPUStorageTextureBindingLayout.)

RENDER_ATTACHMENT

The texture can be used as a color or depth/stencil attachment in a render pass. (Example: as a GPURenderPassColorAttachment.view or GPURenderPassDepthStencilAttachment.view.)

maximum mipLevel count(dimension, size)

Arguments:

  1. Calculate the max dimension value m:

  2. Return floor(log2(m)) + 1.

6.1.3. Texture Creation

createTexture(descriptor)

Creates a GPUTexture.

Called on: GPUDevice this.

Arguments:

Arguments for the GPUDevice.createTexture(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPUTextureDescriptor Description of the GPUTexture to create.

Returns: GPUTexture

Content timeline steps:

  1. ? validate GPUExtent3D shape(descriptor.size).

  2. ? Validate texture format required features of descriptor.format with this.[[device]].

  3. ? Validate texture format required features of each element of descriptor.viewFormats with this.[[device]].

  4. Let t be a new GPUTexture object.

  5. Set t.width to descriptor.size.width.

  6. Set t.height to descriptor.size.height.

  7. Set t.depthOrArrayLayers to descriptor.size.depthOrArrayLayers.

  8. Set t.mipLevelCount to descriptor.mipLevelCount.

  9. Set t.sampleCount to descriptor.sampleCount.

  10. Set t.dimension to descriptor.dimension.

  11. Set t.format to descriptor.format.

  12. Set t.usage to descriptor.usage.

  13. Issue the initialization steps on the Device timeline of this.

  14. Return t.

Device timeline initialization steps:
  1. If any of the following conditions are unsatisfied generate a validation error, make t invalid, and stop.

  2. Set t.[[size]] to descriptor.size.

  3. Set t.[[viewFormats]] to descriptor.viewFormats.

validating GPUTextureDescriptor(GPUDevice this, GPUTextureDescriptor descriptor):

Return true if all of the following requirements are met, and false otherwise:

Creating a 16x16, RGBA, 2D texture with one array layer and one mip level:
const texture = gpuDevice.createTexture({
    size: { width: 16, height: 16 },
    format: 'rgba8unorm',
    usage: GPUTextureUsage.TEXTURE_BINDING,
});

6.1.4. Texture Destruction

An application that no longer requires a GPUTexture can choose to lose access to it before garbage collection by calling destroy().

Note: This allows the user agent to reclaim the GPU memory associated with the GPUTexture once all previously submitted operations using it are complete.

destroy()

Destroys the GPUTexture.

Called on: GPUTexture this.

Returns: undefined

Content timeline steps:

  1. Set this.[[destroyed]] to true.

6.2. GPUTextureView

A GPUTextureView is a view onto some subset of the texture subresources defined by a particular GPUTexture.

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUTextureView {
};
GPUTextureView includes GPUObjectBase;

GPUTextureView has the following internal slots:

[[texture]]

The GPUTexture into which this is a view.

[[descriptor]]

The GPUTextureViewDescriptor describing this texture view.

All optional fields of GPUTextureViewDescriptor are defined.

[[renderExtent]]

For renderable views, this is the effective GPUExtent3DDict for rendering.

Note: this extent depends on the baseMipLevel.

The set of subresources of a texture view view, with [[descriptor]] desc, is the subset of the subresources of view.[[texture]] for which each subresource s satisfies the following:

Two GPUTextureView objects are texture-view-aliasing if and only if their sets of subresources intersect.

6.2.1. Texture View Creation

dictionary GPUTextureViewDescriptor
         : GPUObjectDescriptorBase {
    GPUTextureFormat format;
    GPUTextureViewDimension dimension;
    GPUTextureAspect aspect = "all";
    GPUIntegerCoordinate baseMipLevel = 0;
    GPUIntegerCoordinate mipLevelCount;
    GPUIntegerCoordinate baseArrayLayer = 0;
    GPUIntegerCoordinate arrayLayerCount;
};

GPUTextureViewDescriptor has the following members:

format, of type GPUTextureFormat

The format of the texture view. Must be either the format of the texture or one of the viewFormats specified during its creation.

dimension, of type GPUTextureViewDimension

The dimension to view the texture as.

aspect, of type GPUTextureAspect, defaulting to "all"

Which aspect(s) of the texture are accessible to the texture view.

baseMipLevel, of type GPUIntegerCoordinate, defaulting to 0

The first (most detailed) mipmap level accessible to the texture view.

mipLevelCount, of type GPUIntegerCoordinate

How many mipmap levels, starting with baseMipLevel, are accessible to the texture view.

baseArrayLayer, of type GPUIntegerCoordinate, defaulting to 0

The index of the first array layer accessible to the texture view.

arrayLayerCount, of type GPUIntegerCoordinate

How many array layers, starting with baseArrayLayer, are accessible to the texture view.

enum GPUTextureViewDimension {
    "1d",
    "2d",
    "2d-array",
    "cube",
    "cube-array",
    "3d",
};
"1d"

The texture is viewed as a 1-dimensional image.

Corresponding WGSL types:

  • texture_1d

  • texture_storage_1d

"2d"

The texture is viewed as a single 2-dimensional image.

Corresponding WGSL types:

  • texture_2d

  • texture_storage_2d

  • texture_multisampled_2d

  • texture_depth_2d

  • texture_depth_multisampled_2d

"2d-array"

The texture view is viewed as an array of 2-dimensional images.

Corresponding WGSL types:

  • texture_2d_array

  • texture_storage_2d_array

  • texture_depth_2d_array

"cube"

The texture is viewed as a cubemap. The view has 6 array layers, corresponding to the [+X, -X, +Y, -Y, +Z, -Z] faces of the cube. Sampling is done seamlessly across the faces of the cubemap.

Corresponding WGSL types:

  • texture_cube

  • texture_depth_cube

"cube-array"

The texture is viewed as a packed array of n cubemaps, each with 6 array layers corresponding to the [+X, -X, +Y, -Y, +Z, -Z] faces of the cube. Sampling is done seamlessly across the faces of the cubemaps.

Corresponding WGSL types:

  • texture_cube_array

  • texture_depth_cube_array

"3d"

The texture is viewed as a 3-dimensional image.

Corresponding WGSL types:

  • texture_3d

  • texture_storage_3d

Each GPUTextureAspect value corresponds to a set of aspects. The set of aspects are defined for each value below.

enum GPUTextureAspect {
    "all",
    "stencil-only",
    "depth-only",
};
"all"

All available aspects of the texture format will be accessible to the texture view. For color formats the color aspect will be accessible. For combined depth-stencil formats both the depth and stencil aspects will be accessible. Depth-or-stencil formats with a single aspect will only make that aspect accessible.

The set of aspects is [color, depth, stencil].

"stencil-only"

Only the stencil aspect of a depth-or-stencil format format will be accessible to the texture view.

The set of aspects is [stencil].

"depth-only"

Only the depth aspect of a depth-or-stencil format format will be accessible to the texture view.

The set of aspects is [depth].

createView(descriptor)

Creates a GPUTextureView.

Note: By default createView() will create a view with a dimension that can represent the entire texture. For example, calling createView() without specifying a dimension on a "2d" texture with more than one layer will create a "2d-array" GPUTextureView, even if an arrayLayerCount of 1 is specified.

For textures created from sources where the layer count is unknown at the time of development it is recommended that calls to createView() are provided with an explicit dimension to ensure shader compatibility.

Called on: GPUTexture this.

Arguments:

Arguments for the GPUTexture.createView(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPUTextureViewDescriptor Description of the GPUTextureView to create.

Returns: view, of type GPUTextureView.

Content timeline steps:

  1. ? Validate texture format required features of descriptor.format with this.[[device]].

  2. Let view be a new GPUTextureView object.

  3. Issue the initialization steps on the Device timeline of this.

  4. Return view.

Device timeline initialization steps:
  1. Set descriptor to the result of resolving GPUTextureViewDescriptor defaults for this with descriptor.

  2. If any of the following conditions are unsatisfied generate a validation error, make view invalid, and stop.

  3. Let view be a new GPUTextureView object.

  4. Set view.[[texture]] to this.

  5. Set view.[[descriptor]] to descriptor.

  6. If this.usage contains RENDER_ATTACHMENT:

    1. Let renderExtent be compute render extent(this.[[size]], descriptor.baseMipLevel).

    2. Set view.[[renderExtent]] to renderExtent.

When resolving GPUTextureViewDescriptor defaults for GPUTextureView texture with a GPUTextureViewDescriptor descriptor run the following steps:
  1. Let resolved be a copy of descriptor.

  2. If resolved.format is not provided:

    1. Let format be the result of resolving GPUTextureAspect( format, descriptor.aspect).

    2. If format is null:

      Otherwise:

      • Set resolved.format to format.

  3. If resolved.mipLevelCount is not provided: set resolved.mipLevelCount to texture.mipLevelCountresolved.baseMipLevel.

  4. If resolved.dimension is not provided and texture.dimension is:

    "1d"

    Set resolved.dimension to "1d".

    "2d"

    If the array layer count of texture is 1:

    Otherwise:

    "3d"

    Set resolved.dimension to "3d".

  5. If resolved.arrayLayerCount is not provided and resolved.dimension is:

    "1d", "2d", or "3d"

    Set resolved.arrayLayerCount to 1.

    "cube"

    Set resolved.arrayLayerCount to 6.

    "2d-array" or "cube-array"

    Set resolved.arrayLayerCount to the array layer count of textureresolved.baseArrayLayer.

  6. Return resolved.

To determine the array layer count of GPUTexture texture, run the following steps:
  1. If texture.dimension is:

    "1d" or "3d"

    Return 1.

    "2d"

    Return texture.depthOrArrayLayers.

6.3. Texture Formats

The name of the format specifies the order of components, bits per component, and data type for the component.

If the format has the -srgb suffix, then sRGB conversions from gamma to linear and vice versa are applied during the reading and writing of color values in the shader. Compressed texture formats are provided by features. Their naming should follow the convention here, with the texture name as a prefix. e.g. etc2-rgba8unorm.

The texel block is a single addressable element of the textures in pixel-based GPUTextureFormats, and a single compressed block of the textures in block-based compressed GPUTextureFormats.

The texel block width and texel block height specifies the dimension of one texel block.

The texel block copy footprint of an aspect of a GPUTextureFormat is the number of bytes one texel block occupies during an image copy, if applicable.

Note: The texel block memory cost of a GPUTextureFormat is the number of bytes needed to store one texel block. It is not fully defined for all formats. This value is informative and non-normative.

enum GPUTextureFormat {
    // 8-bit formats
    "r8unorm",
    "r8snorm",
    "r8uint",
    "r8sint",

    // 16-bit formats
    "r16uint",
    "r16sint",
    "r16float",
    "rg8unorm",
    "rg8snorm",
    "rg8uint",
    "rg8sint",

    // 32-bit formats
    "r32uint",
    "r32sint",
    "r32float",
    "rg16uint",
    "rg16sint",
    "rg16float",
    "rgba8unorm",
    "rgba8unorm-srgb",
    "rgba8snorm",
    "rgba8uint",
    "rgba8sint",
    "bgra8unorm",
    "bgra8unorm-srgb",
    // Packed 32-bit formats
    "rgb9e5ufloat",
    "rgb10a2unorm",
    "rg11b10ufloat",

    // 64-bit formats
    "rg32uint",
    "rg32sint",
    "rg32float",
    "rgba16uint",
    "rgba16sint",
    "rgba16float",

    // 128-bit formats
    "rgba32uint",
    "rgba32sint",
    "rgba32float",

    // Depth/stencil formats
    "stencil8",
    "depth16unorm",
    "depth24plus",
    "depth24plus-stencil8",
    "depth32float",

    // "depth32float-stencil8" feature
    "depth32float-stencil8",

    // BC compressed formats usable if "texture-compression-bc" is both
    // supported by the device/user agent and enabled in requestDevice.
    "bc1-rgba-unorm",
    "bc1-rgba-unorm-srgb",
    "bc2-rgba-unorm",
    "bc2-rgba-unorm-srgb",
    "bc3-rgba-unorm",
    "bc3-rgba-unorm-srgb",
    "bc4-r-unorm",
    "bc4-r-snorm",
    "bc5-rg-unorm",
    "bc5-rg-snorm",
    "bc6h-rgb-ufloat",
    "bc6h-rgb-float",
    "bc7-rgba-unorm",
    "bc7-rgba-unorm-srgb",

    // ETC2 compressed formats usable if "texture-compression-etc2" is both
    // supported by the device/user agent and enabled in requestDevice.
    "etc2-rgb8unorm",
    "etc2-rgb8unorm-srgb",
    "etc2-rgb8a1unorm",
    "etc2-rgb8a1unorm-srgb",
    "etc2-rgba8unorm",
    "etc2-rgba8unorm-srgb",
    "eac-r11unorm",
    "eac-r11snorm",
    "eac-rg11unorm",
    "eac-rg11snorm",

    // ASTC compressed formats usable if "texture-compression-astc" is both
    // supported by the device/user agent and enabled in requestDevice.
    "astc-4x4-unorm",
    "astc-4x4-unorm-srgb",
    "astc-5x4-unorm",
    "astc-5x4-unorm-srgb",
    "astc-5x5-unorm",
    "astc-5x5-unorm-srgb",
    "astc-6x5-unorm",
    "astc-6x5-unorm-srgb",
    "astc-6x6-unorm",
    "astc-6x6-unorm-srgb",
    "astc-8x5-unorm",
    "astc-8x5-unorm-srgb",
    "astc-8x6-unorm",
    "astc-8x6-unorm-srgb",
    "astc-8x8-unorm",
    "astc-8x8-unorm-srgb",
    "astc-10x5-unorm",
    "astc-10x5-unorm-srgb",
    "astc-10x6-unorm",
    "astc-10x6-unorm-srgb",
    "astc-10x8-unorm",
    "astc-10x8-unorm-srgb",
    "astc-10x10-unorm",
    "astc-10x10-unorm-srgb",
    "astc-12x10-unorm",
    "astc-12x10-unorm-srgb",
    "astc-12x12-unorm",
    "astc-12x12-unorm-srgb",
};

The depth component of the "depth24plus" and "depth24plus-stencil8" formats may be implemented as either a 24-bit depth value or a "depth32float" value.

The stencil8 format may be implemented as either a real "stencil8", or "depth24stencil8", where the depth aspect is hidden and inaccessible.

Note: While the precision of depth32float channels is strictly higher than the precision of 24-bit depth channels for all values in the representable range (0.0 to 1.0), note that the set of representable values is not an exact superset.

A format is renderable if it is either a color renderable format, or a depth-or-stencil format. If a format is listed in § 26.1.1 Plain color formats with RENDER_ATTACHMENT capability, it is a color renderable format. Any other format is not a color renderable format. All depth-or-stencil formats are renderable.

A renderable format is also blendable if it can be used with render pipeline blending. See § 26.1 Texture Format Capabilities.

A format is filterable if it supports the GPUTextureSampleType "float" (not just "unfilterable-float"); that is, it can be used with "filtering" GPUSamplers. See § 26.1 Texture Format Capabilities.

resolving GPUTextureAspect(format, aspect)

Arguments:

Returns: GPUTextureFormat or null

  1. If aspect is:

    "all"

    Return format.

    "depth-only"
    "stencil-only"

    If format is a depth-stencil-format: Return the aspect-specific format of format according to § 26.1.2 Depth-stencil formats or null if the aspect is not present in format.

  2. Return null.

Use of some texture formats require a feature to be enabled on the GPUDevice. Because new formats can be added to the specification, those enum values may not be known by the implementation. In order to normalize behavior across implementations, attempting to use a format that requires a feature will throw an exception if the associated feature is not enabled on the device. This makes the behavior the same as when the format is unknown to the implementation.

See § 26.1 Texture Format Capabilities for information about which GPUTextureFormats require features.

Validate texture format required features of a GPUTextureFormat format with logical device device by running the following steps:
  1. If format requires a feature and device.[[features]] does not contain the feature:

    1. Throw a TypeError.

6.4. GPUExternalTexture

A GPUExternalTexture is a sampleable 2D texture wrapping an external video object. The contents of a GPUExternalTexture object are a snapshot and may not change, either from inside WebGPU (it is only sampleable) or from outside WebGPU (e.g. due to video frame advancement).

They are bound into bind group layouts using the externalTexture bind group layout entry member. External textures use several binding slots: see Exceeds the binding slot limits.

Note: External textures can be implemented without creating a copy of the imported source, but this depends implementation-defined factors. Ownership of the underlying representation may either be exclusive or shared with other owners (such as a video decoder), but this is not visible to the application.

The underlying representation of an external texture is unobservable (except for sampling behavior) but typically may include

The configuration used may not be stable across time, systems, user agents, media sources, or frames within a single video source. In order to account for many possible representations, the binding conservatively uses the following, for each external texture:

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUExternalTexture {
};
GPUExternalTexture includes GPUObjectBase;

GPUExternalTexture has the following internal slots:

[[expired]], of type boolean

Indicates whether the object has expired (can no longer be used). Initially set to false.

Note: Unlike similar \[[destroyed]] slots, this can change from true back to false.

[[descriptor]], of type GPUExternalTextureDescriptor

The descriptor with which the texture was created.

6.4.1. Importing External Textures

An external texture is created from an external video object using importExternalTexture().

An external texture created from an HTMLVideoElement expires (is destroyed) automatically in a task after it is imported, instead of manually or upon garbage collection like other resources. When an external texture expires, its [[expired]] slot changes to true.

An external texture created from a VideoFrame expires (is destroyed) when, and only when, the source VideoFrame is closed, either explicitly by close(), or by other means.

Note: As noted in decode(), authors should call close() on output VideoFrames to avoid decoder stalls. If an imported VideoFrame is dropped without being closed, the imported GPUExternalTexture object will keep it alive until it is also dropped. The VideoFrame cannot be garbage collected until both objects are dropped. Garbage collection is unpredictable, so this may still stall the video decoder.

Once the GPUExternalTexture expires, importExternalTexture() must be called again. However, the user agent may un-expire and return the same GPUExternalTexture again, instead of creating a new one. This will commonly happen unless the execution of the application is scheduled to match the video’s frame rate (e.g. using requestVideoFrameCallback()). If the same object is returned again, it will compare equal, and GPUBindGroups, GPURenderBundles, etc. referencing the previous object can still be used.

dictionary GPUExternalTextureDescriptor
         : GPUObjectDescriptorBase {
    required (HTMLVideoElement or VideoFrame) source;
    PredefinedColorSpace colorSpace = "srgb";
};
importExternalTexture(descriptor)

Creates a GPUExternalTexture wrapping the provided image source.

Called on: GPUDevice this.

Arguments:

Arguments for the GPUDevice.importExternalTexture(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPUExternalTextureDescriptor Provides the external image source object (and any creation options).

Returns: GPUExternalTexture

Content timeline steps:

  1. Let source be descriptor.source.

  2. If the current image contents of source are the same as the most recent importExternalTexture() call with the same descriptor (ignoring label), and the user agent chooses to reuse it:

    1. Let previousResult be the GPUExternalTexture returned previously.

    2. Set previousResult.[[expired]] to false, renewing ownership of the underlying resource.

    3. Let result be previousResult.

    Note: This allows the application to detect duplicate imports and avoid re-creating dependent objects (such as GPUBindGroups). Implementations still need to be able to handle a single frame being wrapped by multiple GPUExternalTexture, since import metadata like colorSpace can change even for the same frame.

    Otherwise:

    1. If source is not origin-clean, throw a SecurityError and stop.

    2. Let usability be ? check the usability of the image argument(source).

    3. If usability is not good:

      1. Generate a validation error.

      2. Return an invalid GPUExternalTexture.

    4. Let data be the result of converting the current image contents of source into the color space descriptor.colorSpace with unpremultiplied alpha.

      This may result in values outside of the range [0, 1]. If clamping is desired, it may be performed after sampling.

      Note: This is described like a copy, but may be implemented as a reference to read-only underlying data plus appropriate metadata to perform conversion later.

    5. Let result be a new GPUExternalTexture object wrapping data.

  3. If source is an HTMLVideoElement, queue an automatic expiry task with device this and the following steps:

    1. Set result.[[expired]] to true, releasing ownership of the underlying resource.

    Note: An HTMLVideoElement should be imported in the same task that samples the texture (which should generally be scheduled using requestVideoFrameCallback or requestAnimationFrame() depending on the application). Otherwise, a texture could get destroyed by these steps before the application is finished using it.

  4. If source is a VideoFrame, then when source is closed, run the following steps:

    1. Set result.[[expired]] to true.

  5. Set result.label to descriptor.label.

  6. Return result.

Rendering using an video element external texture at the page animation frame rate:
const videoElement = document.createElement('video');
// ... set up videoElement, wait for it to be ready...

function frame() {
    requestAnimationFrame(frame);

    // Always re-import the video on every animation frame, because the
    // import is likely to have expired.
    // The browser may cache and reuse a past frame, and if it does it
    // may return the same GPUExternalTexture object again.
    // In this case, old bind groups are still valid.
    const externalTexture = gpuDevice.importExternalTexture({
        source: videoElement
    });

    // ... render using externalTexture...
}
requestAnimationFrame(frame);
Rendering using an video element external texture at the video’s frame rate, if requestVideoFrameCallback is available:
const videoElement = document.createElement('video');
// ... set up videoElement...

function frame() {
    videoElement.requestVideoFrameCallback(frame);

    // Always re-import, because we know the video frame has advanced
    const externalTexture = gpuDevice.importExternalTexture({
        source: videoElement
    });

    // ... render using externalTexture...
}
videoElement.requestVideoFrameCallback(frame);

6.4.2. Sampling External Textures

External textures are represented in WGSL with texture_external and may be read using textureLoad and textureSampleBaseClampToEdge.

The sampler provided to textureSampleBaseClampToEdge is used to sample the underlying textures. The result is in the color space set by colorSpace. It is implementation-dependent whether, for any given external texture, the sampler (and filtering) is applied before or after conversion from underlying values into the specified color space.

Note: If the internal representation is an RGBA plane, sampling behaves as on a regular 2D texture. If there are several underlying planes (e.g. Y+UV), the sampler is used to sample each underlying texture separately, prior to conversion from YUV to the specified color space.

7. Samplers

7.1. GPUSampler

A GPUSampler encodes transformations and filtering information that can be used in a shader to interpret texture resource data.

GPUSamplers are created via createSampler().

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUSampler {
};
GPUSampler includes GPUObjectBase;

GPUSampler has the following internal slots:

[[descriptor]], of type GPUSamplerDescriptor, readonly

The GPUSamplerDescriptor with which the GPUSampler was created.

[[isComparison]], of type boolean

Whether the GPUSampler is used as a comparison sampler.

[[isFiltering]], of type boolean

Whether the GPUSampler weights multiple samples of a texture.

7.1.1. GPUSamplerDescriptor

A GPUSamplerDescriptor specifies the options to use to create a GPUSampler.

dictionary GPUSamplerDescriptor
         : GPUObjectDescriptorBase {
    GPUAddressMode addressModeU = "clamp-to-edge";
    GPUAddressMode addressModeV = "clamp-to-edge";
    GPUAddressMode addressModeW = "clamp-to-edge";
    GPUFilterMode magFilter = "nearest";
    GPUFilterMode minFilter = "nearest";
    GPUMipmapFilterMode mipmapFilter = "nearest";
    float lodMinClamp = 0;
    float lodMaxClamp = 32;
    GPUCompareFunction compare;
    [Clamp] unsigned short maxAnisotropy = 1;
};
addressModeU, of type GPUAddressMode, defaulting to "clamp-to-edge"
addressModeV, of type GPUAddressMode, defaulting to "clamp-to-edge"
addressModeW, of type GPUAddressMode, defaulting to "clamp-to-edge"

Specifies the address modes for the texture width, height, and depth coordinates, respectively.

magFilter, of type GPUFilterMode, defaulting to "nearest"

Specifies the sampling behavior when the sample footprint is smaller than or equal to one texel.

minFilter, of type GPUFilterMode, defaulting to "nearest"

Specifies the sampling behavior when the sample footprint is larger than one texel.

mipmapFilter, of type GPUMipmapFilterMode, defaulting to "nearest"

Specifies behavior for sampling between mipmap levels.

lodMinClamp, of type float, defaulting to 0
lodMaxClamp, of type float, defaulting to 32

Specifies the minimum and maximum levels of detail, respectively, used internally when sampling a texture.

compare, of type GPUCompareFunction

When provided the sampler will be a comparison sampler with the specified GPUCompareFunction.

Note: Comparison samplers may use filtering, but the sampling results will be implementation-dependent and may differ from the normal filtering rules.

maxAnisotropy, of type unsigned short, defaulting to 1

Specifies the maximum anisotropy value clamp used by the sampler.

Note: Most implementations support maxAnisotropy values in range between 1 and 16, inclusive. The used value of maxAnisotropy will be clamped to the maximum value that the platform supports.

explain how LOD is calculated and if there are differences here between platforms.

explain what anisotropic sampling is

GPUAddressMode describes the behavior of the sampler if the sample footprint extends beyond the bounds of the sampled texture.

Describe a "sample footprint" in greater detail.

enum GPUAddressMode {
    "clamp-to-edge",
    "repeat",
    "mirror-repeat",
};
"clamp-to-edge"

Texture coordinates are clamped between 0.0 and 1.0, inclusive.

"repeat"

Texture coordinates wrap to the other side of the texture.

"mirror-repeat"

Texture coordinates wrap to the other side of the texture, but the texture is flipped when the integer part of the coordinate is odd.

GPUFilterMode and GPUMipmapFilterMode describe the behavior of the sampler if the sample footprint does not exactly match one texel.

enum GPUFilterMode {
    "nearest",
    "linear",
};

enum GPUMipmapFilterMode {
    "nearest",
    "linear",
};
"nearest"

Return the value of the texel nearest to the texture coordinates.

"linear"

Select two texels in each dimension and return a linear interpolation between their values.

GPUCompareFunction specifies the behavior of a comparison sampler. If a comparison sampler is used in a shader, an input value is compared to the sampled texture value, and the result of this comparison test (0.0f for pass, or 1.0f for fail) is used in the filtering operation.

describe how filtering interacts with comparison sampling.

enum GPUCompareFunction {
    "never",
    "less",
    "equal",
    "less-equal",
    "greater",
    "not-equal",
    "greater-equal",
    "always",
};
"never"

Comparison tests never pass.

"less"

A provided value passes the comparison test if it is less than the sampled value.

"equal"

A provided value passes the comparison test if it is equal to the sampled value.

"less-equal"

A provided value passes the comparison test if it is less than or equal to the sampled value.

"greater"

A provided value passes the comparison test if it is greater than the sampled value.

"not-equal"

A provided value passes the comparison test if it is not equal to the sampled value.

"greater-equal"

A provided value passes the comparison test if it is greater than or equal to the sampled value.

"always"

Comparison tests always pass.

7.1.2. Sampler Creation

createSampler(descriptor)

Creates a GPUSampler.

Called on: GPUDevice this.

Arguments:

Arguments for the GPUDevice.createSampler(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPUSamplerDescriptor Description of the GPUSampler to create.

Returns: GPUSampler

Content timeline steps:

  1. Let s be a new GPUSampler object.

  2. Issue the initialization steps on the Device timeline of this.

  3. Return s.

Device timeline initialization steps:
  1. If any of the following conditions are unsatisfied generate a validation error, make s invalid, and stop.

  2. Set s.[[descriptor]] to descriptor.

  3. Set s.[[isComparison]] to false if the compare attribute of s.[[descriptor]] is null or undefined. Otherwise, set it to true.

  4. Set s.[[isFiltering]] to false if none of minFilter, magFilter, or mipmapFilter has the value of "linear". Otherwise, set it to true.

Creating a GPUSampler that does trilinear filtering and repeats texture coordinates:
const sampler = gpuDevice.createSampler({
    addressModeU: 'repeat',
    addressModeV: 'repeat',
    magFilter: 'linear',
    minFilter: 'linear',
    mipmapFilter: 'linear',
});

8. Resource Binding

8.1. GPUBindGroupLayout

A GPUBindGroupLayout defines the interface between a set of resources bound in a GPUBindGroup and their accessibility in shader stages.

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUBindGroupLayout {
};
GPUBindGroupLayout includes GPUObjectBase;

GPUBindGroupLayout has the following internal slots:

[[descriptor]], of type GPUBindGroupLayoutDescriptor

8.1.1. Bind Group Layout Creation

A GPUBindGroupLayout is created via GPUDevice.createBindGroupLayout().

dictionary GPUBindGroupLayoutDescriptor
         : GPUObjectDescriptorBase {
    required sequence<GPUBindGroupLayoutEntry> entries;
};

A GPUBindGroupLayoutEntry describes a single shader resource binding to be included in a GPUBindGroupLayout.

dictionary GPUBindGroupLayoutEntry {
    required GPUIndex32 binding;
    required GPUShaderStageFlags visibility;

    GPUBufferBindingLayout buffer;
    GPUSamplerBindingLayout sampler;
    GPUTextureBindingLayout texture;
    GPUStorageTextureBindingLayout storageTexture;
    GPUExternalTextureBindingLayout externalTexture;
};

GPUBindGroupLayoutEntry dictionaries have the following members:

binding, of type GPUIndex32

A unique identifier for a resource binding within the GPUBindGroupLayout, corresponding to a GPUBindGroupEntry.binding and a @binding attribute in the GPUShaderModule.

visibility, of type GPUShaderStageFlags

A bitset of the members of GPUShaderStage. Each set bit indicates that a GPUBindGroupLayoutEntry's resource will be accessible from the associated shader stage.

buffer, of type GPUBufferBindingLayout

When provided, indicates the binding resource type for this GPUBindGroupLayoutEntry is GPUBufferBinding.

sampler, of type GPUSamplerBindingLayout

When provided, indicates the binding resource type for this GPUBindGroupLayoutEntry is GPUSampler.

texture, of type GPUTextureBindingLayout

When provided, indicates the binding resource type for this GPUBindGroupLayoutEntry is GPUTextureView.

storageTexture, of type GPUStorageTextureBindingLayout

When provided, indicates the binding resource type for this GPUBindGroupLayoutEntry is GPUTextureView.

externalTexture, of type GPUExternalTextureBindingLayout

When provided, indicates the binding resource type for this GPUBindGroupLayoutEntry is GPUExternalTexture.

typedef [EnforceRange] unsigned long GPUShaderStageFlags;
[Exposed=(Window, DedicatedWorker), SecureContext]
namespace GPUShaderStage {
    const GPUFlagsConstant VERTEX   = 0x1;
    const GPUFlagsConstant FRAGMENT = 0x2;
    const GPUFlagsConstant COMPUTE  = 0x4;
};

GPUShaderStage contains the following flags, which describe which shader stages a corresponding GPUBindGroupEntry for this GPUBindGroupLayoutEntry will be visible to:

VERTEX

The bind group entry will be accessible to vertex shaders.

FRAGMENT

The bind group entry will be accessible to fragment shaders.

COMPUTE

The bind group entry will be accessible to compute shaders.

The binding member of a GPUBindGroupLayoutEntry is determined by which member of the GPUBindGroupLayoutEntry is defined: buffer, sampler, texture, storageTexture, or externalTexture. Only one may be defined for any given GPUBindGroupLayoutEntry. Each member has an associated GPUBindingResource type and each binding type has an associated internal usage, given by this table:

Binding member Resource type Binding type
Binding usage
buffer GPUBufferBinding "uniform" constant
"storage" storage
"read-only-storage" storage-read
sampler GPUSampler "filtering" constant
"non-filtering"
"comparison"
texture GPUTextureView "float" constant
"unfilterable-float"
"depth"
"sint"
"uint"
storageTexture GPUTextureView "write-only" storage
externalTexture GPUExternalTexture constant
The list of GPUBindGroupLayoutEntry values entries exceeds the binding slot limits of supported limits limits if the number of slots used toward a limit exceeds the supported value in limits. Each entry may use multiple slots toward multiple limits.
  1. For each entry in entries, if:

    entry.buffer?.type is "uniform" and entry.buffer?.hasDynamicOffset is true

    Consider 1 maxDynamicUniformBuffersPerPipelineLayout slot to be used.

    entry.buffer?.type is "storage" and entry.buffer?.hasDynamicOffset is true

    Consider 1 maxDynamicStorageBuffersPerPipelineLayout slot to be used.

  2. For each shader stage stage in « VERTEX, FRAGMENT, COMPUTE »:

    1. For each entry in entries for which entry.visibility contains stage, if:

      entry.buffer?.type is "uniform"

      Consider 1 maxUniformBuffersPerShaderStage slot to be used.

      entry.buffer?.type is "storage" or "read-only-storage"

      Consider 1 maxStorageBuffersPerShaderStage slot to be used.

      entry.sampler is provided

      Consider 1 maxSamplersPerShaderStage slot to be used.

      entry.texture is provided

      Consider 1 maxSampledTexturesPerShaderStage slot to be used.

      entry.storageTexture is provided

      Consider 1 maxStorageTexturesPerShaderStage slot to be used.

      entry.externalTexture is provided

      Consider 4 maxSampledTexturesPerShaderStage slot, 1 maxSamplersPerShaderStage slot, and 1 maxUniformBuffersPerShaderStage slot to be used.

enum GPUBufferBindingType {
    "uniform",
    "storage",
    "read-only-storage",
};

dictionary GPUBufferBindingLayout {
    GPUBufferBindingType type = "uniform";
    boolean hasDynamicOffset = false;
    GPUSize64 minBindingSize = 0;
};

GPUBufferBindingLayout dictionaries have the following members:

type, of type GPUBufferBindingType, defaulting to "uniform"

Indicates the type required for buffers bound to this bindings.

hasDynamicOffset, of type boolean, defaulting to false

Indicates whether this binding requires a dynamic offset.

minBindingSize, of type GPUSize64, defaulting to 0

Indicates the minimum size of a buffer binding used with this bind point.

Bindings are always validated against this size in createBindGroup().

If this is not 0, pipeline creation additionally validates that this value ≥ the minimum buffer binding size of the variable.

If this is 0, it is ignored by pipeline creation, and instead draw/dispatch commands validate that each binding in the GPUBindGroup satisfies the minimum buffer binding size of the variable.

Note: Similar execution-time validation is theoretically possible for other binding-related fields specified for early validation, like sampleType and format, which currently can only be validated in pipeline creation. However, such execution-time validation could be costly or unnecessarily complex, so it is available only for minBindingSize which is expected to have the most ergonomic impact.

enum GPUSamplerBindingType {
    "filtering",
    "non-filtering",
    "comparison",
};

dictionary GPUSamplerBindingLayout {
    GPUSamplerBindingType type = "filtering";
};

GPUSamplerBindingLayout dictionaries have the following members:

type, of type GPUSamplerBindingType, defaulting to "filtering"

Indicates the required type of a sampler bound to this bindings.

enum GPUTextureSampleType {
    "float",
    "unfilterable-float",
    "depth",
    "sint",
    "uint",
};

dictionary GPUTextureBindingLayout {
    GPUTextureSampleType sampleType = "float";
    GPUTextureViewDimension viewDimension = "2d";
    boolean multisampled = false;
};

GPUTextureBindingLayout dictionaries have the following members:

sampleType, of type GPUTextureSampleType, defaulting to "float"

Indicates the type required for texture views bound to this binding.

viewDimension, of type GPUTextureViewDimension, defaulting to "2d"

Indicates the required dimension for texture views bound to this binding.

multisampled, of type boolean, defaulting to false

Indicates whether or not texture views bound to this binding must be multisampled.

enum GPUStorageTextureAccess {
    "write-only",
};

dictionary GPUStorageTextureBindingLayout {
    GPUStorageTextureAccess access = "write-only";
    required GPUTextureFormat format;
    GPUTextureViewDimension viewDimension = "2d";
};

GPUStorageTextureBindingLayout dictionaries have the following members:

access, of type GPUStorageTextureAccess, defaulting to "write-only"

The access mode for this binding, indicating readability and writability.

Note: There is currently only one access mode, "write-only", but this will expand in the future.

format, of type GPUTextureFormat

The required format of texture views bound to this binding.

viewDimension, of type GPUTextureViewDimension, defaulting to "2d"

Indicates the required dimension for texture views bound to this binding.

dictionary GPUExternalTextureBindingLayout {
};

A GPUBindGroupLayout object has the following internal slots:

[[entryMap]], of type ordered map<GPUSize32, GPUBindGroupLayoutEntry>

The map of binding indices pointing to the GPUBindGroupLayoutEntrys, which this GPUBindGroupLayout describes.

[[dynamicOffsetCount]], of type GPUSize32

The number of buffer bindings with dynamic offsets in this GPUBindGroupLayout.

[[exclusivePipeline]], of type GPUPipelineBase?, initially null

The pipeline that created this GPUBindGroupLayout, if it was created as part of a default pipeline layout. If not null, GPUBindGroups created with this GPUBindGroupLayout can only be used with the specified GPUPipelineBase.

createBindGroupLayout(descriptor)

Creates a GPUBindGroupLayout.

Called on: GPUDevice this.

Arguments:

Arguments for the GPUDevice.createBindGroupLayout(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPUBindGroupLayoutDescriptor Description of the GPUBindGroupLayout to create.

Returns: GPUBindGroupLayout

Content timeline steps:

  1. For each GPUBindGroupLayoutEntry entry in descriptor.entries:

    1. If entry.storageTexture is provided:

      1. ? Validate texture format required features for entry.storageTexture.format with this.[[device]].

  2. Let layout be a new GPUBindGroupLayout object.

  3. Issue the initialization steps on the Device timeline of this.

  4. Return layout.

Device timeline initialization steps:
  1. If any of the following conditions are unsatisfied generate a validation error, make layout invalid, and stop.

  2. Set layout.[[descriptor]] to descriptor.

  3. Set layout.[[dynamicOffsetCount]] to the number of entries in descriptor where buffer is provided and buffer.hasDynamicOffset is true.

  4. For each GPUBindGroupLayoutEntry entry in descriptor.entries:

    1. Insert entry into layout.[[entryMap]] with the key of entry.binding.

8.1.2. Compatibility

Two GPUBindGroupLayout objects a and b are considered group-equivalent if and only if all of the following conditions are satisfied:

If bind groups layouts are group-equivalent they can be interchangeably used in all contents.

8.2. GPUBindGroup

A GPUBindGroup defines a set of resources to be bound together in a group and how the resources are used in shader stages.

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUBindGroup {
};
GPUBindGroup includes GPUObjectBase;

A GPUBindGroup object has the following internal slots:

[[layout]], of type GPUBindGroupLayout, readonly

The GPUBindGroupLayout associated with this GPUBindGroup.

[[entries]], of type sequence<GPUBindGroupEntry>, readonly

The set of GPUBindGroupEntrys this GPUBindGroup describes.

[[usedResources]], of type ordered map<subresource, list<internal usage>>, readonly

The set of buffer and texture subresources used by this bind group, associated with lists of the internal usage flags.

The bound buffer ranges of a GPUBindGroup bindGroup, given list<GPUBufferDynamicOffset> dynamicOffsets, are computed as follows:
  1. Let result be a new set<(GPUBindGroupLayoutEntry, GPUBufferBinding)>.

  2. Let dynamicOffsetIndex be 0.

  3. For each GPUBindGroupEntry bindGroupEntry in bindGroup.[[entries]], sorted by bindGroupEntry.binding:

    1. Let bindGroupLayoutEntry be bindGroup.[[layout]].[[entryMap]][bindGroupEntry.binding].

    2. If bindGroupLayoutEntry.buffer is not provided, continue.

    3. Let bound be a copy of bindGroupEntry.resource.

    4. Assert bound is a GPUBufferBinding.

    5. If bindGroupLayoutEntry.buffer.hasDynamicOffset:

      1. Increment bound.offset by dynamicOffsets[dynamicOffsetIndex].

      2. Increment dynamicOffsetIndex by 1.

    6. Append (bindGroupLayoutEntry, bound) to result.

  4. Return result.

8.2.1. Bind Group Creation

A GPUBindGroup is created via GPUDevice.createBindGroup().

dictionary GPUBindGroupDescriptor
         : GPUObjectDescriptorBase {
    required GPUBindGroupLayout layout;
    required sequence<GPUBindGroupEntry> entries;
};

GPUBindGroupDescriptor dictionaries have the following members:

layout, of type GPUBindGroupLayout

The GPUBindGroupLayout the entries of this bind group will conform to.

entries, of type sequence<GPUBindGroupEntry>

A list of entries describing the resources to expose to the shader for each binding described by the layout.

typedef (GPUSampler or GPUTextureView or GPUBufferBinding or GPUExternalTexture) GPUBindingResource;

dictionary GPUBindGroupEntry {
    required GPUIndex32 binding;
    required GPUBindingResource resource;
};

A GPUBindGroupEntry describes a single resource to be bound in a GPUBindGroup, and has the following members:

binding, of type GPUIndex32

A unique identifier for a resource binding within the GPUBindGroup, corresponding to a GPUBindGroupLayoutEntry.binding and a @binding attribute in the GPUShaderModule.

resource, of type GPUBindingResource

The resource to bind, which may be a GPUSampler, GPUTextureView, GPUExternalTexture, or GPUBufferBinding.

dictionary GPUBufferBinding {
    required GPUBuffer buffer;
    GPUSize64 offset = 0;
    GPUSize64 size;
};

A GPUBufferBinding describes a buffer and optional range to bind as a resource, and has the following members:

buffer, of type GPUBuffer

The GPUBuffer to bind.

offset, of type GPUSize64, defaulting to 0

The offset, in bytes, from the beginning of buffer to the beginning of the range exposed to the shader by the buffer binding.

size, of type GPUSize64

The size, in bytes, of the buffer binding. If not provided, specifies the range starting at offset and ending at the end of buffer.

createBindGroup(descriptor)

Creates a GPUBindGroup.

Called on: GPUDevice this.

Arguments:

Arguments for the GPUDevice.createBindGroup(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPUBindGroupDescriptor Description of the GPUBindGroup to create.

Returns: GPUBindGroup

Content timeline steps:

  1. Let bindGroup be a new GPUBindGroup object.

  2. Issue the initialization steps on the Device timeline of this.

  3. Return bindGroup.

Device timeline initialization steps:
  1. Let limits be this.[[device]].[[limits]].

  2. If any of the following conditions are unsatisfied generate a validation error, make bindGroup invalid, and stop.

    For each GPUBindGroupEntry bindingDescriptor in descriptor.entries:

  3. Let bindGroup.[[layout]] = descriptor.layout.

  4. Let bindGroup.[[entries]] = descriptor.entries.

  5. Let bindGroup.[[usedResources]] = {}.

  6. For each GPUBindGroupEntry bindingDescriptor in descriptor.entries:

    1. Let internalUsage be the binding usage for layoutBinding.

    2. Each subresource seen by resource is added to [[usedResources]] as internalUsage.

effective buffer binding size(binding)
  1. If binding.size is not provided:

    1. Return max(0, binding.buffer.size - binding.offset);

  2. Return binding.size.

Two GPUBufferBinding objects a and b are considered buffer-binding-aliasing if and only if all of the following are true:

Note: When doing this calculation, any dynamic offsets have already been applied to the ranges.

8.3. GPUPipelineLayout

A GPUPipelineLayout defines the mapping between resources of all GPUBindGroup objects set up during command encoding in setBindGroup(), and the shaders of the pipeline set by GPURenderCommandsMixin.setPipeline or GPUComputePassEncoder.setPipeline.

The full binding address of a resource can be defined as a trio of:

  1. shader stage mask, to which the resource is visible

  2. bind group index

  3. binding number

The components of this address can also be seen as the binding space of a pipeline. A GPUBindGroup (with the corresponding GPUBindGroupLayout) covers that space for a fixed bind group index. The contained bindings need to be a superset of the resources used by the shader at this bind group index.

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUPipelineLayout {
};
GPUPipelineLayout includes GPUObjectBase;

GPUPipelineLayout has the following internal slots:

[[bindGroupLayouts]], of type list<GPUBindGroupLayout>

The GPUBindGroupLayout objects provided at creation in GPUPipelineLayoutDescriptor.bindGroupLayouts.

Note: using the same GPUPipelineLayout for many GPURenderPipeline or GPUComputePipeline pipelines guarantees that the user agent doesn’t need to rebind any resources internally when there is a switch between these pipelines.

GPUComputePipeline object X was created with GPUPipelineLayout.bindGroupLayouts A, B, C. GPUComputePipeline object Y was created with GPUPipelineLayout.bindGroupLayouts A, D, C. Supposing the command encoding sequence has two dispatches:
  1. setBindGroup(0, ...)

  2. setBindGroup(1, ...)

  3. setBindGroup(2, ...)

  4. setPipeline(X)

  5. dispatchWorkgroups()

  6. setBindGroup(1, ...)

  7. setPipeline(Y)

  8. dispatchWorkgroups()

In this scenario, the user agent would have to re-bind the group slot 2 for the second dispatch, even though neither the GPUBindGroupLayout at index 2 of GPUPipelineLayout.bindGroupLayouts, or the GPUBindGroup at slot 2, change.

Note: the expected usage of the GPUPipelineLayout is placing the most common and the least frequently changing bind groups at the "bottom" of the layout, meaning lower bind group slot numbers, like 0 or 1. The more frequently a bind group needs to change between draw calls, the higher its index should be. This general guideline allows the user agent to minimize state changes between draw calls, and consequently lower the CPU overhead.

8.3.1. Pipeline Layout Creation

A GPUPipelineLayout is created via GPUDevice.createPipelineLayout().

dictionary GPUPipelineLayoutDescriptor
         : GPUObjectDescriptorBase {
    required sequence<GPUBindGroupLayout> bindGroupLayouts;
};

GPUPipelineLayoutDescriptor dictionaries define all the GPUBindGroupLayouts used by a pipeline, and have the following members:

bindGroupLayouts, of type sequence<GPUBindGroupLayout>

A list of GPUBindGroupLayouts the pipeline will use. Each element corresponds to a @group attribute in the GPUShaderModule, with the Nth element corresponding with @group(N).

createPipelineLayout(descriptor)

Creates a GPUPipelineLayout.

Called on: GPUDevice this.

Arguments:

Arguments for the GPUDevice.createPipelineLayout(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPUPipelineLayoutDescriptor Description of the GPUPipelineLayout to create.

Returns: GPUPipelineLayout

Content timeline steps:

  1. Let pl be a new GPUPipelineLayout object.

  2. Issue the initialization steps on the Device timeline of this.

  3. Return pl.

Device timeline initialization steps:
  1. Let limits be this.[[device]].[[limits]].

  2. Let allEntries be the result of concatenating bgl.[[descriptor]].entries for all bgl in descriptor.bindGroupLayouts.

  3. If any of the following conditions are unsatisfied generate a validation error, make pl invalid, and stop.

  4. Set the pl.[[bindGroupLayouts]] to descriptor.bindGroupLayouts.

Note: two GPUPipelineLayout objects are considered equivalent for any usage if their internal [[bindGroupLayouts]] sequences contain GPUBindGroupLayout objects that are group-equivalent.

8.4. Example

Create a GPUBindGroupLayout that describes a binding with a uniform buffer, a texture, and a sampler. Then create a GPUBindGroup and a GPUPipelineLayout using the GPUBindGroupLayout.
const bindGroupLayout = gpuDevice.createBindGroupLayout({
    entries: [{
        binding: 0,
        visibility: GPUShaderStage.VERTEX | GPUShaderStage.FRAGMENT,
        buffer: {}
    }, {
        binding: 1,
        visibility: GPUShaderStage.FRAGMENT,
        texture: {}
    }, {
        binding: 2,
        visibility: GPUShaderStage.FRAGMENT,
        sampler: {}
    }]
});

const bindGroup = gpuDevice.createBindGroup({
    layout: bindGroupLayout,
    entries: [{
        binding: 0,
        resource: { buffer: buffer },
    }, {
        binding: 1,
        resource: texture
    }, {
        binding: 2,
        resource: sampler
    }]
});

const pipelineLayout = gpuDevice.createPipelineLayout({
    bindGroupLayouts: [bindGroupLayout]
});

9. Shader Modules

9.1. GPUShaderModule

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUShaderModule {
    Promise<GPUCompilationInfo> getCompilationInfo();
};
GPUShaderModule includes GPUObjectBase;

GPUShaderModule is a reference to an internal shader module object.

9.1.1. Shader Module Creation

dictionary GPUShaderModuleDescriptor
         : GPUObjectDescriptorBase {
    required USVString code;
    object sourceMap;
    record<USVString, GPUShaderModuleCompilationHint> hints;
};
code, of type USVString

The WGSL source code for the shader module.

sourceMap, of type object

If defined MAY be interpreted as a source-map-v3 format.

Source maps are optional, but serve as a standardized way to support dev-tool integration such as source-language debugging [SourceMap]. WGSL names (identifiers) in source maps follow the rules defined in WGSL identifier comparison.

hints, of type record<USVString, GPUShaderModuleCompilationHint>

If defined maps an entry point name from the shader to a GPUShaderModuleCompilationHint. No validation is performed with any of these GPUShaderModuleCompilationHint. Implementations should use any information present in the GPUShaderModuleCompilationHint to perform as much compilation as is possible within createShaderModule(). Entry point names follow the rules defined in WGSL identifier comparison.

Note: Supplying information in hints does not have any observable effect, other than performance. Because a single shader module can hold multiple entry points, and multiple pipelines can be created from a single shader module, it can be more performant for an implementation to do as much compilation as possible once in createShaderModule() rather than multiple times in the multiple calls to createComputePipeline() / createRenderPipeline().

createShaderModule(descriptor)

Creates a GPUShaderModule.

Called on: GPUDevice this.

Arguments:

Arguments for the GPUDevice.createShaderModule(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPUShaderModuleDescriptor Description of the GPUShaderModule to create.

Returns: GPUShaderModule

Content timeline steps:

  1. Let sm be a new GPUShaderModule object.

  2. Issue the initialization steps on the Device timeline of this.

  3. Return sm.

Device timeline initialization steps:
  1. Let result be the result of shader module creation with the WGSL source descriptor.code.

  2. If any of the following requirements are unmet, generate a validation error, make sm invalid, and return.

    Note: Uncategorized errors cannot arise from shader module creation. Implementations which detect such errors during shader module creation must behave as if the shader module is valid, and defer surfacing the error until pipeline creation.

Describe remaining createShaderModule() validation and algorithm steps.

Note: User agents should not include detailed compiler error messages or shader text in the message text of validation errors arising here: these details are accessible via getCompilationInfo(). User agents should surface human-readable, formatted error details to developers for easier debugging (for example as a warning in the browser developer console, expandable to show full shader source).

As shader compilation errors should be rare in production applications, user agents could choose to surface them to developers regardless of error handling (GPU error scopes or uncapturederror event handlers), e.g. as an expandable warning. If not, they should provide and document another way for developers to access human-readable error details, for example by adding a checkbox to show errors unconditionally, or by showing human-readable details when logging a GPUCompilationInfo object to the console.

Create a GPUShaderModule from WGSL code:
// A simple vertex and fragment shader pair that will fill the viewport with red.
const shaderSource = `
    var<private> pos : array<vec2<f32>, 3> = array<vec2<f32>, 3>(
        vec2(-1.0, -1.0), vec2(-1.0, 3.0), vec2(3.0, -1.0));

    @vertex
    fn vertexMain(@builtin(vertex_index) vertexIndex : u32) -> @builtin(position) vec4<f32> {
        return vec4(pos[vertexIndex], 1.0, 1.0);
    }

    @fragment
    fn fragmentMain() -> @location(0) vec4<f32> {
        return vec4(1.0, 0.0, 0.0, 1.0);
    }
`;

const shaderModule = gpuDevice.createShaderModule({
    code: shaderSource,
});
9.1.1.1. Shader Module Compilation Hints

Shader module compilation hints are optional, additional information indicating how a given GPUShaderModule entry point is intended to be used in the future. For some implementations this information may aid in compiling the shader module earlier, potentially increasing performance.

dictionary GPUShaderModuleCompilationHint {
    (GPUPipelineLayout or GPUAutoLayoutMode) layout;
};
layout, of type (GPUPipelineLayout or GPUAutoLayoutMode)

A GPUPipelineLayout that the GPUShaderModule may be used with in a future createComputePipeline() or createRenderPipeline() call. If set to "auto" the layout will be the default pipeline layout for the entry point associated with this hint will be used.

Note: If possible, authors should be supplying the same information to createShaderModule() and createComputePipeline() / createRenderPipeline().

If an author is unable to provide hint information at the time of calling createShaderModule(), they should usually not delay calling createShaderModule(); but should instead just omit the unknown information from hints or GPUShaderModuleCompilationHint. Omitting this information may cause compilation to be deferred to createComputePipeline() / createRenderPipeline().

If an author is not confident that the hint information passed to createShaderModule() will match the information later passed to createComputePipeline() / createRenderPipeline() with that same module, they should avoid passing that information to createShaderModule(), as passing mismatched information to createShaderModule() may cause unnecessary compilations to occur.

9.1.2. Shader Module Compilation Information

enum GPUCompilationMessageType {
    "error",
    "warning",
    "info",
};

[Exposed=(Window, DedicatedWorker), Serializable, SecureContext]
interface GPUCompilationMessage {
    readonly attribute DOMString message;
    readonly attribute GPUCompilationMessageType type;
    readonly attribute unsigned long long lineNum;
    readonly attribute unsigned long long linePos;
    readonly attribute unsigned long long offset;
    readonly attribute unsigned long long length;
};

[Exposed=(Window, DedicatedWorker), Serializable, SecureContext]
interface GPUCompilationInfo {
    readonly attribute FrozenArray<GPUCompilationMessage> messages;
};

A GPUCompilationMessage is an informational, warning, or error message generated by the GPUShaderModule compiler. The messages are intended to be human readable to help developers diagnose issues with their shader code. Each message may correspond to either a single point in the shader code, a substring of the shader code, or may not correspond to any specific point in the code at all.

GPUCompilationMessage has the following attributes:

message, of type DOMString, readonly

The human-readable, localizable text for this compilation message.

Note: The message should follow the best practices for language and direction information. This includes making use of any future standards which may emerge regarding the reporting of string language and direction metadata.

Editorial: At the time of this writing, no language/direction recommendation is available that provides compatibility and consistency with legacy APIs, but when there is, adopt it formally.

type, of type GPUCompilationMessageType, readonly

The severity level of the message.

If the type is "error", it corresponds to a shader-creation error.

lineNum, of type unsigned long long, readonly

The line number in the shader code the message corresponds to. Value is one-based, such that a lineNum of 1 indicates the first line of the shader code. Lines are delimited by line breaks.

If the message corresponds to a substring this points to the line on which the substring begins. Must be 0 if the message does not correspond to any specific point in the shader code.

linePos, of type unsigned long long, readonly

The offset, in UTF-16 code units, from the beginning of line lineNum of the shader code to the point or beginning of the substring that the message corresponds to. Value is one-based, such that a linePos of 1 indicates the first code unit of the line.

If message corresponds to a substring this points to the first UTF-16 code unit of the substring. Must be 0 if the message does not correspond to any specific point in the shader code.

offset, of type unsigned long long, readonly

The offset from the beginning of the shader code in UTF-16 code units to the point or beginning of the substring that message corresponds to. Must reference the same position as lineNum and linePos. Must be 0 if the message does not correspond to any specific point in the shader code.

length, of type unsigned long long, readonly

The number of UTF-16 code units in the substring that message corresponds to. If the message does not correspond with a substring then length must be 0.

Note: GPUCompilationMessage.lineNum and GPUCompilationMessage.linePos are one-based since the most common use for them is expected to be printing human readable messages that can be correlated with the line and column numbers shown in many text editors.

Note: GPUCompilationMessage.offset and GPUCompilationMessage.length are appropriate to pass to substr() in order to retrieve the substring of the shader code the message corresponds to.

getCompilationInfo()

Returns any messages generated during the GPUShaderModule's compilation.

The locations, order, and contents of messages are implementation-defined. In particular, messages may not be ordered by lineNum.

Called on: GPUShaderModule this

Returns: Promise<GPUCompilationInfo>

Content timeline steps:

  1. Let contentTimeline be the current Content timeline.

  2. Let promise be a new promise.

  3. Issue the synchronization steps on the Device timeline of this.

  4. Return promise.

Device timeline synchronization steps:
  1. When the device timeline becomes informed that shader module creation has completed for this:

    1. Let messages be a list of any errors, warnings, or informational messages generated during shader module creation for this.

    2. Issue the subsequent steps on contentTimeline.

Content timeline steps:
  1. Let info be a new GPUCompilationInfo.

  2. For each message in messages:

    1. Let m be a new GPUCompilationMessage.

    2. Set m.message to be the text of message.

    3. If message is a shader-creation error:

      Set m.type to "error"

      If message is a warning:

      Set m.type to "warning"

      Otherwise:

      Set m.type to "info"

    4. If message is associated with a specific substring or position within the shader code:
      1. Set m.lineNum to the one-based number of the first line that the message refers to.

      2. Set m.linePos to the one-based number of the first UTF-16 code units on m.lineNum that the message refers to, or 1 if the message refers to the entire line.

      3. Set m.offset to the number of UTF-16 code units from the beginning of the shader to beginning of the substring or position that message refers to.

      4. Set m.length the length of the substring in UTF-16 code units that message refers to, or 0 if message refers to a position

      Otherwise:
      1. Set m.lineNum to 0.

      2. Set m.linePos to 0.

      3. Set m.offset to 0.

      4. Set m.length to 0.

    5. Append m to info.messages.

  3. Resolve promise with info.

10. Pipelines

A pipeline, be it GPUComputePipeline or GPURenderPipeline, represents the complete function done by a combination of the GPU hardware, the driver, and the user agent, that process the input data in the shape of bindings and vertex buffers, and produces some output, like the colors in the output render targets.

Structurally, the pipeline consists of a sequence of programmable stages (shaders) and fixed-function states, such as the blending modes.

Note: Internally, depending on the target platform, the driver may convert some of the fixed-function states into shader code, and link it together with the shaders provided by the user. This linking is one of the reason the object is created as a whole.

This combination state is created as a single object (a GPUComputePipeline or GPURenderPipeline) and switched using one command (GPUComputePassEncoder.setPipeline() or GPURenderCommandsMixin.setPipeline() respectively).

There are two ways to create pipelines:

immediate pipeline creation

createComputePipeline() and createRenderPipeline() return a pipeline object which can be used immediately in a pass encoder.

When this fails, the pipeline object will be invalid and the call will generate either a validation error or an internal error.

Note: A handle object is returned immediately, but actual pipeline creation is not synchronous. If pipeline creation takes a long time, this can incur a stall in the device timeline at some point between the creation call and execution of the submit() in which it is first used. The point is unspecified, but most likely to be one of: at creation, at the first usage of the pipeline in setPipeline(), at the corresponding finish() of that GPUCommandEncoder or GPURenderBundleEncoder, or at submit() of that GPUCommandBuffer.

async pipeline creation

createComputePipelineAsync() and createRenderPipelineAsync() return a Promise which resolves to a pipeline object when creation of the pipeline has completed.

When this fails, the Promise rejects with a GPUPipelineError.

GPUPipelineError describes a pipeline creation failure.

[Exposed=(Window, DedicatedWorker), SecureContext, Serializable]
interface GPUPipelineError : DOMException {
    constructor(optional DOMString message = "", GPUPipelineErrorInit options);
    readonly attribute GPUPipelineErrorReason reason;
};

dictionary GPUPipelineErrorInit {
    required GPUPipelineErrorReason reason;
};

enum GPUPipelineErrorReason {
    "validation",
    "internal",
};

GPUPipelineError constructor:

constructor()
Arguments for the GPUPipelineError.constructor() method.
Parameter Type Nullable Optional Description
message DOMString Error message of the base DOMException.
options GPUPipelineErrorInit Options specific to GPUPipelineError.
  1. Set this.name to "GPUPipelineError".

  2. Set this.message to message.

  3. Set this.reason to options.reason.

GPUPipelineError has the following attributes:

reason, of type GPUPipelineErrorReason, readonly

A read-only slot-backed attribute exposing the type of error encountered in pipeline creation as a GPUPipelineErrorReason:

GPUPipelineError objects are serializable objects.

Their serialization steps, given value and serialized, are:
  1. Run the DOMException serialization steps given value and serialized.

Their deserialization steps, given value and serialized, are:
  1. Run the DOMException deserialization steps given value and serialized.

10.1. Base pipelines

enum GPUAutoLayoutMode {
    "auto",
};

dictionary GPUPipelineDescriptorBase
         : GPUObjectDescriptorBase {
    required (GPUPipelineLayout or GPUAutoLayoutMode) layout;
};
layout, of type (GPUPipelineLayout or GPUAutoLayoutMode)

The GPUPipelineLayout for this pipeline, or "auto" to generate the pipeline layout automatically.

Note: If "auto" is used the pipeline cannot share GPUBindGroups with any other pipelines.

interface mixin GPUPipelineBase {
    [NewObject] GPUBindGroupLayout getBindGroupLayout(unsigned long index);
};

GPUPipelineBase has the following internal slots:

[[layout]], of type GPUPipelineLayout

The definition of the layout of resources which can be used with this.

GPUPipelineBase has the following methods:

getBindGroupLayout(index)

Gets a GPUBindGroupLayout that is compatible with the GPUPipelineBase's GPUBindGroupLayout at index.

Called on: GPUPipelineBase this

Arguments:

Arguments for the GPUPipelineBase.getBindGroupLayout(index) method.
Parameter Type Nullable Optional Description
index unsigned long Index into the pipeline layout’s [[bindGroupLayouts]] sequence.

Returns: GPUBindGroupLayout

Content timeline steps:

  1. Let layout be a new GPUBindGroupLayout object.

  2. Issue the initialization steps on the Device timeline of this.

  3. Return layout.

Device timeline initialization steps:
  1. If any of the following conditions are unsatisfied generate a validation error, make layout invalid, and stop.

  2. Initialize layout so it is a copy of this.[[layout]].[[bindGroupLayouts]][index].

    Note: GPUBindGroupLayout is only ever used by-value, not by-reference, so this is equivalent to returning the same internal object in a new wrapper. A new GPUBindGroupLayout wrapper is returned each time to avoid a round-trip between the Content timeline and the Device timeline.

10.1.1. Default pipeline layout

A GPUPipelineBase object that was created with a layout set to "auto" has a default layout created and used instead.

Note: Default layouts are provided as a convenience for simple pipelines, but use of explicit layouts is recommended in most cases. Bind groups created from default layouts cannot be used with other pipelines, and the structure of the default layout may change when altering shaders, causing unexpected bind group creation errors.

To create a default pipeline layout for GPUPipelineBase pipeline, run the following steps:

  1. Let groupCount be 0.

  2. Let groupDescs be a sequence of device.[[limits]].maxBindGroups new GPUBindGroupLayoutDescriptor objects.

  3. For each groupDesc in groupDescs:

    1. Set groupDesc.entries to an empty sequence.

  4. For each GPUProgrammableStage stageDesc in the descriptor used to create pipeline:

    1. Let shaderStage be the GPUShaderStageFlags for stageDesc.entryPoint in stageDesc.module.

    2. For each resource resource statically used by stageDesc:

      1. Let group be resource’s "group" decoration.

      2. Let binding be resource’s "binding" decoration.

      3. Let entry be a new GPUBindGroupLayoutEntry.

      4. Set entry.binding to binding.

      5. Set entry.visibility to shaderStage.

      6. If resource is for a sampler binding:

        1. Let samplerLayout be a new GPUSamplerBindingLayout.

        2. Set entry.sampler to samplerLayout.

      7. If resource is for a comparison sampler binding:

        1. Let samplerLayout be a new GPUSamplerBindingLayout.

        2. Set samplerLayout.type to "comparison".

        3. Set entry.sampler to samplerLayout.

      8. If resource is for a buffer binding:

        1. Let bufferLayout be a new GPUBufferBindingLayout.

        2. Set bufferLayout.minBindingSize to resource’s minimum buffer binding size.

        3. If resource is for a read-only storage buffer:

          1. Set bufferLayout.type to "read-only-storage".

        4. If resource is for a storage buffer:

          1. Set bufferLayout.type to "storage".

        5. Set entry.buffer to bufferLayout.

      9. If resource is for a sampled texture binding:

        1. Let textureLayout be a new GPUTextureBindingLayout.

        2. If resource is a depth texture binding:

          Else if the sampled type of resource is:

          f32 and there exists a static use of resource with a textureSample* builtin in

          Set textureLayout.sampleType to "float"

          f32 otherwise

          Set textureLayout.sampleType to "unfilterable-float"

          i32

          Set textureLayout.sampleType to "sint"

          u32

          Set textureLayout.sampleType to "uint"

        3. Set textureLayout.viewDimension to resource’s dimension.

        4. If resource is for a multisampled texture:

          1. Set textureLayout.multisampled to true.

        5. Set entry.texture to textureLayout.

      10. If resource is for a storage texture binding:

        1. Let storageTextureLayout be a new GPUStorageTextureBindingLayout.

        2. Set storageTextureLayout.format to resource’s format.

        3. Set storageTextureLayout.viewDimension to resource’s dimension.

        4. If resource is for a write-only storage texture:

          1. Set storageTextureLayout.access to "write-only".

        5. Set entry.storageTexture to storageTextureLayout.

      11. Set groupCount to max(groupCount, group + 1).

      12. If groupDescs[group] has an entry previousEntry with binding equal to binding:

        1. If entry has different visibility than previousEntry:

          1. Add the bits set in entry.visibility into previousEntry.visibility

        2. If resource is for a buffer binding and entry has greater buffer.minBindingSize than previousEntry:

          1. Set previousEntry.buffer.minBindingSize to entry.buffer.minBindingSize.

        3. If resource is a sampled texture binding and entry has different texture.sampleType than previousEntry and both entry and previousEntry have texture.sampleType of either "float" or "unfilterable-float":

          1. Set previousEntry.texture.sampleType to "float".

        4. If any other property is unequal between entry and previousEntry:

          1. Return null (which will cause the creation of the pipeline to fail).

      13. Else

        1. Append entry to groupDescs[group].

  5. Let groupLayouts be a new list.

  6. For each i from 0 to groupCount - 1, inclusive:

    1. Let groupDesc be groupDescs[i].

    2. Let bindGroupLayout be the result of calling device.createBindGroupLayout()(groupDesc).

    3. Set bindGroupLayout.[[exclusivePipeline]] to pipeline.

    4. Append bindGroupLayout to groupLayouts.

  7. Let desc be a new GPUPipelineLayoutDescriptor.

  8. Set desc.bindGroupLayouts to groupLayouts.

  9. Return device.createPipelineLayout()(desc).

10.1.2. GPUProgrammableStage

A GPUProgrammableStage describes the entry point in the user-provided GPUShaderModule that controls one of the programmable stages of a pipeline. Entry point names follow the rules defined in WGSL identifier comparison.

dictionary GPUProgrammableStage {
    required GPUShaderModule module;
    required USVString entryPoint;
    record<USVString, GPUPipelineConstantValue> constants;
};

typedef double GPUPipelineConstantValue; // May represent WGSL’s bool, f32, i32, u32, and f16 if enabled.

GPUProgrammableStage has the following members:

module, of type GPUShaderModule

The GPUShaderModule containing the code that this programmable stage will execute.

entryPoint, of type USVString

The name of the function in module that this stage will use to perform its work.

constants, of type record<USVString, GPUPipelineConstantValue>

Specifies the values of pipeline-overridable constants in the shader module module.

Each such pipeline-overridable constant is uniquely identified by a single pipeline-overridable constant identifier string (representing the numeric ID of the constant, if one is specified, and otherwise the constant’s identifier name). WGSL names (identifiers) in source maps follow the rules defined in WGSL identifier comparison.

The key of each key-value pair must equal the identifier string of one such constant. When the pipeline is executed, that constant will have the specified value.

Values are specified as GPUPipelineConstantValue, which is a double. They are converted to WGSL type of the pipeline-overridable constant (bool/i32/u32/f32/f16). If conversion fails, a validation error is generated.

Pipeline-overridable constants defined in WGSL:
@id(0)      override has_point_light: bool = true;  // Algorithmic control.
@id(1200)   override specular_param: f32 = 2.3;     // Numeric control.
@id(1300)   override gain: f32;                     // Must be overridden.
            override width: f32 = 0.0;              // Specifed at the API level
                                                    //   using the name "width".
            override depth: f32;                    // Specifed at the API level
                                                    //   using the name "depth".
                                                    //   Must be overridden.
            override height = 2 * depth;            // The default value
                                                    // (if not set at the API level),
                                                    // depends on another
                                                    // overridable constant.

Corresponding JavaScript code, providing only the overrides which are required (have no defaults):

{
    // ...
    constants: {
        1300: 2.0,  // "gain"
        depth: -1,  // "depth"
    }
}

Corresponding JavaScript code, overriding all constants:

{
    // ...
    constants: {
        0: false,   // "has_point_light"
        1200: 3.0,  // "specular_param"
        1300: 2.0,  // "gain"
        width: 20,  // "width"
        depth: -1,  // "depth"
        height: 15, // "height"
    }
}
validating GPUProgrammableStage(stage, descriptor, layout)

Arguments:

Return true if all of the following conditions are met, and false otherwise:

validating shader binding(binding, layout)

Arguments:

Let bindGroup be the bind group index, and bindIndex be the binding index, of the shader binding declaration variable.

Return true if all of the following conditions are satisfied:

The minimum buffer binding size for a buffer binding variable var is computed as follows:
  1. Let T be the store type of var.

  2. If T is a runtime-sized array, or contains a runtime-sized array, replace that array<E> with array<E, 1>.

    Note: This ensures there’s always enough memory for one element, which allows array indices to be clamped to the length of the array resulting in an in-memory access.

  3. Return SizeOf(T).

Note: Enforcing this lower bound ensures reads and writes via the buffer variable only access memory locations within the bound region of the buffer.

A resource binding, pipeline-overridable constant, shader stage input, or shader stage output is considered to be statically used by a GPUProgrammableStage if it is present in the interface of the shader stage of the specified entryPoint, in the specified shader module.

10.2. GPUComputePipeline

A GPUComputePipeline is a kind of pipeline that controls the compute shader stage, and can be used in GPUComputePassEncoder.

Compute inputs and outputs are all contained in the bindings, according to the given GPUPipelineLayout. The outputs correspond to buffer bindings with a type of "storage" and storageTexture bindings with a type of "write-only".

Stages of a compute pipeline:

  1. Compute shader

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPUComputePipeline {
};
GPUComputePipeline includes GPUObjectBase;
GPUComputePipeline includes GPUPipelineBase;

10.2.1. Compute Pipeline Creation

A GPUComputePipelineDescriptor describes a compute pipeline. See § 23.2 Computing for additional details.

dictionary GPUComputePipelineDescriptor
         : GPUPipelineDescriptorBase {
    required GPUProgrammableStage compute;
};

GPUComputePipelineDescriptor has the following members:

compute, of type GPUProgrammableStage

Describes the compute shader entry point of the pipeline.

createComputePipeline(descriptor)

Creates a GPUComputePipeline using immediate pipeline creation.

Called on: GPUDevice this.

Arguments:

Arguments for the GPUDevice.createComputePipeline(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPUComputePipelineDescriptor Description of the GPUComputePipeline to create.

Returns: GPUComputePipeline

Content timeline steps:

  1. Let pipeline be a new GPUComputePipeline object.

  2. Issue the initialization steps on the Device timeline of this.

  3. Return pipeline.

Device timeline initialization steps:
  1. Let layout be a new default pipeline layout for pipeline if descriptor.layout is "auto", and descriptor.layout otherwise.

  2. If any of the following conditions are unsatisfied generate a validation error, make pipeline invalid, and stop.

  3. If any pipeline-creation uncategorized errors result from the implementation of pipeline creation, generate an internal error, make pipeline invalid, and stop.

    Note: Even if the implementation detected uncategorized errors in shader module creation, the error is surfaced here.

  4. Set pipeline.[[layout]] to layout.

createComputePipelineAsync(descriptor)

Creates a GPUComputePipeline using async pipeline creation. The returned Promise resolves when the created pipeline is ready to be used without additional delay.

If pipeline creation fails, the returned Promise rejects with an GPUPipelineError.

Note: Use of this method is preferred whenever possible, as it prevents blocking the queue timeline work on pipeline compilation.

Called on: GPUDevice this.

Arguments:

Arguments for the GPUDevice.createComputePipelineAsync(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPUComputePipelineDescriptor Description of the GPUComputePipeline to create.

Returns: Promise<GPUComputePipeline>

Content timeline steps:

  1. Let contentTimeline be the current Content timeline.

  2. Let promise be a new promise.

  3. Issue the initialization steps on the Device timeline of this.

  4. Return promise.

Device timeline initialization steps:
  1. Let pipeline be a new GPUComputePipeline created as if this.createComputePipeline() was called with descriptor;

  2. When pipeline is ready to be used or has been made invalid, issue the subsequent steps on contentTimeline.

Content timeline steps:
  1. If pipeline...

    valid

    Resolve promise with pipeline.

    invalid due to an internal error

    Reject promise with a GPUPipelineError with reason "internal".

    invalid due to an validation error

    Reject promise with a GPUPipelineError with reason "validation".

Creating a simple GPUComputePipeline:
const computePipeline = gpuDevice.createComputePipeline({
    layout: pipelineLayout,
    compute: {
        module: computeShaderModule,
        entryPoint: 'computeMain',
    }
});

10.3. GPURenderPipeline

A GPURenderPipeline is a kind of pipeline that controls the vertex and fragment shader stages, and can be used in GPURenderPassEncoder as well as GPURenderBundleEncoder.

Render pipeline inputs are:

Render pipeline outputs are:

A render pipeline is comprised of the following render stages:

  1. Vertex fetch, controlled by GPUVertexState.buffers

  2. Vertex shader, controlled by GPUVertexState

  3. Primitive assembly, controlled by GPUPrimitiveState

  4. Rasterization, controlled by GPUPrimitiveState, GPUDepthStencilState, and GPUMultisampleState

  5. Fragment shader, controlled by GPUFragmentState

  6. Stencil test and operation, controlled by GPUDepthStencilState

  7. Depth test and write, controlled by GPUDepthStencilState

  8. Output merging, controlled by GPUFragmentState.targets

[Exposed=(Window, DedicatedWorker), SecureContext]
interface GPURenderPipeline {
};
GPURenderPipeline includes GPUObjectBase;
GPURenderPipeline includes GPUPipelineBase;

GPURenderPipeline has the following internal slots:

[[descriptor]], of type GPURenderPipelineDescriptor

The GPURenderPipelineDescriptor describing this pipeline.

All optional fields of GPURenderPipelineDescriptor are defined.

[[writesDepth]], of type boolean

True if the pipeline writes to the depth component of the depth/stencil attachment

[[writesStencil]], of type boolean

True if the pipeline writes to the stencil component of the depth/stencil attachment

10.3.1. Render Pipeline Creation

A GPURenderPipelineDescriptor describes a render pipeline by configuring each of the render stages. See § 23.3 Rendering for additional details.

dictionary GPURenderPipelineDescriptor
         : GPUPipelineDescriptorBase {
    required GPUVertexState vertex;
    GPUPrimitiveState primitive = {};
    GPUDepthStencilState depthStencil;
    GPUMultisampleState multisample = {};
    GPUFragmentState fragment;
};

GPURenderPipelineDescriptor has the following members:

vertex, of type GPUVertexState

Describes the vertex shader entry point of the pipeline and its input buffer layouts.

primitive, of type GPUPrimitiveState, defaulting to {}

Describes the primitive-related properties of the pipeline.

depthStencil, of type GPUDepthStencilState

Describes the optional depth-stencil properties, including the testing, operations, and bias.

multisample, of type GPUMultisampleState, defaulting to {}

Describes the multi-sampling properties of the pipeline.

fragment, of type GPUFragmentState

Describes the fragment shader entry point of the pipeline and its output colors. If not provided, the § 23.3.8 No Color Output mode is enabled.

createRenderPipeline(descriptor)

Creates a GPURenderPipeline using immediate pipeline creation.

Called on: GPUDevice this.

Arguments:

Arguments for the GPUDevice.createRenderPipeline(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPURenderPipelineDescriptor Description of the GPURenderPipeline to create.

Returns: GPURenderPipeline

Content timeline steps:

  1. If descriptor.fragment is provided:

    1. For each non-null colorState of descriptor.fragment.targets:

      1. ? Validate texture format required features of colorState.format with this.[[device]].

  2. If descriptor.depthStencil is provided:

    1. ? Validate texture format required features of descriptor.depthStencil.format with this.[[device]].

  3. Let pipeline be a new GPURenderPipeline object.

  4. Issue the initialization steps on the Device timeline of this.

  5. Return pipeline.

Device timeline initialization steps:
  1. Let layout be a new default pipeline layout for pipeline if descriptor.layout is "auto", and descriptor.layout otherwise.

  2. If any of the following conditions are unsatisfied: generate a validation error, make pipeline invalid, and stop.

  3. If any pipeline-creation uncategorized errors result from the implementation of pipeline creation, generate an internal error, make pipeline invalid, and stop.

    Note: Even if the implementation detected uncategorized errors in shader module creation, the error is surfaced here.

  4. Set pipeline.[[descriptor]] to descriptor.

  5. Set pipeline.[[writesDepth]] to false.

  6. Set pipeline.[[writesStencil]] to false.

  7. Let depthStencil be descriptor.depthStencil.

  8. If depthStencil is not null:

    1. Set pipeline.[[writesDepth]] to depthStencil.depthWriteEnabled.

    2. If depthStencil.stencilWriteMask is not 0:

      1. Let stencilFront be depthStencil.stencilFront.

      2. Let stencilBack be depthStencil.stencilBack.

      3. Let cullMode be descriptor.primitive.cullMode.

      4. If cullMode is not "front", and any of stencilFront.passOp, stencilFront.depthFailOp, or stencilFront.failOp is not "keep":

        1. Set pipeline.[[writesStencil]] to true.

      5. If cullMode is not "back", and any of stencilBack.passOp, stencilBack.depthFailOp, or stencilBack.failOp is not "keep":

        1. Set pipeline.[[writesStencil]] to true.

  9. Set pipeline.[[layout]] to layout.

need description of the render states.

createRenderPipelineAsync(descriptor)

Creates a GPURenderPipeline using async pipeline creation. The returned Promise resolves when the created pipeline is ready to be used without additional delay.

If pipeline creation fails, the returned Promise rejects with an GPUPipelineError.

Note: Use of this method is preferred whenever possible, as it prevents blocking the queue timeline work on pipeline compilation.

Called on: GPUDevice this.

Arguments:

Arguments for the GPUDevice.createRenderPipelineAsync(descriptor) method.
Parameter Type Nullable Optional Description
descriptor GPURenderPipelineDescriptor Description of the GPURenderPipeline to create.

Returns: Promise<GPURenderPipeline>

Content timeline steps:

  1. Let contentTimeline be the current Content timeline.

  2. Let promise be a new promise.

  3. Issue the initialization steps on the Device timeline of this.

  4. Return promise.

Device timeline initialization steps:
  1. Let pipeline be a new GPURenderPipeline created as if this.createRenderPipeline() was called with descriptor;

  2. When pipeline is ready to be used or has been made invalid, issue the subsequent steps on contentTimeline.

Content timeline steps:
  1. If pipeline is...

    valid

    Resolve promise with pipeline.

    invalid due to an internal error

    Reject promise with a GPUPipelineError with reason "internal".

    invalid due to an validation error

    Reject promise with a GPUPipelineError with reason "validation".

validating GPURenderPipelineDescriptor(descriptor, layout, device)

Arguments:

Return true if all of the following conditions are satisfied: