Scalable Video Coding (SVC) Extension for WebRTC

W3C Working Draft

More details about this document
This version:
https://www.w3.org/TR/2022/WD-webrtc-svc-20220303/
Latest published version:
https://www.w3.org/TR/webrtc-svc/
Latest editor's draft:
https://w3c.github.io/webrtc-svc/
History:
https://www.w3.org/standards/history/webrtc-svc
Commit history
Editor:
Bernard Aboba (Microsoft Corporation)
Former editor:
Peter Thatcher (Google) - Until
Feedback:
GitHub w3c/webrtc-svc (pull requests, new issue, open issues)
public-webrtc@w3.org with subject line [webrtc-svc] … message topic … (archives)
Participate
Mailing list
IETF AVTCORE Working Group

Abstract

This document defines a set of ECMAScript APIs in WebIDL to extend the WebRTC 1.0 API to enable user agents to support scalable video coding (SVC).

Status of This Document

This section describes the status of this document at the time of its publication. A list of current W3C publications and the latest revision of this technical report can be found in the W3C technical reports index at https://www.w3.org/TR/.

The API is based on preliminary work done in the W3C ORTC Community Group.

This document was published by the Web Real-Time Communications Working Group as a Working Draft using the Recommendation track.

Publication as a Working Draft does not imply endorsement by W3C and its Members.

This is a draft document and may be updated, replaced or obsoleted by other documents at any time. It is inappropriate to cite this document as other than work in progress.

This document was produced by a group operating under the W3C Patent Policy. W3C maintains a public list of any patent disclosures made in connection with the deliverables of the group; that page also includes instructions for disclosing a patent. An individual who has actual knowledge of a patent which the individual believes contains Essential Claim(s) must disclose the information in accordance with section 6 of the W3C Patent Policy.

This document is governed by the 2 November 2021 W3C Process Document.

1. Introduction

This section is non-normative.

This specification extends the WebRTC specification [WEBRTC] to enable configuration of encoding parameters, as well as the discovery of Scalable Video Coding (SVC) encoder capabilities. The discovery of decoder capabilities and configuration of decoding parameters is not supported.

Since this specification does not change the behavior of WebRTC objects and methods, restrictions relating to Offer/Answer negotiation and encoding parameters remain, as described in [WEBRTC] Section 5.2: "setParameters() does not cause SDP renegotiation and can only be used to change what the media stack is sending or receiving within the envelope negotiated by Offer/Answer."

The configuration of SVC-capable codecs implemented in browsers fits within this restriction. Codecs such as VP8 [RFC6386], VP9 [VP9] and AV1 [AV1] do not negotiate SVC support within Offer/Answer, enabling encoding parameters to be used for SVC configuration.

2. Conformance

As well as sections marked as non-normative, all authoring guidelines, diagrams, examples, and notes in this specification are non-normative. Everything else in this specification is normative.

The key words MUST and SHOULD in this document are to be interpreted as described in BCP 14 [RFC2119] [RFC8174] when, and only when, they appear in all capitals, as shown here.

This specification defines conformance criteria that apply to a single product: the user agent that implements the interfaces that it contains.

Conformance requirements phrased as algorithms or specific steps may be implemented in any manner, so long as the end result is equivalent. In particular, the algorithms defined in this specification are intended to be easy to follow, and not intended to be performant.

Implementations that use ECMAScript to implement the APIs defined in this specification MUST implement them in a manner consistent with the ECMAScript Bindings defined in the Web IDL specification [WEBIDL], as this specification uses that specification and terminology.

3. Terminology

The term "simulcast envelope" refers to the maximum number of simulcast streams and the order of the encoding parameters.

This specification references objects, methods, internal slots and dictionaries defined in [WEBRTC].

For Scalable Video Coding (SVC), the terms "single-session transmission" (SST) and "multi-session transmission" (MST) are defined in [RFC6190]. This specification only supports SST but not MST.

The term "Single Real-time Transport Protocol stream Single Transport" (SRST), defined in [RFC7656] Section 3.7, refers to SVC implementations that transmit all layers within a single transport, using a single Real-time Transport Protocol (RTP) stream and synchronization source (SSRC). The term "Multiple RTP stream Single Transport" (MRST), also defined in [RFC7656] Section 3.7, refers to implementations that transmit all layers within a single transport, using multiple RTP streams with a distinct SSRC for each layer. This specification only supports SRST, not MRST. Codecs with RTP payload specifications supporting SRST include VP8 [RFC7741], VP9 [VP9-PAYLOAD], AV1 [AV1-RTP] and H.264/SVC [RFC6190].

The term "S mode" refers to a scalability mode in which multiple encodings are sent on the same SSRC. This includes the "S2T1", "S2T1h", "S2T2", "S2T2h", "S2T3", "S2T3h", "S3T1", "S3T1h", "S3T2", "S3T2h", "S3T3" and "S3T3h" scalabilityMode values.

The term Selective Forwarding Middlebox (SFM) is defined in Section 3.7 of [RFC7667].

spatialScalability is a proposed addition to VideoConfiguration in the Media Capabilities API.

4. Configuration

This specification enables the configuration of encoding parameters for SVC by extending the RTCRtpEncodingParameters dictionary.

4.1 RTCRtpEncodingParameters Dictionary Extensions

WebIDLpartial dictionary RTCRtpEncodingParameters {
             DOMString scalabilityMode;
};

Dictionary RTCRtpEncodingParameters Members

scalabilityMode of type DOMString

A case-sensitive identifier of the scalability mode to be used for this stream. Scalability modes are defined in Section 6.

4.2 Behavior

[WEBRTC] describes error handling in addTransceiver() (Section 5.1) and setParameters() (Section 5.2), including use of RTCError to indicate a "hardware-encoder-error" due to an unsupported encoding parameter, as well as other errors. Implementations utilize RTCError and other errors in the prescribed manner when an invalid scalabilityMode value is provided to setParameters() or addTransceiver().

[WEBRTC] Section 5.1 describes validation of sendEncodings within addTransceiver(). To validate scalabilityMode, add the following validation steps after step 3 (within step 8):

  1. If sendEncodings contains any encoding whose scalabilityMode value is not supported by any codec in RTCRtpSender.getCapabilities(kind).codecs, throw an OperationError.
  2. If the number of RTCRtpEncodingParameters stored in sendEncodings is more than 1, and sendEncodings contains any encoding whose scalabilityMode value represents an "S mode" throw an OperationError.

When the addTransceiver() and setCodecPreferences() methods are called prior to conclusion of the Offer/Answer negotiation, the negotiated codec and its capabilities may not be known. In this situation the scalabilityMode values configured in sendEncodings may not be supported by the eventually negotiated codec. However, an error will only result if the requested scalabilityMode value is invalid for any supported codec, or if mixed simulcast transport is requested.

So as to ensure that the desired scalabilityMode values can be applied, setCodecPreferences() can be used to prefer or only include codecs supporting the desired configuration. For example, if temporal scalability is desired along with spatial simulcast, when addTransceiver() is called, sendEncodings can be configured to send multiple simulcast streams with different resolutions, with each stream utilizing temporal scalability. If only the VP8, VP9 and AV1 codec implementations support temporal scalability, setCodecPreferences() can be used to remove the H.264/AVC codec from the Offer, improving the chances that a codec supporting temporal scalability is negotiated.

When sendEncodings is used to request the sending of multiple simulcast streams using addTransceiver(), an "S mode" cannot be requested. The browser may only be configured to send simulcast encodings with multiple SSRCs and RIDs, or alternatively, to send all simulcast encodings on a single RTP stream. Simultaneously using both simulcast transport techniques is not permitted.

[WEBRTC] Section 5.2 describes validation of parameters within setParameters(). Add the following to the conditions under which the operation causes a promise rejected with an InvalidModificationError (step 4 within step 6):

  1. Before initial negotiation has concluded, encodings contains any encoding whose scalabilityMode value is not supported by any codec in RTCRtpSender.getCapabilities(kind).codecs.
  2. After initial negotiation has concluded, encodings contains an encoding whose scalabilityMode value is not supported by the most preferred codec.
  3. N is greater than 1, and encodings contains an encoding whose scalabilityMode value represents an "S mode".

Before the initial negotiation has completed, getParameters() returns the scalabilityMode value for each encoding in encodings, as last set by addTransceiver() or setParameters(). If no scalabilityMode value was provided for an encoding in encodings, or if a value was not successfully set, then getParameters() will not return a scalabilityMode value for that encoding.

After the initial negotiation has completed, getParameters() returns the currently configured scalabilityMode value for each encoding in encodings. This may be different from the values requested in addTransceiver() or setParameters(). If the configuration is not satisfactory, setParameters() can be used to change it.

If addTransceiver() or setParameters() did not provide an scalabilityMode value for an encoding in encodings, then after the initial negotiation has completed, getParameters() returns the default scalabilityMode of the most preferred codec for that encoding. The most preferred codec and the default scalabilityMode for each ecodec are both implementation dependent. The default scalabilityMode SHOULD be one of the temporal scalability modes (e.g. "L1T1","L1T2","L1T3", etc.).

5. Discovery

The SVC capabilities of an encoder can be discovered using extensions to the RTCRtpCodecCapability dictionary.

RTCRtpSender.getCapabilities(kind) provides information on what codecs and scalability modes and scalability modes an application can send. The SFM can provide information on the codecs and scalability modes it can decode by providing its receiver capabilities. After exchanging capabilities, the application can compute the intersection of codecs and scalabilityMode values supported by the browser's RTCRtpSender and the SFM's receiver. This can be used to determine the arguments passed to the browser's addTransceiver() and setParameters() methods.

The [Media-Capabilities] API provides information on decoder support for spatial scalablity modes. spatialScalability indicates whether a decoder has the ability to support spatial prediction, which requires the ability to use frames of a resolution different than the current resolution as a dependency. If spatialScalability is set to true, the decoder can decode any scalabilityMode value supported by the encoder. If spatialScalability is set to false or is absent, the decoder cannot decode spatial scalability modes, but can can decode all other scalabilityMode values supported by the encoder.

5.1 RTCRtpCodecCapability Dictionary Extensions

WebIDLpartial dictionary RTCRtpCodecCapability {
             sequence<DOMString> scalabilityModes;
};

Dictionary RTCRtpCodecCapability Members

scalabilityModes of type sequence<DOMString>

A sequence of the scalability modes (defined in Section 6) supported by the encoder implementation.

In response to RTCRtpSender.getCapabilities(kind), implementations of this specification MUST return a sequence of scalability modes supported by each codec of that kind. If a codec does not support encoding of scalability modes other than "L1T1", then the scalabilityModes member is not provided. The "L1T1" scalability mode enables SVC encoding to be turned off using setParameters(), so that it MUST be included within the sequence of scalability modes returned by RTCRtpSender.getCapabilities(kind) in order for it to be considered valid within setParameters(). If "L1T1" is set using setParameters() then it will be returned in response to getParameters().

In response to RTCRtpReceiver.getCapabilities(kind), scalabilityModes are not provided.

5.2 Negotiation

There are situations where an SFM may only support reception of a subset of codecs and scalability modes. For example, an SFM that parses codec payloads may only support the H.264/AVC codec without scalability and the VP8 codec with temporal scalability. On the other hand, the browser may be able to encode VP8 with temporal scalability, VP9 with temporal and spatial scalability and or H.264/AVC with temporal scalability. In such a situation, an application desiring to use SVC would only be able to encode VP8 with temporal scalability.

Since sending simulcast encodings on a single stream is not negotiated within Offer/Answer, an application using SDP signaling needs to determine whether single stream simulcast transport is supported prior to the Offer/Answer negotiation. This can be handled by having the SFM send it's receiver capabilities to the application prior to Offer/Answer. This allows the application to determine whether single stream simulcast is supported, and if so, what scalability modes the SFM can handle. For example, an SFM that can only support reception of a maximum of 2 simulcast encodings on a single SSRC with the AV1 codec would only indicate support for the "S2T1" and "S2T1h" scalability modes in its receiver capabilities.

For an SFM the supported scalabilityModes may depend on the negotiated RTP header extensions. For example, if the SFM cannot parse codec payloads (either because it is not designed to do so, or because the payloads are encrypted), then negotiation of an RTP header extension (such as the AV1 Dependency Descriptor defined in Appendix A of [AV1-RTP]) could be a prerequisite for the SFM to forward scalabilityModes. As a result, the scalabilityModes supported by an SFM may not be determined until completion of the Offer/Answer negotiation.

6. Scalability modes

The scalability modes supported in this specification, as well as their associated identifiers and characteristics, are provided in the table below. The names of the scability modes (which are case sensitive) are provided, along with the scalability mode identifiers assigned in [AV1] Section 6.7.5, and links to dependency diagrams provided in Section 9.

While the [AV1] and VP9 [VP9] specifications support all the modes defined in the table, other codec specifications do not. For example, VP8 [RFC6386] only supports temporal scalability (e.g. "L1T2", "L1T3"); H.264/SVC [RFC6190], which supports both temporal and spatial scalability, only permits transport of simulcast on distinct SSRCs, so that it does not support the "S" modes, where multiple encodings are transported on a single RTP stream.

Scalability Mode Identifier Spatial Layers Resolution Ratio Temporal Layers Inter-layer dependency AV1 scalability_mode_idc
"L1T1" 1 1 N/A
"L1T2" 1 2 SCALABILITY_L1T2
"L1T3" 1 3 SCALABILITY_L1T3
"L2T1" 2 2:1 1 Yes SCALABILITY_L2T1
"L2T2" 2 2:1 2 Yes SCALABILITY_L2T2
"L2T3" 2 2:1 3 Yes SCALABILITY_L2T3
"L3T1" 3 2:1 1 Yes SCALABILITY_L3T1
"L3T2" 3 2:1 2 Yes SCALABILITY_L3T2
"L3T3" 3 2:1 3 Yes SCALABILITY_L3T3
"L2T1h" 2 1.5:1 1 Yes SCALABILITY_L2T1h
"L2T2h" 2 1.5:1 2 Yes SCALABILITY_L2T2h
"L2T3h" 2 1.5:1 3 Yes SCALABILITY_L2T3h
"S2T1" 2 2:1 1 No SCALABILITY_S2T1
"S2T2" 2 2:1 2 No SCALABILITY_S2T2
"S2T3" 2 2:1 3 No SCALABILITY_S2T3
"S2T1h" 2 1.5:1 1 No SCALABILITY_S2T1h
"S2T2h" 2 1.5:1 2 No SCALABILITY_S2T2h
"S2T3h" 2 1.5:1 3 No SCALABILITY_S2T3h
"S3T1" 3 2:1 1 No SCALABILITY_S3T1
"S3T2" 3 2:1 2 No SCALABILITY_S3T2
"S3T3" 3 2:1 3 No SCALABILITY_S3T3
"S3T1h" 3 1.5:1 1 No SCALABILITY_S3T1h
"S3T2h" 3 1.5:1 2 No SCALABILITY_S3T2h
"S3T3h" 3 1.5:1 3 No SCALABILITY_S3T3h
"L2T2_KEY" 2 2:1 2 Yes SCALABILITY_L3T2_KEY
"L2T2_KEY_SHIFT" 2 2:1 2 Yes SCALABILITY_L3T2_KEY_SHIFT
"L2T3_KEY" 2 2:1 3 Yes SCALABILITY_L3T3_KEY
"L2T3_KEY_SHIFT" 2 2:1 3 Yes SCALABILITY_L3T3_KEY_SHIFT
"L3T2_KEY" 3 2:1 2 Yes SCALABILITY_L4T5_KEY
"L3T2_KEY_SHIFT" 3 2:1 2 Yes SCALABILITY_L4T5_KEY_SHIFT
"L3T3_KEY" 3 2:1 3 Yes SCALABILITY_L4T7_KEY
"L3T3_KEY_SHIFT" 3 2:1 3 Yes SCALABILITY_L4T7_KEY_SHIFT

6.1 Guidelines for addition of scalabilityMode values

When proposing a scalabilityMode value, the following principles should be followed:

  1. The proposed scalabilityMode MUST define entries to the table in Section 6, including values for the Scalabilty Mode Identifier, spatial and temporal layers, Resolution Ratio, Inter-layer dependency and the corresponding AV1 scalability_mode_idc value (if assigned).
  2. The Scalability Mode Identifier SHOULD be consistent with the existing naming scheme, which utilizes LxTy to denote a scalabilityMode with x spatial layers using a 2:1 resolution ratio and y temporal layers. LxTyh denotes x spatial layers with a 1.5:1 resolution ratio and y temporal layers. SxTy denotes a scalabilityMode with x simulcast encodings with a 2:1 resolution ratio, with each simulcast encoding containing y temporal layers. SxTyh denotes a 1.5:1 resolution ratio. LxTy_KEY denotes a scalabilityMode with x spatial layers using a 2:1 resolution ratio and y temporal layers in which spatial layers only depend on lower spatial layers at a key frame. LxTy_KEY_SHIFT modes denotes a scalabilityMode with x spatial layers using a 2:1 resolution ratio and y temporal layers in which spatial layers only depend on lower spatial layers at a key frame and subsequent frames have their temporal identifier shifted upward.
  3. A dependency diagram MUST be supplied, in the format provided in Section 9.

7. Examples

7.1 Spatial Simulcast and Temporal Scalability

This section is non-normative.

This example extends [WEBRTC] Section 7.1 (Example 1) to demonstrate sending three spatial simulcast layers each with three temporal layers, using an SSRC and RID for each simulcast layer. Only the "sendEncodings" attribute is changed from the original example.

const signaling = new SignalingChannel(); // handles JSON.stringify/parse
const constraints = {audio: true, video: true};
const configuration = {'iceServers': [{'urls': 'stun:stun.example.org'}]};
let pc;

// call start() to initiate
async function start() {
  pc = new RTCPeerConnection(configuration);

  // let the "negotiationneeded" event trigger offer generation
  pc.onnegotiationneeded = async () => {
    try {
      await pc.setLocalDescription();
      // send the offer to the other peer
      signaling.send({description: pc.localDescription});
    } catch (err) {
      console.error(err);
    }
  };

  try {
    // get a local stream, show it in a self-view and add it to be sent
    const stream = await navigator.mediaDevices.getUserMedia(constraints);
    selfView.srcObject = stream;
    pc.addTransceiver(stream.getAudioTracks()[0], {direction: 'sendonly'});
    pc.addTransceiver(stream.getVideoTracks()[0], {
      direction: 'sendonly',
      sendEncodings: [
        {rid: 'q', scaleResolutionDownBy: 4.0, scalabilityMode: 'L1T3'}
        {rid: 'h', scaleResolutionDownBy: 2.0, scalabilityMode: 'L1T3'},
        {rid: 'f', scalabilityMode: 'L1T3'},
      ]    
    });
  } catch (err) {
    console.error(err);
  }
}

signaling.onmessage = async ({data: {description, candidate}}) => {
  try {
    if (description) {
      await pc.setRemoteDescription(description);
      // if we got an offer, we need to reply with an answer
      if (description.type == 'offer') {
        await pc.setLocalDescription();
        signaling.send({description: pc.localDescription});
      }
    } else if (candidate) {
      await pc.addIceCandidate(candidate);
    }
  } catch (err) {
    console.error(err);
  }
};

This is an example with two spatial layers (with a 2:1 ratio) and three temporal layers.

let sendEncodings = [
  {scalabilityMode: 'L2T3'}
];

This is an example with three spatial simulcast layers each with three temporal layers on a single SSRC.

let sendEncodings = [
  {scalabilityMode: 'S3T3'}
]

7.2 SVC Encoder Capabilities

This section is non-normative.

This is an example of RTCRtpSender.getCapabilities}}('video').codecs[] returned by a browser implementing [WEBRTC]. Only the scalabilityModes attribute is defined in this specification.

[
    {
      "clockRate": 90000,
      "mimeType": "video/VP8",
      "scalabilityModes": [
        "L1T1",
        "L1T2",
        "L1T3"
      ]
    },
    {
      "clockRate": 90000,
      "mimeType": "video/rtx"
    },
    {
      "clockRate": 90000,
      "mimeType": "video/VP9",
      "scalabilityModes": [
        "L1T1",
        "L1T2",
        "L1T3"
      ],
      "sdpFmtpLine": "profile-id=0"
    },
    {
      "clockRate": 90000,
      "mimeType": "video/VP9",
      "sdpFmtpLine": "profile-id=2"
    },
    {
      "clockRate": 90000,
      "mimeType": "video/H264",
      "sdpFmtpLine": "level-asymmetry-allowed=1;packetization-mode=1;profile-level-id=42001f"
    },
    {
      "clockRate": 90000,
      "mimeType": "video/H264",
      "sdpFmtpLine": "level-asymmetry-allowed=1;packetization-mode=0;profile-level-id=42001f"
    },
    {
      "clockRate": 90000,
      "mimeType": "video/H264",
      "sdpFmtpLine": "level-asymmetry-allowed=1;packetization-mode=1;profile-level-id=42e01f"
    },
    {
      "clockRate": 90000,
      "mimeType": "video/H264",
      "sdpFmtpLine": "level-asymmetry-allowed=1;packetization-mode=0;profile-level-id=42e01f"
    },
    {
      "clockRate": 90000,
      "mimeType": "video/AV1",
      "scalabilityModes": [
        "L1T1",
        "L1T2",
        "L1T3",
        "L2T1",
        "L2T1h",
        "L2T1_KEY",
        "L2T2",
        "L2T2_KEY",
        "L2T2_KEY_SHIFT",
        "L3T1",
        "L3T3",
        "L3T3_KEY",
        "S2T1"
      ]
    },
    {
      "clockRate": 90000,
      "mimeType": "video/H264",
      "sdpFmtpLine": "level-asymmetry-allowed=1;packetization-mode=1;profile-level-id=4d001f"
    },
    {
      "clockRate": 90000,
      "mimeType": "video/H264",
      "sdpFmtpLine": "level-asymmetry-allowed=1;packetization-mode=1;profile-level-id=64001f"
    },
    {
      "clockRate": 90000,
      "mimeType": "video/red"
    },
    {
      "clockRate": 90000,
      "mimeType": "video/ulpfec"
    }
]

7.3 SFM Capabilities

This section is non-normative.

This is an example of receiver capabilities returned by an SFM that only supports forwarding of VP8, VP9 and AV1 temporal scalability modes.

 "codecs": [
    {
      "clockRate": 90000,
      "mimeType": "video/VP8",
      "scalabilityModes": [
        "L1T1",
        "L1T2",
        "L1T3"
      ]
    },
    {
      "clockRate": 90000,
      "mimeType": "video/VP9",
      "scalabilityModes": [
        "L1T1",
        "L1T2",
        "L1T3",
        "L1T2h",
        "L1T3h"
      ]
    },
    {
      "clockRate": 90000,
      "mimeType": "video/AV1",
      "scalabilityModes": [
        "L1T1",
        "L1T2",
        "L1T3",
        "L1T2h",
        "L1T3h"
      ]
    }
]

8. Privacy Considerations

This section is non-normative.

This section is non-normative; it specifies no new behaviour, but instead summarizes information already present in other parts of the specification. The privacy considerations for the WebRTC APIs are described in [WEBRTC] Section 13.

8.1 Persistent information

The WebRTC API exposes information about the underlying media system via the RTCRtpSender.getCapabilities() method, including detailed and ordered information about the codecs that the system is able to produce. The WebRTC-SVC extension adds the scalabilityModes supported by the RTCRtpSender to that information, which is persistent across time, therefore increasing the fingerprint surface. Additional information is not provided relating to the RTCRtpReceiver.

Since for SVC codecs implemented in WebRTC the use of scalable coding tools is not negotiated and is independent of the supported profiles, and since SVC is rarely supported in hardware encoders, knowledge of the scalabilityModes supported by the RTCRtpSender does not provide additional information on the underlying hardware. However, since browsers may differ in their support for SVC modes, the supported scalabilityModes may permit differentiation between browsers. This additional fingerprint surface is expected to decrease over time as this specification is more widely implemented.

9. Security Considerations

This section is non-normative.

This section is non-normative; it specifies no new behaviour, but instead summarizes information already present in other parts of the specification. WebRTC protocol security considerations are described in [RTCWEB-SECURITY-ARCH] and the security and privacy considerations for the WebRTC APIs are described in [WEBRTC] Section 13.

10. Scalability Mode Dependency Diagrams

Dependency diagrams for the scability modes defined in this specification are provided below.

10.1 L1T1

L1T1: a single layer
Figure 1 L1T1: 1-layer encoding

10.2 L1T2 and L1T2h

L1T2 and L1T2h: 2-layer temporal scalability encoding
Figure 2 L1T2 and L1T2h: 1-layer spatial and 2-layer temporal scalability encoding

10.3 L1T3 and L1T3h

L1T3 and L1T3h: 3-layer temporal scalability encoding
Figure 3 L1T3 and L1T3h: 1-layer spatial and 3-layer temporal scalability encoding

10.4 L2T1 and L2T1h

L2T1 and L2T1h: 2-layer spatial and 1-layer temporal scalability encoding
Figure 4 L2T1 and L2T1h: 2-layer spatial and 1-layer temporal scalability encoding

10.5 L2T1_KEY

L2T1_KEY: 2-layer spatial and 1-layer temporal scalability K-SVC encoding
Figure 5 L2T1_KEY: 2-layer spatial and 1-layer temporal scalability K-SVC encoding

10.6 L2T2 and L2T2h

L2T2 and L2T2h: 2-layer spatial and 2-layer temporal scalability encoding
Figure 6 L2T2 and L2T2h: 2-layer spatial and 2-layer temporal scalability encoding

10.7 L2T2_KEY

L2T2_KEY: 2-layer spatial and 2-layer temporal scalability K-SVC encoding
Figure 7 L2T2_KEY: 2-layer spatial and 2-layer temporal scalability K-SVC encoding

10.8 L2T2_KEY_SHIFT

L2T2_KEY_SHIFT: 2-layer spatial and 2-layer temporal scalability K-SVC shifted encoding with temporal shift
Figure 8 L2T2_KEY_SHIFT: 2-layer spatial and 2-layer temporal scalability K-SVC encoding with temporal shift

10.9 L2T3 and L2T3h

L2T3 and L2T3h: 2-layer spatial and 3-layer temporal scalability encoding
Figure 9 L2T3 and L2T3h: 2-layer spatial and 3-layer temporal scalability encoding

10.10 L2T3_KEY

L2T3_KEY: 2-layer spatial and 3-layer temporal scalability K-SVC encoding
Figure 10 L2T3_KEY: 2-layer spatial and 3-layer temporal scalability K-SVC encoding

10.11 L2T3_KEY_SHIFT

L2T3_KEY_SHIFT: 2-layer spatial and 3-layer temporal scalability K-SVC shifted encoding with temporal shift
Figure 11 L2T3_KEY_SHIFT: 2-layer spatial and 3-layer temporal scalability K-SVC encoding with temporal shift

10.12 L3T1 and L3T1h

L3T1 and L3T1h: 3-layer spatial and 1-layer temporal scalability encoding
Figure 12 L3T1 and L3T1h: 3-layer spatial and 1-layer temporal scalability encoding

10.13 L3T1_KEY

L3T1_KEY: 3-layer spatial and 1-layer temporal scalability K-SVC encoding
Figure 13 L3T1_KEY: 3-layer spatial and 1-layer temporal scalability K-SVC encoding

10.14 L3T2 and L3T2h

L3T2 and L3T2h: 3-layer spatial and 2-layer temporal scalability encoding
Figure 14 L3T2 and L3T2h: 3-layer spatial and 2-layer temporal scalability encoding

10.15 L3T2_KEY

L3T2_KEY: 3-layer spatial and 2-layer temporal scalability K-SVC encoding
Figure 15 L3T2_KEY: 3-layer spatial and 2-layer temporal scalability K-SVC encoding

10.16 L3T2_KEY_SHIFT

L3T2_KEY_SHIFT: 3-layer spatial and 2-layer temporal scalability K-SVC with temporal shift
Figure 16 L3T2_KEY_SHIFT: 3-layer spatial and 2-layer temporal scalability K-SVC with temporal shift

10.17 L3T3 and L3T3h

L3T3 and L3T3h: 3-layer spatial and 3-layer temporal scalability encoding
Figure 17 L3T3 and L3T3h: 3-layer spatial and 3-layer temporal scalability encoding

10.18 L3T3_KEY

L3T3_KEY: 3-layer spatial and 3-layer temporal scalability K-SVC encoding
Figure 18 L3T3_KEY: 3-layer spatial and 3-layer temporal scalability K-SVC encoding

10.19 L3T3_KEY_SHIFT

L3T3_KEY_SHIFT: 3-layer spatial and 3-layer temporal scalability K-SVC with temporal shift
Figure 19 L3T3_KEY_SHIFT: 3-layer spatial and 3-layer temporal scalability K-SVC with temporal shift

10.20 S2T1 and S2T1h

S2T1 and S2T1h: 2-layer spatial simulcast encoding
Figure 20 S2T1 and S2T1h: 2-layer spatial simulcast encoding

10.21 S2T2 and S2T2h

S2T2 and S2T2h: 2-layer spatial simulcast and 2-layer temporal scalability encoding
Figure 21 S2T2 and S2T2h: 2-layer spatial simulcast and 2-layer temporal scalability encoding

10.22 S2T3 and S2T3h

S2T3 and S2T3h: 2-layer spatial simulcast and 3-layer temporal scalability encoding
Figure 22 S2T3 and S2T3h: 2-layer spatial simulcast and 3-layer temporal scalability encoding

10.23 S3T1 and S3T1h

S3T1 and S3T1h: 3-layer spatial simulcast encoding
Figure 23 S3T1 and S3T1h: 3-layer spatial simulcast encoding

10.24 S3T2 and S3T2h

S3T2 and S3T2h: 3-layer spatial simulcast and 2-layer temporal scalability encoding
Figure 24 S3T2 and S3T2h: 3-layer spatial simulcast and 2-layer temporal scalability encoding

10.25 S3T3 and S3T3h

S3T3 and S3T3h: 3-layer spatial simulcast and 3-layer temporal scalability encoding
Figure 25 S3T3 and S3T3h: 3-layer spatial simulcast and 3-layer temporal scalability encoding

A. Acknowledgements

The editors wish to thank Robin Raymond, Michael Horowitz, Harald Alvestrand, Chris Cunningham, Danil Chapovalov and Florent Castelli for their contributions to this specification, which evolved from the ORTC API developed in the W3C ORTC CG.

B. References

B.1 Normative references

[AV1-RTP]
RTP Payload Format for AV1. AV1 RTC SG. Alliance for Open Media. Standard. URL: https://aomediacodec.github.io/av1-rtp-spec/
[Media-Capabilities]
Media Capabilities. Mounir Lamouri; Chris Cunningham; Vi Nguyen. W3C. 11 January 2022. W3C Working Draft. URL: https://www.w3.org/TR/media-capabilities/
[RFC2119]
Key words for use in RFCs to Indicate Requirement Levels. S. Bradner. IETF. March 1997. Best Current Practice. URL: https://www.rfc-editor.org/rfc/rfc2119
[RFC7656]
A Taxonomy of Semantics and Mechanisms for Real-Time Transport Protocol (RTP) Sources. J. Lennox; K. Gross; S. Nandakumar; G. Salgueiro; B. Burman, Ed.. IETF. November 2015. Informational. URL: https://www.rfc-editor.org/rfc/rfc7656
[RFC7667]
RTP Topologies. M. Westerlund; S. Wenger. IETF. November 2015. RFC. URL: https://datatracker.ietf.org/doc/html/rfc7667
[RFC8174]
Ambiguity of Uppercase vs Lowercase in RFC 2119 Key Words. B. Leiba. IETF. May 2017. Best Current Practice. URL: https://www.rfc-editor.org/rfc/rfc8174
[WEBIDL]
Web IDL Standard. Edgar Chen; Timothy Gu. WHATWG. Living Standard. URL: https://webidl.spec.whatwg.org/
[WEBRTC]
WebRTC 1.0: Real-Time Communication Between Browsers. Cullen Jennings; Henrik Boström; Jan-Ivar Bruaroey. W3C. 26 January 2021. W3C Recommendation. URL: https://www.w3.org/TR/webrtc/

B.2 Informative references

[AV1]
AV1 Bitstream & Decoding Process Specification. Peter de Rivaz; Jack Haughton. AOM. 8 January 2019. Standard. URL: https://aomediacodec.github.io/av1-spec/av1-spec.pdf
[RFC6190]
RTP Payload Format for Scalable Video Coding. S. Wenger; Y.-K. Wang; T. Schierl; A. Eleftheriadis. IETF. May 2011. Proposed Standard. URL: https://www.rfc-editor.org/rfc/rfc6190
[RFC6386]
VP8 Data Format and Decoding Guide. J. Bankoski; J. Koleszar; L. Quillio; J. Salonen; P. Wilkins; Y. Xu. IETF. November 2011. Informational. URL: https://www.rfc-editor.org/rfc/rfc6386
[RFC7741]
RTP Payload Format for VP8 Video. P. Westin; H. Lundin; M. Glover; J. Uberti; F. Galligan. IETF. March 2016. Proposed Standard. URL: https://www.rfc-editor.org/rfc/rfc7741
[RTCWEB-SECURITY-ARCH]
WebRTC Security Architecture. E. Rescorla. IETF. January 2021. Proposed Standard. URL: https://www.rfc-editor.org/rfc/rfc8827
[VP9]
VP9 Bitstream & Decoding Process Specification. A. Grange; P. de Rivaz; J. Hunt. Google. February 2016. Version 0.6. URL: https://storage.googleapis.com/downloads.webmproject.org/docs/vp9/vp9-bitstream-specification-v0.6-20160331-draft.pdf
[VP9-PAYLOAD]
RTP Payload Format for VP9 Video. J. Uberti; S. Holmer; M. Flodman; J. Lennox; D. Hong. IETF. 10 June 2021. Internet Draft (work in progress). URL: https://datatracker.ietf.org/doc/html/draft-ietf-payload-vp9