Initial Author of this Specification was Ian Hickson, Google Inc., with the following copyright statement:
 © Copyright 2004-2011 Apple Computer, Inc., Mozilla Foundation, and Opera Software ASA. You are granted a license to use, reproduce and create derivative works of this document.
All subsequent changes since 26 July 2011 done by the W3C WebRTC Working Group and the Device APIs Working Group are under the following Copyright:
© 2011-2013 W3C® (MIT, ERCIM, Keio, Beihang), All Rights Reserved. Document use  rules apply.
For the entire publication on the W3C site the liability and trademark rules apply.
This document defines a set of JavaScript APIs that allow local media, including audio and video, to be requested from a platform.
This section describes the status of this document at the time of its publication. Other documents may supersede this document. A list of current W3C publications and the latest revision of this technical report can be found in the W3C technical reports index at http://www.w3.org/TR/.
This document incorporates significant changes in the design of the getUserMedia API over the past few months, including the possibility of tracking and identifying the source of a local media stream, and the possibility to select or modify the mode of operation of a given media source.
This document is not complete. While the Working Groups are trying to stabilize as much of the API as possible, it is still subject to major changes and, while early experimentations are encouraged, it is therefore not intended for widespread implementation. The API is based on preliminary work done in the WHATWG.
This document was published by the Web Real-Time Communication Working Group and Device APIs Working Group as a Working Draft. This document is intended to become a W3C Recommendation. If you wish to make comments regarding this document, please send them to public-media-capture@w3.org (subscribe, archives). All comments are welcome.
Publication as a Working Draft does not imply endorsement by the W3C Membership. This is a draft document and may be updated, replaced or obsoleted by other documents at any time. It is inappropriate to cite this document as other than work in progress.
This document was produced by a group operating under the 5 February 2004 W3C Patent Policy. W3C maintains a public list of any patent disclosures (Web Real-Time Communication Working Group, Device APIs Working Group) made in connection with the deliverables of the group; that page also includes instructions for disclosing a patent. An individual who has actual knowledge of a patent which the individual believes contains Essential Claim(s) must disclose the information in accordance with section 6 of the W3C Patent Policy.
This section is non-normative.
Access to multimedia streams (video, audio, or both) from local devices (video cameras, microphones, Web cams) can have a number of uses, such as real-time communication, recording, and surveillance.
This document defines the APIs used to get access to local devices that can generate multimedia stream data. This document also defines the stream API by which JavaScript is able to manipulate the stream data or otherwise process it.
As well as sections marked as non-normative, all authoring guidelines, diagrams, examples, and notes in this specification are non-normative. Everything else in this specification is normative.
The key words MUST, MUST NOT, REQUIRED, SHOULD, SHOULD NOT, RECOMMENDED, MAY, and OPTIONAL in this specification are to be interpreted as described in [RFC2119].
This specification defines conformance criteria that apply to a single product: the user agent that implements the interfaces that it contains.
Implementations that use ECMAScript to implement the APIs defined in this specification must implement them in a manner consistent with the ECMAScript Bindings defined in the Web IDL specification [WEBIDL], as this specification uses that specification and terminology.
            The 
            EventHandler interface represents a callback used for event
            handlers as defined in [HTML5].
          
The concepts queue a task and fires a simple event are defined in [HTML5].
The terms event handlers and event handler event types are defined in [HTML5].
A source is the "thing" providing the source of a media stream track. The source is the broadcaster of the media itself. A source can be a physical webcam, microphone, local video or audio file from the user's hard drive, network resource, or static image.
Some sources have an identifier which must be unique to the application (un-guessable by another application) and persistent between application sessions (e.g., the identifier for a given source device/application must stay the same, but not be guessable by another application). Sources that must have an identifier are camera and microphone sources; local file sources are not required to have an identifier. Source identifiers let the application save, identify the availability of, and directly request specific sources.
Other than the identifier, other bits of source identity are never directly available to the application until the user agent connects a source to a track. Once a source has been "released" to the application (either via a permissions UI, pre-configured allow-list, or some other release mechanism) the application will be able discover additional source-specific capabilities.
Sources do not have constraints -- tracks have constraints. When a source is connected to a track, it must conform to the constraints present on that track (or set of tracks).
Sources will be released (un-attached) from a track when the track is ended for any reason.
On the  object,
          sources are represented by a MediaStreamTracksourceType
          attribute. The behavior of APIs associated with the source's
          capabilities and state change depending on the source
          type.
Sources have capabilities
          and state. The capabilities and state
          are "owned" by the source and are common to any (multiple)
          tracks that happen to be using the same source (e.g., if two
          different tracks objects bound to the same source ask for
          the same capability or state information, they will get back
          the same answer).
State refers to the immediate, current value of the source's (optionally constrained) capabilities. State is always read-only.
A source's state can change dynamically over time due to environmental conditions, sink configurations, or constraint changes. A source's state must always conform to the current set of mandatory constraints that all of the tracks it is bound to have defined, and should do its best to conform to the set of optional constraints specified.
A source's state is directly exposed to audio and video track objects through individual read-only attributes. These attributes share the same name as their corresponding capabilities and constraints.
Events are available that signal to the application that source state has changed.
A conforming user-agent must support all the state names defined in this spec.
Source capabilities are the intrinsic "features" of a source object. For each source state, there is a corresponding capability that describes whether it is supported by the source and if so, what the range of supported values are. Capabilities are expressed as either a series of states (for enumerated-type capabilities) or as a min/max range.
The values of the supported capabilities must be normalized to the ranges and enumerated types defined in this specification.
Capabilities return the same underlying per-source capabilities, regardless of any user-supplied constraints present on the source (capabilities are independent of constraints).
Source capabilities are effectively constant. Applications should be able to depend on a specific source having the same capabilities for any session.
Constraints are an optional feature for restricting the range of allowed variability on a source. Without provided constraints, implementations are free to select a source's state from the full range of its supported capabilities, and to adjust that state at any time for any reason.
Constraints may be optional or mandatory. Optional constraints are represented by an ordered list, mandatory constraints are an unordered set. The order of the optional constraints is from most important (at the head of the list) to least important (at the tail of the list).
Constraints are stored on the track object, not the source. Each track can be optionally initialized with constraints, or constraints can be added afterward through the constraint APIs defined in this spec.
Applying track level constraints to a source is conditional based on the type of source. For example, read-only sources will ignore any specified constraints on the track.
It is possible for two tracks that share a unique source to apply contradictory constraints. Under such contradictions, the implementation will mute both tracks and notify them that they are over-constrained.
Events are available that allow the application to know when constraints cannot be met by the user agent. These typically occur when the application applies constraints beyond the capability of a source, contradictory constraints, or in some cases when a source cannot sustain itself in over-constrained scenarios (overheating, etc.).
Constraints that are intended for video sources will be ignored by audio sources and vice-versa. Similarly, constraints that are not recognized will be preserved in the constraint structure, but ignored by the application. This will allow future constraints to be defined in a backward compatible manner.
A correspondingly-named constraint exists for each corresponding source state name and capability name. In general, user agents will have more flexibility to optimize the media streaming experience the fewer constraints are applied.
The  interface is used to represent
      streams of media data, typically (but not necessarily) of audio and/or
      video content, e.g. from a local camera. The data from a
      MediaStream object does not necessarily have a
      canonical binary form; for example, it could just be "the video currently
      coming from the user’s video camera". This allows user agents to
      manipulate media streams in whatever fashion is most suitable on the
      user’s platform.MediaStream
Each  object can contain zero or more
      tracks, in particular audio and video tracks. All tracks in a MediaStream
      are intended to be synchronized when rendered. Different MediaStreams do
      not need to be synchronized.MediaStream
Each track in a MediaStream object has a corresponding
       object.MediaStreamTrack
A  represents content comprising
      one or more channels, where the channels have a defined well known
      relationship to each other (such as a stereo or 5.1 audio signal).MediaStreamTrack
A channel is the smallest unit considered in this API specification.
A  object has an input and an output.
      The input depends on how the object was created: a
      MediaStream object generated by a MediaStreamgetUserMedia() call (which is
      described later in this document), for instance, might take its input
      from the user’s local camera. The output of the object controls how the
      object is used, e.g., what is saved if the object is written to a file or
      what is displayed if the object is used in a video
      element.
Each track in a  object can be
      disabled, meaning that it is muted in the object’s output. All tracks are
      initially enabled.MediaStream
A  can be finished, indicating
      that its inputs have forever stopped providing data.MediaStream
The output of a  object MUST correspond
      to the tracks in its input. Muted audio tracks MUST be replaced with
      silence. Muted video tracks MUST be replaced with blackness.MediaStream
A new  object can be created from
      accessible media sources (that does not require any additional
      permissions) using the MediaStreamMediaStream()
       constructor. The constructor argument can either be an existing
       object, in which case all the tracks of
      the given stream are added to the new MediaStream
      object, or an array of MediaStream objects. The
      latter from makes it possible to compose a stream from different source
      streams.MediaStreamTrack

Both  and MediaStream
       objects can be cloned. This allows for
      greater control since the separate instances can be manipulated and
      consumed individually. A cloned MediaStreamTrack
       contains clones of all member tracks from the
      original stream.MediaStream
When a  object is being generated
      from a local file (as opposed to a live audio/video source), the user
      agent SHOULD stream the data from the file in real time, not all at once.
      The MediaStreamMediaStream object is also used in contexts outside
      getUserMedia, such as [WEBRTC10]. In both cases, ensuring
      a realtime stream reduces the ease with which pages can distinguish live
      video from pre-recorded video, which can help protect the user’s
      privacy.
The  MediaStream()
      constructor composes a new stream out of existing tracks. It takes zero
      or one argument. If the argument is supplied, it
      MUST either be of type  or an array of
      MediaStream objects. When the constructor is
      invoked, the UA must run the following steps:MediaStreamTrack
Let stream be a newly constructed 
           object.MediaStream
Initialize stream’s id attribute to a newly generated
          value.
If the constructor’s argument is present, run the sub steps that corresponds to the argument type.
Array of 
              objects:MediaStreamTrack
Run the following sub steps for each 
               in the array:MediaStreamTrack
Add track: Let track be the 
                    about to be processed.MediaStreamTrack
If track has ended or if there is already
                  a  contained within
                  stream that has the same underlying source as
                  track, then abort these steps and continue with the
                  next track (if any).MediaStreamTrack
Add track to stream's track set.
Run the sub steps labeled Add track (above) for every
               in the argument stream’s
              track set.
              MediaStreamTrack
Return stream.
A  can have multiple audio and video
      sources (e.g. because the user has multiple microphones, or because the
      real source of the stream is a media resource with many media tracks).
      The stream represented by a MediaStream thus has zero
      or more tracks.MediaStream
The tracks of a  are stored in a
      track set. The track set MUST contain the
      MediaStream objects that correspond
      to the tracks of the stream. The relative order of the tracks in the set
      is user agent defined and the API will never put any requirements on the
      order. The proper way to find a specific MediaStreamTrack object in the set is to look it up by its
      MediaStreamTrack
      id.
An object that reads data from the output of a
       is referred to as a
      MediaStream consumer. The list of
      MediaStream consumers currently includes the media
      elements and the MediaStreamPeerConnection API specified in
      [WEBRTC10].
 consumers must be able to
      handle tracks being added and removed. This behavior is specified per
      consumer.MediaStream
A  object is said to be
      finished when all tracks belonging to the stream have
      ended. When this happens for any reason other than the
      MediaStreamstop() method being
      invoked, the user agent MUST queue a task that runs the following
      steps:
If the object’s ended attribute has the value
          true already, then abort these steps. (The stop() method was probably called
          just before the stream stopped for other reasons, e.g. the user
          clicked an in-page stop button and then the user agent provided stop
          button.)
Set the object’s ended attribute to true.
Fire a simple event named ended at the object.
If the end of the stream was reached due to a user request, the task source for this task is the user interaction task source. Otherwise the task source for this task is the networking task source.
typedef sequence<MediaStreamTrack> MediaStreamTrackSequence;MediaStreamTrack> type.[Constructor,
        Constructor (MediaStream stream),
        Constructor (MediaStreamTrackSequence tracks)]
interface MediaStream : EventTarget {
    readonly    attribute DOMString    id;
    sequence<MediaStreamTrack> getAudioTracks ();
    sequence<MediaStreamTrack> getVideoTracks ();
    MediaStreamTrack?          getTrackById (DOMString trackId);
    void                       addTrack (MediaStreamTrack track);
    void                       removeTrack (MediaStreamTrack track);
    MediaStream                clone ();
                attribute boolean      ended;
                attribute EventHandler onended;
                attribute EventHandler onaddtrack;
                attribute EventHandler onremovetrack;
};ended of type boolean,            The MediaStream.ended
          attribute MUST return true if the  has
          finished, and false otherwise.MediaStream
When a  object is created, its
          MediaStreamended attribute
          MUST be set to false, unless it is being created using the
          MediaStream() constructor
          whose arguments are lists of 
          objects that are all ended, in which case the
          MediaStreamTrack object MUST be created with its
          MediaStreamended attribute set
          to true.
id of type DOMString, readonly   When a  object is created, the
          user agent MUST generate a globally unique identifier string, and
          MUST initialize the object’s MediaStreamid attribute to that string.
          Such strings MUST only use characters in the ranges U+0021, U+0023 to
          U+0027, U+002A to U+002B, U+002D to U+002E, U+0030 to U+0039, U+0041
          to U+005A, U+005E to U+007E, and MUST be 36 characters long.
The id
          attribute MUST return the value to which it was initialized when the
          object was created.
onaddtrack of type EventHandler,            addtrack, MUST be supported by
        all objects implementing the MediaStream interface.
        onended of type EventHandler,            ended, MUST be supported by all
        objects implementing the MediaStream
        interface.onremovetrack of type EventHandler,            removetrack, MUST be
        supported by all objects implementing the MediaStream
         interface.addTrackAdds the given  to this
          MediaStreamTrack.MediaStream
When the addTrack() method is
          invoked, the user agent MUST run the following steps:
Let track be the  argument and stream this MediaStreamTrack
              
              object.MediaStream
If stream is finished, throw an
              INVALID_STATE_ERR exception.
If track is already in stream’s track set, then abort these steps.
Add track to stream's track set.
| Parameter | Type | Nullable | Optional | Description | 
|---|---|---|---|---|
| track |  | ✘ | ✘ | 
voidcloneClones the given  and all its
            tracks.MediaStream
When the clone() method
            is invoked, the user agent MUST run the following steps:
Let streamClone be a newly constructed 
                 object.MediaStream
Initialize streamClone's id attribute to a newly
                generated value.
Let trackSetClone be a list that contains the
                result of running
                MediaStreamTrack.clone() on all the tracks in this
                stream.
Let trackSetClone be streamClone's track set.
MediaStreamgetAudioTracksReturns a sequence of  objects
          representing the audio tracks in this stream.MediaStreamTrack
The getAudioTracks()
          method MUST return a sequence that represents a snapshot of all the
           objects in this stream’s
          track set whose MediaStreamTrack
          kind is equal to
          "audio". The conversion from the track set to the sequence is user agent defined and
          the order does not have to stable between calls.
sequence<MediaStreamTrack>getTrackByIdThe getTrackById()
          method MUST return the first 
          object in this stream’s track set whose
          MediaStreamTrackid is equal to
          trackId. The method MUST return null if no track matches
          the trackId argument.
| Parameter | Type | Nullable | Optional | Description | 
|---|---|---|---|---|
| trackId | DOMString | ✘ | ✘ | 
MediaStreamTrack, nullablegetVideoTracksReturns a sequence of  objects
          representing the video tracks in this stream.MediaStreamTrack
The getVideoTracks()
          method MUST return a sequence that represents a snapshot of all the
           objects in this stream’s
          track set whose MediaStreamTrack
          kind is equal to
          "video". The conversion from the track set to the sequence is user agent defined and
          the order does not have to stable between calls.
sequence<MediaStreamTrack>removeTrackRemoves the given  from this
          MediaStreamTrack.MediaStream
When the removeTrack() method
          is invoked, the user agent MUST run the following steps:
Let track be the  argument and stream this MediaStreamTrack
              
              object.MediaStream
If stream is finished, throw an
              INVALID_STATE_ERR exception.
If track is in stream's track set, remove it.
| Parameter | Type | Nullable | Optional | Description | 
|---|---|---|---|---|
| track |  | ✘ | ✘ | 
voidA  object represents a media
      source in the user agent. Several MediaStreamTrack
      objects can represent the same media source, e.g., when the user chooses
      the same camera in the UI shown by two consecutive calls to
      MediaStreamTrackgetUserMedia()
      .
Note that a web application can revoke all given permissions
      with MediaStreamTrack.stop().
A  has three stages in its
        lifecycle. A track begins as MediaStreamTracknew prior to being connected to
        an active source.
Once connected, the started event fires and the track
        becomes live. In the live state, the track is
        active and media is available for rendering at a 
         consumer.MediaStream
A muted or disabled  renders either
        silence (audio), black frames (video), or a zero-information-content
        equivalent. The muted/unmuted state of a track reflects if the source
        provides any media at this moment. The enabled/disabled state is under
        application control and determines if the track outputs media (to its
        consumers). Hence, media from the source only flows when a MediaStreamTrack
         object is both unmuted and enabled.MediaStreamTrack
A  is muted when the source is temporarily unable to
        provide the track with data. A track can be muted by a user. Often this
        action is outside the control of the application. This could be as a
        result of the user hitting a hardware switch, or toggling a control in
        the operating system or browser chrome. A track can also be muted by the
        user agent. For example, a track that is a member of a MediaStreamTrack
        , received via a MediaStreamRTCPeerConnection
         [WEBRTC10], is muted if the application on the other side
        disables the corresponding track in the MediaStream being
        sent.
Applications are able to enable or
        disable a  to prevent it from
        rendering media from the source. A muted track will however, regardless
        of the enabled state, render silence and blackness. A disabled track is
        logically equivalent to a muted track, from a consumer point of view.MediaStreamTrack
A  object is said to end
        when the source of the track is disconnected or exhausted.MediaStreamTrack
When a  object ends for any reason
        (e.g., because the user rescinds the permission for the page to use the
        local camera, or because the data comes from a finite file and the file’s
        end has been reached and the user has not requested that it be looped, or
        because the UA has instructed the track to end for any reason, or because
        the reference count of the track’s underlying media source has reached
        zero, it is said to be ended. When track instance
        track ends for any reason other than the MediaStreamTrackstop() method being invoked on the
         object that represents
        track, the user agent MUST queue a task that runs the
        following steps:MediaStreamTrack
If the track’s readyState attribute
            has the value ended already, then abort these steps.
Set track’s readyState attribute to
            ended.
Fire a simple event named ended at the object.
If the end of the stream was reached due to a user request, the event source for this event is the user interaction event source.
Constraints are independent of sources.  However, if
        the sourceType is "none",
        the readonly attribute
        is true, or the remote
        attribute is true, the track's constraints will
        not be applied to the source.
Whether  were provided at
        track initialization time or need to be established later at
        runtime, the APIs defined below allow the retrieval and
        manipulation of the constraints currently established on a
        track.MediaTrackConstraints
Each track maintains an internal version of
        the  structure, namely a
        mandatory set of constraints (no duplicates), and an optional
        ordered list of individual constraint objects (may contain
        duplicates). The internal stored constraint structure is only
        exposed to the application using the existing
        MediaTrackConstraints,
        MediaTrackConstraintsMediaTrackConstraintSet, MediaTrackConstraint,
        and similarly-derived-type dictionary objects.
When track constraints change, a user agent must queue a task to evaluate those changes when the task queue is next serviced. Similarly, if the sourceType changes, then the user agent should perform the same actions to re-evaluate the constraints of each track affected by that source change.
interface MediaStreamTrack : EventTarget {
    readonly    attribute DOMString             kind;
    readonly    attribute DOMString             id;
    readonly    attribute DOMString             label;
                attribute boolean               enabled;
    readonly    attribute boolean               muted;
                attribute EventHandler          onmute;
                attribute EventHandler          onunmute;
    readonly    attribute boolean               _readonly;
    readonly    attribute boolean               remote;
    readonly    attribute MediaStreamTrackState readyState;
                attribute EventHandler          onstarted;
                attribute EventHandler          onended;
    MediaTrackConstraints?                        constraints ();
    MediaSourceStates                             states ();
    (AllVideoCapabilites or AllAudioCapabilities) capabilities ();
    void                                          applyConstraints (MediaTrackConstraints constraints);
                attribute EventHandler          onoverconstrained;
    MediaStreamTrack                              clone ();
    void                                          stop ();
};enabled of type boolean,            The MediaStreamTrack.enabled
            attribute, on getting, MUST return the last value to which it was
            set. On setting, it MUST be set to the new value, and then, if the
             object is still associated with
            a track, MUST enable the track if the new value is true, and disable
            it otherwise.MediaStreamTrack
Thus, after a  is
            disassociated from its track, its MediaStreamTrackenabled attribute still
            changes value when set; it just doesn’t do anything with that new
            value.
id of type DOMString, readonly   Unless a  object is created as
            a part a of special purpose algorithm that specifies how the track id
            must be initialized, the user agent MUST generate a globally unique
            identifier string and initialize the object’s MediaStreamTrackid attribute to that string.
An example of an algorithm that specifies how the track id must be
            initialized is the algorithm to represent an incoming network
            component with a  object.
            [WEBRTC10]MediaStreamTrack
            MediaStreamTrack.id attribute MUST return the value
            to which it was initialized when the object was created.
kind of type DOMString, readonly   The MediaStreamTrack.kind
            attribute MUST return the string "audio" if the object
            represents an audio track or "video" if object represents
            a video track.
label of type DOMString, readonly   User agents MAY label audio and video sources (e.g., "Internal
            microphone" or "External USB Webcam"). The MediaStreamTrack.label
            attribute MUST return the label of the object’s corresponding track,
            if any. If the corresponding track has or had no label, the attribute
            MUST instead return the empty string.
Thus the kind and label attributes do not
            change value, even if the  object
            is disassociated from its corresponding track.MediaStreamTrack
muted of type boolean, readonly   The MediaStreamTrack.muted
            attribute MUST return true if the track is muted, and false otherwise.
onended of type EventHandler,            ended, MUST be supported by
          all objects implementing the MediaStreamTrack
          interface.onmute of type EventHandler,            mute, MUST be supported by
          all objects implementing the MediaStreamTrack
          interface.onoverconstrained of type EventHandler,            overconstrained, MUST be supported by
          all objects implementing the MediaStreamTrack
          interface.onstarted of type EventHandler,            started, MUST be supported by
          all objects implementing the MediaStreamTrack
          interface.onunmute of type EventHandler,            unmute, MUST be supported by
          all objects implementing the MediaStreamTrack
          interface.readonly of type boolean, readonly   readonly
            attribute MUST return the value true.
            Otherwise, it must return the value false.
          readyState of type MediaStreamTrackState, readonly   The readyState
            attribute represents the state of the track. It MUST return the value
            to which the user agent last set it.
remote of type boolean, readonly   RTCPeerConnection, the remote
            attribute MUST return the value true.
            Otherwise, it must return the value false.
          applyConstraintsThis API will replace all existing constraints with the provided constraints (if existing constraints exist). Otherwise, it will apply the newly provided constraints to the track.
| Parameter | Type | Nullable | Optional | Description | 
|---|---|---|---|---|
| constraints |  | ✘ | ✘ | A new constraint structure to apply to this track. | 
voidcapabilitiesReturns a dictionary with all of the capabilities for the
            track type. If the track type is ,
            the VideoStreamTrackAllVideoCapabilities
            dictionary is returned. If the track type
            is ,
            the AudioStreamTrack
            dictionary is returned.AllAudioCapabilities
Given that implementations of various hardware may not
            exactly map to the same range, an
            implementation should make a reasonable attempt to
            translate and scale the hardware's setting onto the mapping
            provided by this specification. If this is not possible due
            to the user agent's inability to retrieve a given
            capability from a source, then for CapabilityRange-typed
            capabilities, the min and max
            fields will not be present on the returned dictionary, and
            the supported field will
            be false. For -typed
            capabilities, a suitable CapabilityList"notavailable" value
            will be the sole capability in the list.
An example of the user agent providing an alternative mapping: if a source supports a hypothetical fluxCapacitance state whose type is a CapabilityRange, and the state is defined in this specification to be the range from -10 (min) to 10 (max), but the source's (hardware setting) for fluxCapacitance only supports values of "off" "medium" and "full", then the user agent should map the range value of -10 to "off", 10 should map to "full", and 0 should map to "medium". Constraints imposing a strict value of 3 will cause the user agent to attempt to set the value of "medium" on the hardware, and return a fluxCapacitance state of 0, the closest supported setting. No error event is raised in this scenario.
 objects should order their enumerated
            values from minimum to maximum where it makes sense, or in
            the order defined by the enumerated type where
            applicable.CapabilityList
See the 
            and AllVideoCapabilities
            dictionaries for details on the expected types for the various
            supported state names.AllAudioCapabilities
(AllVideoCapabilites or AllAudioCapabilities)cloneClones the given .MediaStreamTrack
When the clone() method
            is invoked, the user agent MUST run the following steps:
Let trackClone be a newly constructed 
                 object.MediaStreamTrack
Initialize trackClone's id attribute to a newly
                generated value.
Let trackClone inherit this track's underlying
                source, 
                kind and 
                label
                attributes.
Return trackClone.
MediaStreamTrackconstraintsmandatory field will not be present (it will be
            undefined). If no optional constraints have been defined,
            the optional field will not be present (it will be
            undefined). If neither optional, nor mandatory constraints have been
            created, the value null is returned.
          MediaTrackConstraints, nullablestatesMediaSourceStatesstopWhen a 
            object’s MediaStreamTrackstop() method
            is invoked, if no source is attached
            (e.g., sourceType is "none"), then this
            call returns immediately (e.g., is a no-op).  Otherwise, the
            user agent MUST queue a task that runs the following steps:
Let track be the current
                 object.MediaStreamTrack
End track. The track starts outputting only silence and/or blackness, as appropriate.
Permanently stop the generation of data for track's source. If the data is being generated from a live source (e.g., a microphone or camera), then the user agent SHOULD remove any active "on-air" indicator for that source. If the data is being generated from a prerecorded source (e.g. a video file), any remaining content in the file is ignored.
The task source for the tasks
            queued for the stop() method is the DOM
            manipulation task source.
voidenum MediaStreamTrackState {
    "new",
    "live",
    "ended"
};| Enumeration description | |
|---|---|
new | The track type is new and has not been initialized (connected to a source of any kind). This state implies that the track's label will be the empty string. | 
live | 
             The track is active (the track’s underlying media source is making a best-effort attempt to provide data in real time). The output of a track in the   | 
ended | 
             The track has ended (the track's underlying media source is no longer providing data, and will never provide more data for this track). Once a track enters this state, it never exits it. For example, a video track in a
              | 
enum SourceTypeEnum {
    "none",
    "camera",
    "microphone"
};| Enumeration description | |
|---|---|
none | This track has no source. This is the case when the track is in the "new" or "ended" readyState. | 
camera | A valid source type only for s. The source is a local video-producing camera source. | 
microphone | A valid source type only for s. The source is a local audio-producing microphone source. | 
The addtrack and
      removetrack events
      use the  interface.MediaStreamTrackEvent
Firing a track event named
      e with a 
      track means that an event with the name e, which
      does not bubble (except where otherwise stated) and is not cancelable
      (except where otherwise stated), and which uses the
      MediaStreamTrack interface with the
      MediaStreamTrackEventtrack
      attribute set to track, MUST be created and dispatched at the
      given target.
dictionary MediaStreamTrackEventInit : EventInit {
    MediaStreamTrack? track;
};
[Constructor(DOMString type, MediaStreamTrackEventInit eventInitDict)]
interface MediaStreamTrackEvent : Event {
    readonly    attribute MediaStreamTrack track;
};track of type MediaStreamTrack, readonly   The track attribute
          represents the  object associated
          with the event.MediaStreamTrack
MediaStreamTrackEventInit Memberstrack of type readonly MediaStreamTrack, nullable
The  object cannot be
      instantiated directly. To create an instance of
      a MediaStreamTrack, one of its derived track
      types may be instantiated. These derived types are defined in
      this section.MediaStreamTrack
Note that the camera's green light
 doesn't come on
      when a new track is created; nor does the user get prompted to
      enable the camera/microphone. Those actions only happen after
      the developer has requested that a media stream
      containing "new" tracks be bound to a source
      via getUserMedia(). Until that point tracks
      are inert.
Video tracks may be instantiated with optional media track constraints. These constraints can be later modified on the track as needed by the application, or created after-the-fact if the initial constraints are unknown to the application.
Example:  
          objects are instantiated in JavaScript using the new
          operator:VideoStreamTrack
new VideoStreamTrack();
          or
          newVideoStreamTrack( { optional: [ {sourceId: "20983-20o198-109283-098-09812" }, {width: { min: 800, max: 1200 }}, {height: { min: 600 }}] });
[Constructor(optional MediaTrackConstraints videoConstraints)]
interface VideoStreamTrack : MediaStreamTrack {
    static sequence<DOMString> getSourceIds ();
};getSourceIds, staticReturns an array of application-unique source
            identifiers. This list will be populated only with local
            sources whose sourceType
            is "camera"
            and if allowed by the user-agent, "readonly"
            variants of the former two types. The video source ids
            returned in the list constitute those sources that the
            user agent can identify at the time the API is called (the
            list can grow/shrink over time as sources may be added or
            removed). As a static
            method, getSourceIds can be queried
            without instantiating
            any  objects or
            without calling VideoStreamTrackgetUserMedia().
Issue:  This information
              deliberately adds to the fingerprinting surface of the
              UA. However, this information will not be identifiable
              outside the scope of this application and could also be
              obtained via other round-about techniques
              using getUserMedia().
sequence<DOMString>Example:  objects are instantiated in JavaScript using the new operator:AudioStreamTrack
new AudioStreamTrack();
              or
              newAudioStreamTrack( { optional: [ {sourceId: "64815-wi3c89-1839dk-x82-392aa" }, {gain: 0.5 }] });
[Constructor(optional MediaTrackConstraints audioConstraints)]
interface AudioStreamTrack : MediaStreamTrack {
    static sequence<DOMString> getSourceIds ();
};getSourceIds, staticgetSourceIds on
            the VideoStreamTrack object. Note that
            the list of source ids
            for AudioStreamTrack will be populated
            only with local sources
            whose sourceType
            is "microphone" and if allowed by the
            user-agent, "readonly" microphone
            variants.sequence<DOMString>There is a variable associated with each capability that
      represents the state of the source with respect to that
      capability, the actual setting in use by the state.  In the same
      way that the current set of constraints can be returned on a
      track using the constraints() method,
      the states() method on a track returns the
      values of the state variables associated with all capabilities,
      as the  object.  Only
      states appropriate to the MediaSourceStatessourceType are
      returned.
Note: The following specific list(s) of states DOES NOT REFLECT CONSENSUS. Many states beyond these have been proposed, and the ones listed do not have universal support. The ones below are included **** ONLY **** to provide a starting point so we can see concrete examples of what real states might look like. The particular set below was chosen to match the **also temporary** set of constraints.
dictionary MediaSourceStates {
    SourceTypeEnum       sourceType;
    DOMString            sourceId;
    unsigned long?       width;
    unsigned long?       height;
    float?               frameRate;
    VideoFacingModeEnum? facingMode;
    unsigned long?       volume;
};MediaSourceStates MembersfacingMode of type VideoFacingModeEnum, nullableframeRate of type float, nullableheight of type unsigned long, nullablesourceId of type DOMStringsourceType of type SourceTypeEnumvolume of type unsigned long, nullablewidth of type unsigned long, nullableEach constraint that is supported by an implementation MUST
      have an associated capability that will be returned in the
      result of a call to the capabilities()
      method.
Capabilities are provided as either a min/max range or a list of enumerated values. Min/max capabilities are always provided for constraints that are not enumerated types. Listed capabilities are always provided for constraints corresponding to enumerated types.
dictionary CapabilityRange {
    any     max;
    any     min;
    boolean supported;
};CapabilityRange Membersmax of type anyThe type of this value is specific to the capability.
If the related capability is not supported by the source,
          then this field will not be provided by the user agent (it
          will be undefined).
min of type anyThe type of this value is specific to the capability.
If the related capability is not supported by the source,
          then this field will not be provided by the user agent (it
          will be undefined).
supported of type booleantrue if the capability is
        supported, false otherwise.typedef sequence<DOMString> CapabilityList;Capability Lists are just an array of
        supported DOMString values from the possible
        superset of values described by each
        state's enumerated type.
Note: The following specific list(s) of capabilities DOES NOT REFLECT CONSENSUS. Many capabilities beyond these have been proposed, and the ones listed do not have universal support. The ones below are included **** ONLY **** to provide a starting point so we can see concrete examples of what real capabilities might look like. The particular sets below were chosen to match the **also temporary** set of constraints.
dictionary AllVideoCapabilities {
    CapabilityList?  sourceType;
    CapabilityList?  sourceId;
    CapabilityRange? width;
    CapabilityRange? height;
    CapabilityRange? frameRate;
    CapabilityList?  facingMode;
};AllVideoCapabilities MembersfacingMode of type CapabilityList, nullableVideoFacingModeEnum) on the source.frameRate of type CapabilityRange, nullableheight of type CapabilityRange, nullablesourceId of type CapabilityList, nullablegetSourceIds() method.sourceType of type CapabilityList, nullableSourceTypeEnum) on the current source.width of type CapabilityRange, nullabledictionary AllAudioCapabilities {
    CapabilityRange? volume;
};AllAudioCapabilities Membersvolume of type CapabilityRange, nullablepartial interface URL {
    static DOMString createObjectURL (MediaStream stream);
};createObjectURL, staticMints a Blob URL to refer to the given
          .MediaStream
When the createObjectURL() method
          is called with a  argument, the user
          agent MUST return a unique Blob URL for the
          given MediaStream. [FILE-API]MediaStream
For audio and video streams, the data exposed on that stream MUST
          be in a format supported by the user agent for use in
          audio and video elements.
A Blob URL is the
          same as what the File API specification calls a Blob URI, except that
          anything in the definition of that feature that refers to
          File and Blob objects is hereby extended to
          also apply to  objects.MediaStream
| Parameter | Type | Nullable | Optional | Description | 
|---|---|---|---|---|
| stream |  | ✘ | ✘ | 
DOMStringA MediaStream may be assigned to media elements as defined in HTML5 [HTML5]
      by calling createObjectURL to obtain a URL for the MediaStream and then
      setting the media elements src attribute to that URL. A MediaStream is
      not preloadable or seekable and represents a simple, potentially infinite,
      linear media timeline. The timeline starts at 0 and increments linearly
      in real time as long as the MediaStream is playing. The timeline does not
      increment when the MediaStream is paused.
Do we also need to support direct assignment and access of the underlying stream?
The nature of the MediaStream places certain restrictions on the
      behavior and attribute values of the associated media element and on the
      operations that can be performed on it, as shown below:
media.readystate to HAVE_ENOUGH_DATA.
        | Attribute Name | Attribute Type | Valid Values | Additional considerations | 
|---|---|---|---|
              src
             | 
            DOMString | 
            a local URI referencing a MediaStream | N.B. Revocation of the URI does not count as a change to this field. | 
              currentSrc
             | 
            DOMString | 
            a local URI referencing a MediaStream | - | 
              preload
             | 
            DOMString | 
            none | 
            A MediaStream cannot be preloaded. | 
              buffered
             | 
            
              TimeRanges
             | 
            buffered.length MUST return 1.buffered.start(0) MUST return 0.buffered.end(0) MUST return 0. | 
            A MediaStream cannot be preloaded. Therefore, the amount buffered is always an empty TimeRange. | 
              currentTime
             | 
            double | 
            Any positive integer. The initial value is 0 and the values increments linearly in real time whenever the stream is playing. | The value is the current stream position, in seconds. The UA MUST ignore attempts to set this attribute. | 
              duration
             | 
            double | 
            Infinity | 
              A MediaStream does not have a pre-defined duration.
               If the underlying MediaStream is destroyed, the UA MUST set this property to the value
              of the last known   | 
          
              seeking
             | 
            boolean | 
            false | A MediaStream is not seekable. Therefore, this attribute
            MUST always have the value
            false. | 
          
              defaultPlaybackRate
             | 
            double | 
            1.0 | A MediaStream is not seekable. Therefore, this attribute
            MUST always have the value
            1.0 and any attempt to alter it MUST fail. | 
          
              playbackRate
             | 
            double | 
            1.0 | A MediaStream is not seekable. Therefore, this attribute
            MUST always have the value
            1.0 and any attempt to alter it MUST fail. | 
          
              played
             | 
            
              TimeRanges
             | 
            
              played.length MUST return 1.played.start(0) MUST return 0.played.end(0) MUST return the last known currentTime.
             | 
            A MediaStream's timeline always consists of a single range, starting at 0 and extending up to the currentTime. | 
              seekable
             | 
            
              TimeRanges
             | 
            
              seekable.length MUST return 0.seekable.start() MUST return currentTime.seekable.end() MUST return currentTime.
             | 
            A MediaStream is not seekable. | 
              startDate
             | 
            Date | 
            Not-a-Number (NaN) | A MediaStream does not specify a timeline offset. | 
              loop
             | 
            boolean | 
            false | A MediaStream has no defined end and therefore cannot be looped. | 
This section is non-normative.
The following event fires on 
           objects:MediaStream
        
| Event name | Interface | Fired when... | 
|---|---|---|
              
                ended
              
             | 
            
              Event
             | 
            The 
                 finished as a result of all tracks in the 
                 ending. | 
          
              addtrack
             | 
            
              
                
             | 
            A new   has been added to
            this stream. Note that this event is not fired when the script
            directly modifies the tracks of a .
           | 
          
              removetrack
              
             | 
            
              
                
             | 
            A  has been removed from
            this stream. Note that this event is not fired when the script
            directly modifies the tracks of a .
             | 
          
The following event fires on 
           objects:MediaStreamTrack
        
| Event name | Interface | Fired when... | 
|---|---|---|
              
                started
              
             | 
            
              Event
             | 
            The 
              object has just transitioned from the
              "new" readyState to another state.
              This event fires before any other corresponding events
              such as "ended" or "statechanged". | 
          
              
                mute
              
             | 
            
              Event
             | 
            The 
                 object's source is temporarily unable to provide
            data. | 
          
              
                unmute
              
             | 
            
              Event
             | 
            The 
                 object's source is live again after having been
            temporarily unable to provide data. | 
          
              
                overconstrained
              
             | 
            
              Event
             | 
            
               This event fires asynchronously for each affected
              track (when multiple tracks share the same source) after
              the user agent has evaluated the current constraints
              against a given  Due to being over-constrained, the user agent must mute each affected track. The affected track(s) will remain un-usable (in
              the  The "overconstrained" event is a simple event of
              type   | 
          
              
                ended
              
             | 
            
              Event
             | 
            The 
                 object's source will no longer provide any data, either
            because the user revoked the permissions, or because the source
            device has been ejected, or because the remote peer stopped
            sending data, or because the 
                stop()
               method was invoked. | 
          
dictionary MediaStreamConstraints {
    (boolean or MediaTrackConstraints) video = false;
    (boolean or MediaTrackConstraints) audio = false;
};MediaStreamConstraints Membersaudio of type (boolean or MediaTrackConstraints), defaulting to falseProvide definition of audio constraints here.
video of type (boolean or MediaTrackConstraints), defaulting to falseProvide definition of video constraints here.
dictionary MediaTrackConstraints {
    MediaTrackConstraintSet? mandatory = null;
    MediaTrackConstraint[]?  _optional = null;
};MediaTrackConstraints Membersmandatory of type MediaTrackConstraintSet, nullable, defaulting to nullProvide definition of mandatory constraints here.
optional of type array of MediaTrackConstraint, nullable, defaulting to nullProvide definition of optional constraints here.
A MediaTrackConstraintSet is a dictionary containing one or more key-value pairs, where each key MUST be a valid registered constraint name in the IANA-hosted RTCWeb Media Constraints registry [RTCWEB-CONSTRAINTS] and its value SHOULD be as defined in the associated reference(s) given in the registry.
A MediaTrackConstraint is a dictionary containing exactly one key-value pair, where the key MUST be a valid registered constraint name in the IANA-hosted RTCWeb Media Constraints registry [RTCWEB-CONSTRAINTS] and the value SHOULD be as defined in the associated reference(s) given in the registry.
Example MediaTrackConstraints value:
{
  mandatory: {
    width: { min: 640 },
    height: { min: 480 }
  },
  optional: [
    { width: 650 },
    { width: { min: 650 }},
    { frameRate: 60 },
    { width: { max: 800 }},
    { facingMode: "user" }
  ]
}                                
      This section is non-normative.
The user agent is encouraged to reserve resources when it has determined that a given call to getUserMedia() will succeed. It is preferable to reserve the resource prior to invoking the success callback provided by the web page. Subsequent calls to getUserMedia() (in this page or any other) should treat the resource that was previously allocated, as well as resources held by other applications, as busy. Resources marked as busy should not be provided as sources to the current web page, unless specified by the user. Optionally, the user agent may choose to provide a stream sourced from a busy source but only to a page whose origin matches the owner of the original stream that is keeping the source busy.
This document recommends that in the permission grant dialog or device selection interace (if one is present), the user be allowed to select any available hardware as a source for the stream requested by the page (provided the resource is able to fulfill mandatory constraints, if any were specified), in addition to the ability to substitute a video or audio source with local files and other media. A file picker may be used to provide this functionality to the user.
This document also recommends that the user be shown all resources that are currently busy as a result of prior calls to getUserMedia() (in this page or any other page that is still alive) and be allowed to terminate that stream and utilize the resource for the current page instead. If possible in the current operating environment, it is also suggested that resources currently held by other applications be presented and treated in the same manner. If the user chooses this option, the track corresponding to the resource that was provided to the page whose stream was affected must be removed. Additionally, if removing a track in this manner causes the stream to contain no more tracks, the onended event must be raised on it.
A MediaStream may contain more than one
        video and audio track. This makes it possible to include video from two
        or more webcams in a single stream object, for example. However, the
        current API does not allow a page to express a need for multiple video
        streams from independent sources.
It is recommended for multiple calls to getUserMedia() from the same page be allowed as a way for pages to request multiple, discrete, video or audio streams.
A single call to getUserMedia() will always return a stream with either zero or one audio tracks, and either zero or one video tracks. If a script calls getUserMedia() multiple times before reaching a stable state, this document advises the UI designer that the permission dialogs should be merged, so that the user can give permission for the use of multiple cameras and/or media sources in one dialog interaction. The constraints on each getUserMedia call can be used to decide which stream gets which media sources.
This sample code exposes a button. When clicked, the button is disabled and the user is prompted to offer a stream. The user can cause the button to be re-enabled by providing a stream (e.g., giving the page access to the local camera) and then disabling the stream (e.g., revoking that access).
<input type="button" value="Start" onclick="start()" id="startBtn"> <script> var startBtn = document.getElementById('startBtn'); function start() { navigator.getUserMedia({audio:true, video:true}, gotStream, logError); startBtn.disabled = true; } function gotStream(stream) { stream.onended = function () { startBtn.disabled = false; }; } function logError(error) { log(error.name + ": " + error.message); } </script>
This example allows people to take photos of themselves from the local video camera. Note that the forthcoming Image Capture specification may provide a simpler way to accomplish this.
<article> <style scoped> video { transform: scaleX(-1); } p { text-align: center; } </style> <h1>Snapshot Kiosk</h1> <section id="splash"> <p id="errorMessage">Loading...</p> </section> <section id="app" hidden> <p><video id="monitor" autoplay></video> <canvas id="photo"></canvas> <p><input type=button value="📷" onclick="snapshot()"> </section> <script> navigator.getUserMedia({video:true}, gotStream, noStream); var video = document.getElementById('monitor'); var canvas = document.getElementById('photo'); function gotStream(stream) { video.src = URL.createObjectURL(stream); video.onerror = function () { stream.stop(); }; stream.onended = noStream; video.onloadedmetadata = function () { canvas.width = video.videoWidth; canvas.height = video.videoHeight; document.getElementById('splash').hidden = true; document.getElementById('app').hidden = false; }; } function noStream() { document.getElementById('errorMessage').textContent = 'No camera available.'; } function snapshot() { canvas.getContext('2d').drawImage(video, 0, 0); } </script> </article>
IANA is requested to register the following constraints as specified in [RTCWEB-CONSTRAINTS]:
Note: The following specific list(s) of constraints DOES NOT REFLECT CONSENSUS. Many constraints beyond these have been proposed, and the ones listed do not have universal support. The ones below are included **** ONLY **** to provide a starting point so we can see concrete examples of what real constraints might look like. Along those lines, there was an attempt to include constraints of a variety of types so sample type definitions could be included as well (e.g., DOMString, MinMaxConstraint).
The following constraint names are defined to apply to both  and VideoStreamTrack objects:AudioStreamTrack
| Constraint Name | Values | Notes | 
|---|---|---|
| sourceType | SourceTypeEnum | 
            Constrain the video or audio source to an exact value from the set of enumerated-type values of the SourceTypeEnum. | 
          
| sourceId | DOMString | Constrain the video or audio source to an exact source identifier value. | 
The following constraint names are defined to apply only to  objects:VideoStreamTrack
| Constraint Name | Values | Notes | 
|---|---|---|
| width | unsigned long or MinMaxConstraint | 
            Constrain the video source to the exact desired width or width range. | 
| height | unsigned long or MinMaxConstraint | 
            Constrain the video source to the exact desired height or height range. | 
| frameRate | float or MinMaxConstraint | 
            Constrain the video source to the exact desired frame rate (fps) or frameRate range. | 
| facingMode | VideoFacingModeEnum | Constrain the video source to an exact value from the set of enumerated-type values of the VideoFacingModeEnum. | 
The following constraint names are defined to apply only to  objects:AudioStreamTrack
| Constraint Name | Values | Notes | 
|---|---|---|
| volume | unsigned long or MinMaxConstraint | 
            Constrain the audio source to the exact desired volume or volume range. | 
For constraints that accept ranges,
         the  dictionary is also
         defined. Note that the type of the value associated
         with MinMaxConstraintmin and max must be the same
         for both. The specific types associated with min
         and max are defined differently for each
         constraint name.
dictionary MinMaxConstraint {
    any max;
    any min;
};MinMaxConstraint Membersmax of type anymin of type anyThis section will be removed before publication.