CARVIEW |
Media Source Extensions
W3C Editor's Draft 14 November 2013
- This version:
- https://dvcs.w3.org/hg/html-media/raw-file/default/media-source/media-source.html
- Latest published version:
- https://www.w3.org/TR/media-source/
- Latest editor's draft:
- https://dvcs.w3.org/hg/html-media/raw-file/default/media-source/media-source.html
- Editors:
- Aaron Colwell, Google Inc.
- Adrian Bateman, Microsoft Corporation
- Mark Watson, Netflix Inc.
Copyright © 2013 W3C® (MIT, ERCIM, Keio, Beihang), All Rights Reserved. W3C liability, trademark and document use rules apply.
Abstract
This specification extends HTMLMediaElement to allow JavaScript to generate media streams for playback. Allowing JavaScript to generate streams facilitates a variety of use cases like adaptive streaming and time shifting live streams.If you wish to make comments or file bugs regarding this document in a manner that is tracked by the W3C, please submit them via our public bug database.
Status of This Document
This section describes the status of this document at the time of its publication. Other documents may supersede this document. A list of current W3C publications and the latest revision of this technical report can be found in the W3C technical reports index at https://www.w3.org/TR/.
The working groups maintains a list of all bug reports that the editors have not yet tried to address. This draft highlights some of the pending issues that are still to be discussed in the working group. No decision has been taken on the outcome of these issues including whether they are valid.
Implementors should be aware that this specification is not stable. Implementors who are not taking part in the discussions are likely to find the specification changing out from under them in incompatible ways. Vendors interested in implementing this specification before it eventually reaches the Candidate Recommendation stage should join the mailing list mentioned below and take part in the discussions.
This document was published by the HTML Working Group as an Editor's Draft. If you wish to make comments regarding this document, please send them to public-html-media@w3.org (subscribe, archives). All comments are welcome.
Publication as an Editor's Draft does not imply endorsement by the W3C Membership. This is a draft document and may be updated, replaced or obsoleted by other documents at any time. It is inappropriate to cite this document as other than work in progress.
This document was produced by a group operating under the 5 February 2004 W3C Patent Policy. W3C maintains a public list of any patent disclosures made in connection with the deliverables of the group; that page also includes instructions for disclosing a patent. An individual who has actual knowledge of a patent which the individual believes contains Essential Claim(s) must disclose the information in accordance with section 6 of the W3C Patent Policy.
Table of Contents
- 1. Introduction
- 2. MediaSource Object
- 3. SourceBuffer Object
- 3.1 Attributes
- 3.2 Methods
- 3.3 Track Buffers
- 3.4 Event Summary
- 3.5 Algorithms
- 3.5.1 Segment Parser Loop
- 3.5.2 Reset Parser State
- 3.5.3 Append Error Algorithm
- 3.5.4 Prepare Append Algorithm
- 3.5.5 Buffer Append Algorithm
- 3.5.6 Stream Append Loop
- 3.5.7 Initialization Segment Received
- 3.5.8 Coded Frame Processing
- 3.5.9 Coded Frame Removal Algorithm
- 3.5.10 Coded Frame Eviction Algorithm
- 3.5.11 Audio Splice Frame Algorithm
- 3.5.12 Audio Splice Rendering Algorithm
- 3.5.13 Text Splice Frame Algorithm
- 4. SourceBufferList Object
- 5. VideoPlaybackQuality Object
- 6. URL Object Extensions
- 7. HTMLMediaElement Extensions
- 8. HTMLVideoElement Extensions
- 9. AudioTrack Extensions
- 10. VideoTrack Extensions
- 11. TextTrack Extensions
- 12. Byte Stream Formats
- 13. Examples
- 14. Acknowledgments
- 15. Revision History
- A. References
1. Introduction
This specification allows JavaScript to dynamically construct media streams for <audio> and <video>. It defines objects that allow JavaScript to pass media segments to an HTMLMediaElement [HTML5]. A buffering model is also included to describe how the user agent acts when different media segments are appended at different times. Byte stream specifications used with these extensions are available in the byte stream format registry.

1.1 Goals
This specification was designed with the following goals in mind:
- Allow JavaScript to construct media streams independent of how the media is fetched.
- Define a splicing and buffering model that facilitates use cases like adaptive streaming, ad-insertion, time-shifting, and video editing.
- Minimize the need for media parsing in JavaScript.
- Leverage the browser cache as much as possible.
- Provide requirements for byte stream format specifications.
- Not require support for any particular media format or codec.
This specification defines:
- Normative behavior for user agents to enable interoperability between user agents and web applications when processing media data.
- Normative requirements to enable other specifications to define media formats to be used within this specification.
1.2 Definitions
- Active Track Buffers
The track buffers that provide coded frames for the
enabled
audioTracks
, theselected
videoTracks
, and the"showing"
or"hidden"
textTracks
. All these tracks are associated withSourceBuffer
objects in theactiveSourceBuffers
list.- Append Window
A presentation timestamp range used to filter out coded frames while appending. The append window represents a single continuous time range with a single start time and end time. Coded frames with presentation timestamp within this range are allowed to be appended to the
SourceBuffer
while coded frames outside this range are filtered out. The append window start and end times are controlled by theappendWindowStart
andappendWindowEnd
attributes respectively.- Coded Frame
A unit of media data that has a presentation timestamp, a decode timestamp, and a coded frame duration.
- Coded Frame Duration
-
The duration of a coded frame. For video and text, the duration indicates how long the video frame or text should be displayed. For audio, the duration represents the sum of all the samples contained within the coded frame. For example, if an audio frame contained 441 samples @44100Hz the frame duration would be 100 milliseconds.
- Coded Frame Group
A group of coded frames that are adjacent and have monotonically increasing decode timestamps without any gaps. Discontinuities detected by the coded frame processing algorithm and
abort()
calls trigger the start of a new coded frame group.- Decode Timestamp
-
The decode timestamp indicates the latest time at which the frame needs to be decoded assuming instantaneous decoding and rendering of this and any dependant frames (this is equal to the presentation timestamp of the earliest frame, in presentation order, that is dependant on this frame). If frames can be decoded out of presentation order, then the decode timestamp must be present in or derivable from the byte stream. The user agent must run the end of stream algorithm with the error parameter set to
"decode"
if this is not the case. If frames cannot be decoded out of presentation order and a decode timestamp is not present in the byte stream, then the decode timestamp is equal to the presentation timestamp. - Displayed Frame Delay
-
The delay, to the nearest microsecond, between a frame's presentation time and the actual time it was displayed. This delay is always greater than or equal to zero since frames must never be displayed before their presentation time. Non-zero delays are a sign of playback jitter and possible loss of A/V sync.
- Initialization Segment
-
A sequence of bytes that contain all of the initialization information required to decode a sequence of media segments. This includes codec initialization data, Track ID mappings for multiplexed segments, and timestamp offsets (e.g. edit lists).
NoteThe byte stream format specifications in the byte stream format registry contain format specific examples.
- Media Segment
-
A sequence of bytes that contain packetized & timestamped media data for a portion of the media timeline. Media segments are always associated with the most recently appended initialization segment.
NoteThe byte stream format specifications in the byte stream format registry contain format specific examples.
- MediaSource object URL
-
A MediaSource object URL is a unique Blob URI [FILE-API] created by
createObjectURL()
. It is used to attach aMediaSource
object to an HTMLMediaElement.These URLs are the same as a Blob URI, except that anything in the definition of that feature that refers to File and Blob objects is hereby extended to also apply to
MediaSource
objects.The origin of the MediaSource object URL is specified by the Origin of Blob URIs [FILE-API].
NoteFor example, the origin of the MediaSource object URL affects the way that the media element is consumed by canvas.
- Parent Media Source
The parent media source of a
SourceBuffer
object is theMediaSource
object that created it.- Presentation Start Time
The presentation start time is the earliest time point in the presentation and specifies the initial playback position and earliest possible position. All presentations created using this specification have a presentation start time of 0.
- Presentation Interval
-
The presentation interval of a coded frame is the time interval from its presentation timestamp to the presentation timestamp plus the coded frame's duration. For example, if a coded frame has a presentation timestamp of 10 seconds and a coded frame duration of 100 milliseconds, then the presentation interval would be [10-10.1). Note that the start of the range is inclusive, but the end of the range is exclusive.
- Presentation Order
-
The order that coded frames are rendered in the presentation. The presentation order is achieved by ordering coded frames in monotonically increasing order by their presentation timestamps.
- Presentation Timestamp
-
A reference to a specific time in the presentation. The presentation timestamp in a coded frame indicates when the frame must be rendered.
- Random Access Point
A position in a media segment where decoding and continuous playback can begin without relying on any previous data in the segment. For video this tends to be the location of I-frames. In the case of audio, most audio frames can be treated as a random access point. Since video tracks tend to have a more sparse distribution of random access points, the location of these points are usually considered the random access points for multiplexed streams.
- SourceBuffer byte stream format specification
The specific byte stream format specification that describes the format of the byte stream accepted by a
SourceBuffer
instance. The byte stream format specification, for aSourceBuffer
object, is selected based on the type passed to theaddSourceBuffer()
call that created the object.- Track Description
A byte stream format specific structure that provides the Track ID, codec configuration, and other metadata for a single track. Each track description inside a single initialization segment has a unique Track ID. The user agent must run the end of stream algorithm with the error parameter set to
"decode"
if the Track ID is not unique within the initialization segment .- Track ID
A Track ID is a byte stream format specific identifier that marks sections of the byte stream as being part of a specific track. The Track ID in a track description identifies which sections of a media segment belong to that track.
2. MediaSource Object
The MediaSource object represents a source of media data for an HTMLMediaElement. It keeps track of the readyState
for this source as well as a list of SourceBuffer
objects that can be used to add media data to the presentation. MediaSource objects are created by the web application and then attached to an HTMLMediaElement. The application uses the SourceBuffer
objects in sourceBuffers
to add media data to this source. The HTMLMediaElement fetches this media data from the MediaSource
object when it is needed during playback.
enum ReadyState {
"closed",
"open",
"ended"
};
Enumeration description | |
---|---|
closed | Indicates the source is not currently attached to a media element. |
open |
The source has been opened by a media element and is ready for data to be appended to the SourceBuffer objects in sourceBuffers .
|
ended |
The source is still attached to a media element, but endOfStream() has been called.
|
enum EndOfStreamError {
"network",
"decode"
};
Enumeration description | |
---|---|
network |
Terminates playback and signals that a network error has occured. Note JavaScript applications should use this status code to terminate playback with a network error. For example, if a network error occurs while fetching media data. |
decode |
Terminates playback and signals that a decoding error has occured. Note JavaScript applications should use this status code to terminate playback with a decode error. For example, if a parsing error occurs while processing out-of-band media data. |
[Constructor]
interface MediaSource : EventTarget
{
readonly attribute SourceBufferList
sourceBuffers;
readonly attribute SourceBufferList
activeSourceBuffers;
readonly attribute ReadyState
readyState;
attribute unrestricted double
duration;
SourceBuffer
addSourceBuffer (DOMString
type);
void
removeSourceBuffer (SourceBuffer
sourceBuffer);
void
endOfStream (optional EndOfStreamError
error);
static boolean
isTypeSupported (DOMString
type);
};
2.1 Attributes
activeSourceBuffers
of typeSourceBufferList
, readonly-
Contains the subset of
sourceBuffers
that are providing the selected video track, the enabled audio tracks, and the "showing" or "hidden" text tracks.NoteThe Changes to selected/enabled track state section describes how this attribute gets updated.
duration
of typeunrestricted double
,-
Allows the web application to set the presentation duration. The duration is initially set to NaN when the
MediaSource
object is created.On getting, run the following steps:
- If the
readyState
attribute is"closed"
then return NaN and abort these steps. - Return the current value of the attribute.
On setting, run the following steps:
- If the value being set is negative or NaN then throw an
INVALID_ACCESS_ERR
exception and abort these steps. - If the
readyState
attribute is not"open"
then throw anINVALID_STATE_ERR
exception and abort these steps. - If the
updating
attribute equals true on anySourceBuffer
insourceBuffers
, then throw anINVALID_STATE_ERR
exception and abort these steps. - Run the duration change algorithm with new duration set to the value being assigned to this attribute.
Note
appendBuffer()
,appendStream()
andendOfStream()
can update the duration under certain circumstances.
- If the
readyState
of typeReadyState
, readonly-
Indicates the current state of the
MediaSource
object. When theMediaSource
is createdreadyState
must be set to"closed"
. sourceBuffers
of typeSourceBufferList
, readonly-
Contains the list of
SourceBuffer
objects associated with thisMediaSource
. WhenreadyState
equals"closed"
this list will be empty. OncereadyState
transitions to"open"
SourceBuffer objects can be added to this list by usingaddSourceBuffer()
.
2.2 Methods
addSourceBuffer
-
Adds a new
SourceBuffer
tosourceBuffers
.Implementations must support at least 1 MediaSource object with the following
SourceBuffer
configurations. MediaSource objects must support each of the configurations below, but they are only required to support one configuration at a time. Supporting multiple configurations at once or additional configurations is a quality of implementation issue.- A single SourceBuffer with 1 audio track and/or 1 video track.
- Two SourceBuffers with one handling a single audio track and the other handling a single video track.
Parameter Type Nullable Optional Description type DOMString
✘ ✘ Return type:SourceBuffer
When this method is invoked, the user agent must run the following steps:
- If type is null or an empty string then throw an
INVALID_ACCESS_ERR
exception and abort these steps. - If type contains a MIME type that is not supported or contains a MIME type that is not supported with the types specified for the other
SourceBuffer
objects insourceBuffers
, then throw aNOT_SUPPORTED_ERR
exception and abort these steps. - If the user agent can't handle any more SourceBuffer objects then throw a
QUOTA_EXCEEDED_ERR
exception and abort these steps.NoteFor example, a user agent may throw a
QUOTA_EXCEEDED_ERR
exception if the media element has reached theHAVE_METADATA
readyState. This can occur if the user agent's media engine does not support adding more tracks during playback. - If the
readyState
attribute is not in the"open"
state then throw anINVALID_STATE_ERR
exception and abort these steps. - Create a new
SourceBuffer
object and associated resources. - Add the new object to
sourceBuffers
and queue a task to fire a simple event namedaddsourcebuffer
atsourceBuffers
. - Return the new object.
endOfStream
-
Signals the end of the stream.
Parameter Type Nullable Optional Description error EndOfStreamError
✘ ✔ Return type:void
When this method is invoked, the user agent must run the following steps:
- If the
readyState
attribute is not in the"open"
state then throw anINVALID_STATE_ERR
exception and abort these steps. - If the
updating
attribute equals true on anySourceBuffer
insourceBuffers
, then throw anINVALID_STATE_ERR
exception and abort these steps. - Run the end of stream algorithm with the error parameter set to error.
- If the
isTypeSupported
, static-
Check to see whether the
MediaSource
is capable of creatingSourceBuffer
objects for the the specified MIME type.NoteIf true is returned from this method, it only indicates that the
MediaSource
implementation is capable of creatingSourceBuffer
objects for the specified MIME type. AnaddSourceBuffer()
call may still fail if sufficient resources are not available to support the addition of a newSourceBuffer
.NoteThis method returning true implies that HTMLMediaElement.canPlayType() will return "maybe" or "probably" since it does not make sense for a
MediaSource
to support a type the HTMLMediaElement knows it cannot play.Parameter Type Nullable Optional Description type DOMString
✘ ✘ Return type:boolean
When this method is invoked, the user agent must run the following steps:
- If type is an empty string, then return false.
- If type does not contain a valid MIME type string, then return false.
- If type contains a media type or media subtype that the MediaSource does not support, then return false.
- If type contains a codec that the MediaSource does not support, then return false.
- If the MediaSource does not support the specified combination of media type, media subtype, and codecs then return false.
- Return true.
removeSourceBuffer
-
Removes a
SourceBuffer
fromsourceBuffers
.Parameter Type Nullable Optional Description sourceBuffer SourceBuffer
✘ ✘ Return type:void
When this method is invoked, the user agent must run the following steps:
- If sourceBuffer is null then throw an
INVALID_ACCESS_ERR
exception and abort these steps. - If sourceBuffer specifies an object that is not in
sourceBuffers
then throw aNOT_FOUND_ERR
exception and abort these steps. - If the sourceBuffer.
updating
attribute equals true, then run the following steps:- Abort the buffer append and stream append loop algorithms if they are running.
- Set the sourceBuffer.
updating
attribute to false. - Queue a task to fire a simple event named
abort
at sourceBuffer. - Queue a task to fire a simple event named
updateend
at sourceBuffer.
- Let SourceBuffer audioTracks list equal the
AudioTrackList
object returned by sourceBuffer.audioTracks
. - If the SourceBuffer audioTracks list is not empty, then run the following steps:
- Let HTMLMediaElement audioTracks list equal the
AudioTrackList
object returned by theaudioTracks
attribute on the HTMLMediaElement. - Let the removed enabled audio track flag equal false.
- For each
AudioTrack
object in the SourceBuffer audioTracks list, run the following steps:- Set the
sourceBuffer
attribute on theAudioTrack
object to null. - If the
enabled
attribute on theAudioTrack
object is true, then set the removed enabled audio track flag to true. - Remove the
AudioTrack
object from the HTMLMediaElement audioTracks list. - Queue a task to fire a trusted event named
removetrack
, that does not bubble and is not cancelable, and that uses theTrackEvent
interface, at the HTMLMediaElement audioTracks list. - Remove the
AudioTrack
object from the SourceBuffer audioTracks list. - Queue a task to fire a trusted event named
removetrack
, that does not bubble and is not cancelable, and that uses theTrackEvent
interface, at the SourceBuffer audioTracks list.
- Set the
- If the removed enabled audio track flag equals true, then queue a task to fire a simple event named
change
at the HTMLMediaElement audioTracks list.
- Let HTMLMediaElement audioTracks list equal the
- Let SourceBuffer videoTracks list equal the
VideoTrackList
object returned by sourceBuffer.videoTracks
. - If the SourceBuffer videoTracks list is not empty, then run the following steps:
- Let HTMLMediaElement videoTracks list equal the
VideoTrackList
object returned by thevideoTracks
attribute on the HTMLMediaElement. - Let the removed selected video track flag equal false.
- For each
VideoTrack
object in the SourceBuffer videoTracks list, run the following steps:- Set the
sourceBuffer
attribute on theVideoTrack
object to null. - If the
selected
attribute on theVideoTrack
object is true, then set the removed selected video track flag to true. - Remove the
VideoTrack
object from the HTMLMediaElement videoTracks list. - Queue a task to fire a trusted event named
removetrack
, that does not bubble and is not cancelable, and that uses theTrackEvent
interface, at the HTMLMediaElement videoTracks list. - Remove the
VideoTrack
object from the SourceBuffer videoTracks list. - Queue a task to fire a trusted event named
removetrack
, that does not bubble and is not cancelable, and that uses theTrackEvent
interface, at the SourceBuffer videoTracks list.
- Set the
- If the removed selected video track flag equals true, then queue a task to fire a simple event named
change
at the HTMLMediaElement videoTracks list.
- Let HTMLMediaElement videoTracks list equal the
- Let SourceBuffer textTracks list equal the
TextTrackList
object returned by sourceBuffer.textTracks
. - If the SourceBuffer textTracks list is not empty, then run the following steps:
- Let HTMLMediaElement textTracks list equal the
TextTrackList
object returned by thetextTracks
attribute on the HTMLMediaElement. - Let the removed enabled text track flag equal false.
- For each
TextTrack
object in the SourceBuffer textTracks list, run the following steps:- Set the
sourceBuffer
attribute on theTextTrack
object to null. - If the
mode
attribute on theTextTrack
object is set to"showing"
or"hidden"
, then set the removed enabled text track flag to true. - Remove the
TextTrack
object from the HTMLMediaElement textTracks list. - Queue a task to fire a trusted event named
removetrack
, that does not bubble and is not cancelable, and that uses theTrackEvent
interface, at the HTMLMediaElement textTracks list. - Remove the
TextTrack
object from the SourceBuffer textTracks list. - Queue a task to fire a trusted event named
removetrack
, that does not bubble and is not cancelable, and that uses theTrackEvent
interface, at the SourceBuffer textTracks list.
- Set the
- If the removed enabled text track flag equals true, then queue a task to fire a simple event named
change
at the HTMLMediaElement textTracks list.
- Let HTMLMediaElement textTracks list equal the
- If sourceBuffer is in
activeSourceBuffers
, then remove sourceBuffer fromactiveSourceBuffers
and queue a task to fire a simple event namedremovesourcebuffer
at theSourceBufferList
returned byactiveSourceBuffers
. - Remove sourceBuffer from
sourceBuffers
and queue a task to fire a simple event namedremovesourcebuffer
at theSourceBufferList
returned bysourceBuffers
. - Destroy all resources for sourceBuffer.
- If sourceBuffer is null then throw an
2.3 Event Summary
Event name | Interface | Dispatched when... |
---|---|---|
sourceopen |
Event |
readyState transitions from "closed" to "open" or from "ended" to "open" . |
sourceended |
Event |
readyState transitions from "open" to "ended" . |
sourceclose |
Event |
readyState transitions from "open" to "closed" or "ended" to "closed" . |
2.4 Algorithms
2.4.1 Attaching to a media element
A MediaSource
object can be attached to a media element by assigning a MediaSource object URL to the media element src
attribute or the src attribute of a <source> inside a media element. A MediaSource object URL is created by passing a MediaSource object to createObjectURL()
.
If the resource fetch algorithm absolute URL matches the MediaSource object URL, run the following steps right before the "Perform a potentially CORS-enabled fetch" step in the resource fetch algorithm.
- If
readyState
is NOT set to"closed"
- Run the "If the media data cannot be fetched at all, due to network errors, causing the user agent to give up trying to fetch the resource" steps of the resource fetch algorithm.
- Otherwise
-
- Set the
readyState
attribute to"open"
. -
Queue a task to fire a simple event named
sourceopen
at theMediaSource
. - Continue the resource fetch algorithm by running the "Perform a potentially CORS-enabled fetch" step. Text in the resource fetch algorithm
that refers to "the download" or "bytes received" refer to data passed in via
appendBuffer()
andappendStream()
. References to HTTP in the resource fetch algorithm do not apply because the HTMLMediaElement does not fetch media data via HTTP when aMediaSource
is attached.
- Set the
2.4.2 Detaching from a media element
The following steps are run in any case where the media element is going to transition to NETWORK_EMPTY and queue a task to fire a simple event named emptied at the media element. These steps must be run right before the transition.
- Set the
readyState
attribute to"closed"
. - Set the
duration
attribute to NaN. - Remove all the
SourceBuffer
objects fromactiveSourceBuffers
. -
Queue a task to fire a simple event named
removesourcebuffer
atactiveSourceBuffers
. - Remove all the
SourceBuffer
objects fromsourceBuffers
. -
Queue a task to fire a simple event named
removesourcebuffer
atsourceBuffers
. -
Queue a task to fire a simple event named
sourceclose
at theMediaSource
.
2.4.3 Seeking
Run the following steps as part of the "Wait until the user agent has established whether or not the media data for the new playback position is available, and, if it is, until it has decoded enough data to play back that position" step of the seek algorithm:
- The media element looks for media segments containing the new playback position in each
SourceBuffer
object inactiveSourceBuffers
.- If one or more of the objects in
activeSourceBuffers
is missing media segments for the new playback position -
- Set the
HTMLMediaElement.readyState
attribute toHAVE_METADATA
. - The media element waits until an
appendBuffer()
or anappendStream()
call causes the coded frame processing algorithm to set theHTMLMediaElement.readyState
attribute to a value greater thanHAVE_METADATA
.NoteThe web application can use
buffered
to determine what the media element needs to resume playback.
- Set the
- Otherwise
- Continue
- If one or more of the objects in
- The media element resets all decoders and initializes each one with data from the appropriate initialization segment.
- The media element feeds coded frames from the active track buffers into the decoders starting with the closest random access point before the the new playback position.
- Resume the seek algorithm at the "Await a stable state" step.
2.4.4 SourceBuffer Monitoring
The following steps are periodically run during playback to make sure that all of the SourceBuffer
objects in activeSourceBuffers
have enough data to ensure uninterrupted playback. Appending new segments and changes to activeSourceBuffers
also cause these steps to run because they affect the conditions that trigger state transitions.
Having enough data to ensure uninterrupted playback is an implementation specific condition where the user agent
determines that it currently has enough data to play the presentation without stalling for a meaningful period of time. This condition is
constantly evaluated to determine when to transition the media element into and out of the HAVE_ENOUGH_DATA
ready state.
These transitions indicate when the user agent believes it has enough data buffered or it needs more data respectively.
An implementation may choose to use bytes buffered, time buffered, the append rate, or any other metric it sees fit to
determine when it has enough data. The metrics used may change during playback so web applications should only rely on the value of
HTMLMediaElement.readyState
to determine whether more data is needed or not.
When the media element needs more data, the user agent should transition it from HAVE_ENOUGH_DATA
to
HAVE_FUTURE_DATA
early enough for a web application to be able to respond without causing an interruption in playback.
For example, transitioning when the current playback position is 500ms before the end of the buffered data gives the application roughly
500ms to append more data before playback stalls.
- If
buffered
for all objects inactiveSourceBuffers
do not containTimeRanges
for the current playback position: -
- Set the
HTMLMediaElement.readyState
attribute toHAVE_METADATA
. - If this is the first transition to
HAVE_METADATA
, then queue a task to fire a simple event namedloadedmetadata
at the media element. - Abort these steps.
- Set the
- If
buffered
for all objects inactiveSourceBuffers
containTimeRanges
that include the current playback position and enough data to ensure uninterrupted playback: -
- Set the
HTMLMediaElement.readyState
attribute toHAVE_ENOUGH_DATA
. -
Queue a task to fire a simple event named
canplaythrough
at the media element. - Playback may resume at this point if it was previously suspended by a transition to
HAVE_CURRENT_DATA
. - Abort these steps.
- Set the
- If
buffered
for at least one object inactiveSourceBuffers
contains aTimeRange
that includes the current playback position but not enough data to ensure uninterrupted playback: -
- Set the
HTMLMediaElement.readyState
attribute toHAVE_FUTURE_DATA
. - If the previous value of
HTMLMediaElement.readyState
was less thanHAVE_FUTURE_DATA
, then queue a task to fire a simple event namedcanplay
at the media element. - Playback may resume at this point if it was previously suspended by a transition to
HAVE_CURRENT_DATA
. - Abort these steps.
- Set the
- If
buffered
for at least one object inactiveSourceBuffers
contains aTimeRange
that ends at the current playback position and does not have a range covering the time immediately after the current position: -
- Set the
HTMLMediaElement.readyState
attribute toHAVE_CURRENT_DATA
. - If this is the first transition to
HAVE_CURRENT_DATA
, then queue a task to fire a simple event namedloadeddata
at the media element. - Playback is suspended at this point since the media element doesn't have enough data to advance the media timeline.
- Abort these steps.
- Set the
2.4.5 Changes to selected/enabled track state
During playback activeSourceBuffers
needs to be updated if the selected video track, the enabled audio tracks, or a text track mode changes. When one or more of these changes occur the following steps need to be followed.
- If the selected video track changes, then run the following steps:
-
- If the
SourceBuffer
associated with the previously selected video track is not associated with any other enabled tracks, run the following steps:- Remove the
SourceBuffer
fromactiveSourceBuffers
. -
Queue a task to fire a simple event named
removesourcebuffer
atactiveSourceBuffers
- Remove the
- If the
SourceBuffer
associated with the newly selected video track is not already inactiveSourceBuffers
, run the following steps:- Add the
SourceBuffer
toactiveSourceBuffers
. -
Queue a task to fire a simple event named
addsourcebuffer
atactiveSourceBuffers
- Add the
- If the
- If an audio track becomes disabled and the
SourceBuffer
associated with this track is not associated with any other enabled or selected track, then run the following steps: -
- Remove the
SourceBuffer
associated with the audio track fromactiveSourceBuffers
-
Queue a task to fire a simple event named
removesourcebuffer
atactiveSourceBuffers
- Remove the
- If an audio track becomes enabled and the
SourceBuffer
associated with this track is not already inactiveSourceBuffers
, then run the following steps: -
- Add the
SourceBuffer
associated with the audio track toactiveSourceBuffers
-
Queue a task to fire a simple event named
addsourcebuffer
atactiveSourceBuffers
- Add the
- If a text track mode becomes "disabled" and the
SourceBuffer
associated with this track is not associated with any other enabled or selected track, then run the following steps: -
- Remove the
SourceBuffer
associated with the text track fromactiveSourceBuffers
-
Queue a task to fire a simple event named
removesourcebuffer
atactiveSourceBuffers
- Remove the
- If a text track mode becomes "showing" or "hidden" and the
SourceBuffer
associated with this track is not already inactiveSourceBuffers
, then run the following steps: -
- Add the
SourceBuffer
associated with the text track toactiveSourceBuffers
-
Queue a task to fire a simple event named
addsourcebuffer
atactiveSourceBuffers
- Add the
2.4.6 Duration change
Follow these steps when duration
needs to change to a new duration.
- If the current value of
duration
is equal to new duration, then return. - Set old duration to the current value of
duration
. - Update
duration
to new duration. - If the new duration is less than old duration, then call
remove(new duration, old duration)
on all objects insourceBuffers
.NoteThis preserves audio frames and text cues that start before and end after the
duration
. - If a user agent is unable to partially render audio frames or text cues that start before and end after the
duration
, then run the following steps:- Update new duration to the highest end timestamp across all
SourceBuffer
objects insourceBuffers
. - Update
duration
to new duration.
- Update new duration to the highest end timestamp across all
- Update the
media controller duration
to new duration and run the HTMLMediaElement duration change algorithm.
2.4.7 End of stream algorithm
This algorithm gets called when the application signals the end of stream via an endOfStream()
call or an algorithm needs to
signal a decode error. This algorithm takes an error parameter that indicates whether an error will be signalled.
- Change the
readyState
attribute value to"ended"
. -
Queue a task to fire a simple event named
sourceended
at theMediaSource
. - If error is not set, is null, or is an empty string
-
- Run the duration change algorithm with new duration set to the highest end timestamp across all
SourceBuffer
objects insourceBuffers
.
NoteThis allows the duration to properly reflect the end of the appended media segments. For example, if the duration was explicitly set to 10 seconds and only media segments for 0 to 5 seconds were appended before endOfStream() was called, then the duration will get updated to 5 seconds.
- Notify the media element that it now has all of the media data.
- Run the duration change algorithm with new duration set to the highest end timestamp across all
- If error is set to
"network"
-
- If the
HTMLMediaElement.readyState
attribute equalsHAVE_NOTHING
- Run the "If the media data cannot be fetched at all, due to network errors, causing the user agent to give up trying to fetch the resource" steps of the resource fetch algorithm.
- If the
HTMLMediaElement.readyState
attribute is greater thanHAVE_NOTHING
- Run the "If the connection is interrupted after some media data has been received, causing the user agent to give up trying to fetch the resource" steps of the resource fetch algorithm.
- If the
- If error is set to
"decode"
-
- If the
HTMLMediaElement.readyState
attribute equalsHAVE_NOTHING
- Run the "If the media data can be fetched but is found by inspection to be in an unsupported format, or can otherwise not be rendered at all" steps of the resource fetch algorithm.
- If the
HTMLMediaElement.readyState
attribute is greater thanHAVE_NOTHING
- Run the media data is corrupted steps of the resource fetch algorithm.
- If the
- Otherwise
- Throw an
INVALID_ACCESS_ERR
exception.
3. SourceBuffer Object
enum AppendMode {
"segments",
"sequence"
};
Enumeration description | |
---|---|
segments |
The timestamps in the media segment determine where the coded frames are placed in the presentation. Media segments can be appended in any order. |
sequence |
Media segments will be treated as adjacent in time independent of the timestamps in the media segment. Coded frames in a new media segment will be placed immediately after the coded
frames in the previous media segment. The |
interface SourceBuffer : EventTarget
{
attribute AppendMode
mode;
readonly attribute boolean
updating;
readonly attribute TimeRanges
buffered;
attribute double
timestampOffset;
readonly attribute AudioTrackList
audioTracks;
readonly attribute VideoTrackList
videoTracks;
readonly attribute TextTrackList
textTracks;
attribute double
appendWindowStart;
attribute unrestricted double
appendWindowEnd;
void
appendBuffer (ArrayBuffer
data);
void
appendBuffer (ArrayBufferView
data);
void
appendStream (Stream
stream, [EnforceRange] optional unsigned long long
maxSize);
void
abort ();
void
remove (double
start, double
end);
};
3.1 Attributes
appendWindowEnd
of typeunrestricted double
,-
The presentation timestamp for the end of the append window. This attribute is initially set to positive Infinity.
On getting, Return the initial value or the last value that was successfully set.
On setting, run the following steps:
- If this object has been removed from the
sourceBuffers
attribute of the parent media source, then throw anINVALID_STATE_ERR
exception and abort these steps. - If the
updating
attribute equals true, then throw anINVALID_STATE_ERR
exception and abort these steps. - If the new value equals NaN, then throw an
INVALID_ACCESS_ERR
and abort these steps. - If the new value is less than or equal to
appendWindowStart
then throw anINVALID_ACCESS_ERR
exception and abort these steps. - Update the attribute to the new value.
- If this object has been removed from the
appendWindowStart
of typedouble
,-
The presentation timestamp for the start of the append window. This attribute is initially set to 0.
On getting, Return the initial value or the last value that was successfully set.
On setting, run the following steps:
- If this object has been removed from the
sourceBuffers
attribute of the parent media source, then throw anINVALID_STATE_ERR
exception and abort these steps. - If the
updating
attribute equals true, then throw anINVALID_STATE_ERR
exception and abort these steps. - If the new value is less than 0 or greater than or equal to
appendWindowEnd
then throw anINVALID_ACCESS_ERR
exception and abort these steps. - Update the attribute to the new value.
- If this object has been removed from the
audioTracks
of typeAudioTrackList
, readonly-
The list of
AudioTrack
objects created by this object. buffered
of typeTimeRanges
, readonly-
Indicates what
TimeRanges
are buffered in theSourceBuffer
.When the attribute is read the following steps must occur:
- If this object has been removed from the
sourceBuffers
attribute of the parent media source then throw anINVALID_STATE_ERR
exception and abort these steps. - Return a new static normalized TimeRanges object for the media segments buffered.
- If this object has been removed from the
mode
of typeAppendMode
,-
Controls how a sequence of media segments are handled. This attribute is initially set to
"segments"
when the object is created.On getting, Return the initial value or the last value that was successfully set.
On setting, run the following steps:
- Let new mode equal the new value being assigned to this attribute.
- If new mode does not equal
"segments"
or"sequence"
, then throw anINVALID_ACCESS_ERR
exception and abort these steps. - If this object has been removed from the
sourceBuffers
attribute of the parent media source, then throw anINVALID_STATE_ERR
exception and abort these steps. - If the
updating
attribute equals true, then throw anINVALID_STATE_ERR
exception and abort these steps. -
If the
readyState
attribute of the parent media source is in the"ended"
state then run the following steps:- Set the
readyState
attribute of the parent media source to"open"
- Queue a task to fire a simple event named
sourceopen
at the parent media source.
- Set the
- If the append state equals PARSING_MEDIA_SEGMENT, then throw an
INVALID_STATE_ERR
and abort these steps. - If the new mode equals
"sequence"
, then set the group start timestamp to the highest presentation end timestamp. - Update the attribute to new mode.
textTracks
of typeTextTrackList
, readonly-
The list of
TextTrack
objects created by this object. timestampOffset
of typedouble
,-
Controls the offset applied to timestamps inside subsequent media segments that are appended to this
SourceBuffer
. ThetimestampOffset
is initially set to 0 which indicates that no offset is being applied.On getting, Return the initial value or the last value that was successfully set.
On setting, run the following steps:
- Let new timestamp offset equal the new value being assigned to this attribute.
- If this object has been removed from the
sourceBuffers
attribute of the parent media source, then throw anINVALID_STATE_ERR
exception and abort these steps. - If the
updating
attribute equals true, then throw anINVALID_STATE_ERR
exception and abort these steps. -
If the
readyState
attribute of the parent media source is in the"ended"
state then run the following steps:- Set the
readyState
attribute of the parent media source to"open"
- Queue a task to fire a simple event named
sourceopen
at the parent media source.
- Set the
- If the append state equals PARSING_MEDIA_SEGMENT, then throw an
INVALID_STATE_ERR
and abort these steps. - If the
mode
attribute equals"sequence"
, then set the group start timestamp to new timestamp offset. - Update the attribute to new timestamp offset.
updating
of typeboolean
, readonly-
Indicates whether the asynchronous continuation of an
appendBuffer()
,appendStream()
, orremove()
operation is still being processed. This attribute is initially set to false when the object is created. videoTracks
of typeVideoTrackList
, readonly-
The list of
VideoTrack
objects created by this object.
3.2 Methods
abort
-
Aborts the current segment and resets the segment parser.
No parameters.Return type:void
When this method is invoked, the user agent must run the following steps:
- If this object has been removed from the
sourceBuffers
attribute of the parent media source then throw anINVALID_STATE_ERR
exception and abort these steps. - If the
readyState
attribute of the parent media source is not in the"open"
state then throw anINVALID_STATE_ERR
exception and abort these steps. - If the
updating
attribute equals true, then run the following steps:- Abort the buffer append and stream append loop algorithms if they are running.
- Set the
updating
attribute to false. - Queue a task to fire a simple event named
abort
at thisSourceBuffer
object. - Queue a task to fire a simple event named
updateend
at thisSourceBuffer
object.
- Run the reset parser state algorithm.
- Set
appendWindowStart
to 0. - Set
appendWindowEnd
to positive Infinity.
- If this object has been removed from the
appendBuffer
-
Appends the segment data in an
ArrayBuffer
[TYPED-ARRAYS] to the source buffer.The steps for this method are the same as the ArrayBufferView version of
appendBuffer()
.Parameter Type Nullable Optional Description data ArrayBuffer
✘ ✘ Return type:void
appendBuffer
-
Appends the segment data in an
ArrayBufferView
[TYPED-ARRAYS] to the source buffer.Parameter Type Nullable Optional Description data ArrayBufferView
✘ ✘ Return type:void
When this method is invoked, the user agent must run the following steps:
- If data is null then throw an
INVALID_ACCESS_ERR
exception and abort these steps. - Run the prepare append algorithm.
- Add data to the end of the input buffer.
- Set the
updating
attribute to true. - Queue a task to fire a simple event named
updatestart
at thisSourceBuffer
object. - Asynchronously run the buffer append algorithm.
- If data is null then throw an
appendStream
-
Appends segment data to the source buffer from a
Stream
[STREAMS-API].Parameter Type Nullable Optional Description stream Stream
✘ ✘ maxSize unsigned long long
✘ ✔ Return type:void
When this method is invoked, the user agent must run the following steps:
- If stream is null then throw an
INVALID_ACCESS_ERR
exception and abort these steps. - Run the prepare append algorithm.
- Set the
updating
attribute to true. - Queue a task to fire a simple event named
updatestart
at thisSourceBuffer
object. - Asynchronously run the stream append loop algorithm with stream and maxSize.
- If stream is null then throw an
remove
-
Removes media for a specific time range.
Parameter Type Nullable Optional Description start double
✘ ✘ end double
✘ ✘ Return type:void
When this method is invoked, the user agent must run the following steps:
- If start is negative or greater than
duration
, then throw anINVALID_ACCESS_ERR
exception and abort these steps. - If end is less than or equal to start, then throw an
INVALID_ACCESS_ERR
exception and abort these steps. - If this object has been removed from the
sourceBuffers
attribute of the parent media source then throw anINVALID_STATE_ERR
exception and abort these steps. - If the
updating
attribute equals true, then throw anINVALID_STATE_ERR
exception and abort these steps. -
If the
readyState
attribute of the parent media source is in the"ended"
state then run the following steps:- Set the
readyState
attribute of the parent media source to"open"
- Queue a task to fire a simple event named
sourceopen
at the parent media source .
- Set the
- Set the
updating
attribute to true. - Queue a task to fire a simple event named
updatestart
at thisSourceBuffer
object. - Return control to the caller and run the rest of the steps asynchronously.
- Run the coded frame removal algorithm with start and end as the start and end of the removal range.
- Set the
updating
attribute to false. - Queue a task to fire a simple event named
update
at thisSourceBuffer
object. - Queue a task to fire a simple event named
updateend
at thisSourceBuffer
object.
- If start is negative or greater than
3.3 Track Buffers
A track buffer stores the track descriptions and coded frames for an individual
track. The track buffer is updated as initialization segments and media segments are appended to the
SourceBuffer
.
Each track buffer has a last decode timestamp variable that stores the decode timestamp of the last coded frame appended in the current coded frame group. The variable is initially unset to indicate that no coded frames have been appended yet.
Each track buffer has a last frame duration variable that stores the coded frame duration of the last coded frame appended in the current coded frame group. The variable is initially unset to indicate that no coded frames have been appended yet.
Each track buffer has a highest presentation timestamp variable that stores the highest presentation timestamp encountered in a coded frame appended in the current coded frame group. The variable is initially unset to indicate that no coded frames have been appended yet.
Each track buffer has a need random access point flag variable that keeps track of whether the track buffer is waiting for a random access point coded frame. The variable is initially set to true to indicate that random access point coded frame is needed before anything can be added to the track buffer.
3.4 Event Summary
Event name | Interface | Dispatched when... |
---|---|---|
updatestart |
Event |
updating transitions from false to true. |
update |
Event |
The append or remove has successfully completed. updating transitions from true to false. |
updateend |
Event |
The append or remove has ended. |
error |
Event |
An error occurred during the append. updating transitions from true to false. |
abort |
Event |
The append or remove was aborted by an abort() call. updating transitions from true to false. |
3.5 Algorithms
3.5.1 Segment Parser Loop
All SourceBuffer objects have an internal append state variable that keeps track of the high-level segment parsing state. It is initially set to WAITING_FOR_SEGMENT and can transition to the following states as data is appended.
Append state name | Description |
---|---|
WAITING_FOR_SEGMENT | Waiting for the start of an initialization segment or media segment to be appended. |
PARSING_INIT_SEGMENT | Currently parsing an initialization segment. |
PARSING_MEDIA_SEGMENT | Currently parsing a media segment. |
The input buffer is a byte buffer that is used to hold unparsed bytes across appendBuffer()
and appendStream()
calls. The buffer is empty when the SourceBuffer object is created.
The buffer full flag keeps track of whether appendBuffer()
or
appendStream()
is allowed to accept more bytes. It is set to false when the SourceBuffer object is created and gets updated
as data is appended and removed.
The group start timestamp variable keeps track of the starting timestamp for a new
coded frame group in the "sequence"
mode.
It is unset when the SourceBuffer object is created and gets updated when the mode
attribute equals "sequence"
and the
timestampOffset
attribute is set, or the coded frame processing algorithm runs.
The highest presentation end timestamp variable stores the highest presentation end timestamp encountered in the current coded frame group. It is set to 0 when the SourceBuffer object is created and gets updated by the coded frame processing algorithm.
When the segment parser loop algorithm is invoked, run the following steps:
- Loop Top: If the input buffer is empty, then jump to the need more data step below.
- If the input buffer contains bytes that violate the SourceBuffer byte stream format specification, then run the
end of stream algorithm with the error parameter set to
"decode"
and abort this algorithm. - Remove any bytes that the byte stream format specifications say must be ignored from the start of the input buffer.
-
If the append state equals WAITING_FOR_SEGMENT, then run the following steps:
- If the beginning of the input buffer indicates the start of an initialization segment, set the append state to PARSING_INIT_SEGMENT.
- If the beginning of the input buffer indicates the start of an media segment, set append state to PARSING_MEDIA_SEGMENT.
- Jump to the loop top step above.
-
If the append state equals PARSING_INIT_SEGMENT, then run the following steps:
- If the input buffer does not contain a complete initialization segment yet, then jump to the need more data step below.
- Run the initialization segment received algorithm.
- Remove the initialization segment bytes from the beginning of the input buffer.
- Set append state to WAITING_FOR_SEGMENT.
- Jump to the loop top step above.
-
If the append state equals PARSING_MEDIA_SEGMENT, then run the following steps:
- If the first initialization segment flag is false, then run the end of stream algorithm with the error parameter set to
"decode"
and abort this algorithm. - If the input buffer does not contain a complete media segment header yet, then jump to the need more data step below.
- If the input buffer contains one or more complete coded frames, then run the
coded frame processing algorithm.
Note
The frequency at which the coded frame processing algorithm is run is implementation-specific. The coded frame processing algorithm may be called when the input buffer contains the complete media segment or it may be called multiple times as complete coded frames are added to the input buffer.
- If this
SourceBuffer
is full and cannot accept more media data, then set the buffer full flag to true. - If the input buffer does not contain a complete media segment, then jump to the need more data step below.
- Remove the media segment bytes from the beginning of the input buffer.
- Set append state to WAITING_FOR_SEGMENT.
- Jump to the loop top step above.
- If the first initialization segment flag is false, then run the end of stream algorithm with the error parameter set to
- Need more data: Return control to the calling algorithm.
3.5.2 Reset Parser State
When the parser state needs to be reset, run the following steps:
- If the append state equals PARSING_MEDIA_SEGMENT and the input buffer contains some complete coded frames, then run the coded frame processing algorithm until all of these complete coded frames have been processed.
- Unset the last decode timestamp on all track buffers.
- Unset the last frame duration on all track buffers.
- Unset the highest presentation timestamp on all track buffers.
- Set the need random access point flag on all track buffers to true.
- Remove all bytes from the input buffer.
- Set append state to WAITING_FOR_SEGMENT.
3.5.3 Append Error Algorithm
When an error occurs during an append, run the following steps:
- Run the reset parser state algorithm.
- Set the
updating
attribute to false. - Queue a task to fire a simple event named
error
at thisSourceBuffer
object. - Queue a task to fire a simple event named
updateend
at thisSourceBuffer
object.
3.5.4 Prepare Append Algorithm
When an append operation begins, the follow steps are run to validate and prepare the SourceBuffer
.
- If the
SourceBuffer
has been removed from thesourceBuffers
attribute of the parent media source then throw anINVALID_STATE_ERR
exception and abort these steps. - If the
updating
attribute equals true, then throw anINVALID_STATE_ERR
exception and abort these steps. -
If the
readyState
attribute of the parent media source is in the"ended"
state then run the following steps:- Set the
readyState
attribute of the parent media source to"open"
-
Queue a task to fire a simple event named
sourceopen
at the parent media source .
- Set the
- Run the coded frame eviction algorithm.
-
If the buffer full flag equals true, then throw a
QUOTA_EXCEEDED_ERR
exception and abort these step.NoteThis is the signal that the implementation was unable to evict enough data to accomodate the append or the append is too big. The web application should use
remove()
to explicitly free up space and/or reduce the size of the append.
3.5.5 Buffer Append Algorithm
When appendBuffer()
is called, the following steps are run to process the appended data.
- Run the segment parser loop algorithm.
- If the segment parser loop algorithm in the previous step was aborted, then abort this algorithm.
- Set the
updating
attribute to false. - Queue a task to fire a simple event named
update
at thisSourceBuffer
object. - Queue a task to fire a simple event named
updateend
at thisSourceBuffer
object.
3.5.6 Stream Append Loop
When a Stream
[STREAMS-API] is passed to appendStream()
, the following steps are run to transfer data from the
Stream
to the SourceBuffer
. This algorithm is initialized with the stream and maxSize parameters
from the appendStream()
call.
- If maxSize is set, then let bytesLeft equal maxSize.
- Loop Top: If maxSize is set and bytesLeft equals 0, then jump to the loop done step below.
- If stream has been closed, then jump to the loop done step below.
- Read data from stream into data:
- If maxSize is set:
-
- Read up to bytesLeft bytes from stream into data.
- Subtract the number of bytes in data from bytesLeft.
- Otherwise:
- Read all available bytes in stream into data.
- If an error occured while reading from stream, then run the append error algorithm and abort this algorithm.
- Run the coded frame eviction algorithm.
-
If the buffer full flag equals true, then run the append error algorithm and abort this algorithm.
NoteThe web application should use
remove()
to free up space in theSourceBuffer
. - Add data to the end of the input buffer.
- Run the segment parser loop algorithm.
- If the segment parser loop algorithm in the previous step was aborted, then abort this algorithm.
- Jump to the loop top step above.
- Loop Done: Set the
updating
attribute to false. - Queue a task to fire a simple event named
update
at thisSourceBuffer
object. - Queue a task to fire a simple event named
updateend
at thisSourceBuffer
object.
3.5.7 Initialization Segment Received
The following steps are run when the segment parser loop successfully parses a complete initialization segment:
Each SourceBuffer object has an internal first initialization segment flag that tracks whether the first initialization segment has been appended. This flag is set to false when the SourceBuffer is created and updated by the algorithm below.
- Update the
duration
attribute if it currently equals NaN:- If the initialization segment contains a duration:
- Run the duration change algorithm with new duration set to the duration in the initialization segment.
- Otherwise:
- Run the duration change algorithm with new duration set to positive Infinity.
- If the initialization segment has no audio, video, or text tracks, then run the end of stream algorithm with the error parameter set to
"decode"
and abort these steps. - If the first initialization segment flag is true, then run the following steps:
- Verify the following properties. If any of the checks fail then run the end of stream algorithm with the error parameter set to
"decode"
and abort these steps.- The number of audio, video, and text tracks match what was in the first initialization segment.
- The codecs for each track, match what was specified in the first initialization segment.
- If more than one track for a single type are present (ie 2 audio tracks), then the Track IDs match the ones in the first initialization segment.
- Add the appropriate track descriptions from this initialization segment to each of the track buffers.
- Verify the following properties. If any of the checks fail then run the end of stream algorithm with the error parameter set to
- Let active track flag equal false.
-
If the first initialization segment flag is false, then run the following steps:
- If the initialization segment contains tracks with codecs the user agent does not support, then run the end of stream algorithm with the error parameter set to
"decode"
and abort these steps.NoteUser agents may consider codecs, that would otherwise be supported, as "not supported" here if the codecs were not specified in the type parameter passed to
addSourceBuffer()
.
For example, MediaSource.isTypeSupported('video/webm;codecs="vp8,vorbis"') may return true, but ifaddSourceBuffer()
was called with 'video/webm;codecs="vp8"' and a Vorbis track appears in the initialization segment, then the user agent may use this step to trigger a decode error. -
For each audio track in the initialization segment, run following steps:
- Let new audio track be a new
AudioTrack
object. - Generate a unique ID and assign it to the
id
property on new audio track. -
If
audioTracks
.length
equals 0, then run the following steps:- Set the
enabled
property on new audio track to true. - Set active track flag to true.
- Set the
- Add new audio track to the
audioTracks
attribute on thisSourceBuffer
object. - Queue a task to fire a trusted event named
addtrack
, that does not bubble and is not cancelable, and that uses theTrackEvent
interface, at theAudioTrackList
object referenced by theaudioTracks
attribute on thisSourceBuffer
object. - Add new audio track to the
audioTracks
attribute on the HTMLMediaElement. - Queue a task to fire a trusted event named
addtrack
, that does not bubble and is not cancelable, and that uses theTrackEvent
interface, at theAudioTrackList
object referenced by theaudioTracks
attribute on the HTMLMediaElement. - Create a new track buffer to store coded frames for this track.
- Add the track description for this track to the track buffer.
- Let new audio track be a new
-
For each video track in the initialization segment, run following steps:
- Let new video track be a new
VideoTrack
object. - Generate a unique ID and assign it to the
id
property on new video track. -
If
videoTracks
.length
equals 0, then run the following steps:- Set the
selected
property on new video track to true. - Set active track flag to true.
- Set the
- Add new video track to the
videoTracks
attribute on thisSourceBuffer
object. - Queue a task to fire a trusted event named
addtrack
, that does not bubble and is not cancelable, and that uses theTrackEvent
interface, at theVideoTrackList
object referenced by thevideoTracks
attribute on thisSourceBuffer
object. - Add new video track to the
videoTracks
attribute on the HTMLMediaElement. - Queue a task to fire a trusted event named
addtrack
, that does not bubble and is not cancelable, and that uses theTrackEvent
interface, at theVideoTrackList
object referenced by thevideoTracks
attribute on the HTMLMediaElement. - Create a new track buffer to store coded frames for this track.
- Add the track description for this track to the track buffer.
- Let new video track be a new
-
For each text track in the initialization segment, run following steps:
-
Let new text track be a new
TextTrack
object with its properties populated with the appropriate information from the initialization segment. -
If the
mode
property on new text track equals"showing"
or"hidden"
, then set active track flag to true. - Add new text track to the
textTracks
attribute on thisSourceBuffer
object. - Queue a task to fire a trusted event named
addtrack
, that does not bubble and is not cancelable, and that uses theTrackEvent
interface, attextTracks
attribute on thisSourceBuffer
object. - Add new text track to the
textTracks
attribute on the HTMLMediaElement. - Queue a task to fire a trusted event named
addtrack
, that does not bubble and is not cancelable, and that uses theTrackEvent
interface, at theTextTrackList
object referenced by thetextTracks
attribute on the HTMLMediaElement. - Create a new track buffer to store coded frames for this track.
- Add the track description for this track to the track buffer.
-
Let new text track be a new
- If active track flag equals true, then run the following steps:
- Add this
SourceBuffer
toactiveSourceBuffers
. - Queue a task to fire a simple event named
addsourcebuffer
atactiveSourceBuffers
- Add this
- Set first initialization segment flag to true.
- If the initialization segment contains tracks with codecs the user agent does not support, then run the end of stream algorithm with the error parameter set to
-
If the
HTMLMediaElement.readyState
attribute isHAVE_NOTHING
, then run the following steps:-
If one or more objects in
sourceBuffers
have first initialization segment flag set to false, then abort these steps. - Set the
HTMLMediaElement.readyState
attribute toHAVE_METADATA
. - Queue a task to fire a simple event named
loadedmetadata
at the media element.
-
If one or more objects in
-
If the active track flag equals true and the
HTMLMediaElement.readyState
attribute is greater thanHAVE_CURRENT_DATA
, then set theHTMLMediaElement.readyState
attribute toHAVE_METADATA
.
3.5.8 Coded Frame Processing
When complete coded frames have been parsed by the segment parser loop then the following steps are run:
-
For each coded frame in the media segment run the following steps:
- Loop Top: Let presentation timestamp be a double precision floating point representation of the coded frame's presentation timestamp in seconds.
Note
Special processing may be needed to determine the presentation and decode timestamps for timed text frames since this information may not be explicilty present in the underlying format or may be dependent on the order of the frames. Some metadata text tracks, like MPEG2-TS PSI data, may only have implied timestamps. Format specific rules for these situations should be in the byte stream format specifications or in separate extension specifications.
- Let decode timestamp be a double precision floating point representation of the coded frame's decode timestamp in seconds.
Note
Implementations don't have to internally store timestamps in a double precision floating point representation. This representation is used here because it is the represention for timestamps in the HTML spec. The intention here is to make the behavior clear without adding unnecessary complexity to the algorithm to deal with the fact that adding a timestampOffset may cause a timestamp rollover in the underlying timestamp representation used by the byte stream format. Implementations can use any internal timestamp representation they wish, but the addition of timestampOffset should behave in a similar manner to what would happen if a double precision floating point representation was used.
- Let frame duration be a double precision floating point representation of the coded frame's duration in seconds.
- If
mode
equals"sequence"
and group start timestamp is set, then run the following steps:- Set
timestampOffset
equal to group start timestamp - presentation timestamp. - Set highest presentation end timestamp equal to group start timestamp.
- Set the need random access point flag on all track buffers to true.
- Unset group start timestamp.
- Set
-
If
timestampOffset
is not 0, then run the following steps:- Add
timestampOffset
to the presentation timestamp. - Add
timestampOffset
to the decode timestamp. - If the presentation timestamp or decode timestamp is less than the presentation start time,
then run the end of stream algorithm with the error parameter set to
"decode"
, and abort these steps.
- Add
- Let track buffer equal the track buffer that the coded frame will be added to.
-
- If last decode timestamp for track buffer is set and decode timestamp is less than last decode timestamp:
- OR
- If last decode timestamp for track buffer is set and the difference between decode timestamp and last decode timestamp is greater than 2 times last frame duration:
-
-
- If
mode
equals"segments"
: - Set highest presentation end timestamp to presentation timestamp.
- If
mode
equals"sequence"
: - Set group start timestamp equal to the highest presentation end timestamp.
- If
- Unset the last decode timestamp on all track buffers.
- Unset the last frame duration on all track buffers.
- Unset the highest presentation timestamp on all track buffers.
- Set the need random access point flag on all track buffers to true.
- Jump to the Loop Top step above to restart processing of the current coded frame.
-
- Otherwise:
- Continue.
- Let frame end timestamp equal the sum of presentation timestamp and frame duration.
- If presentation timestamp is less than
appendWindowStart
, then set the need random access point flag to true, drop the coded frame, and jump to the top of the loop to start processing the next coded frame.NoteSome implementations may choose to collect some of these coded frames that are outside the append window and use them to generate a splice at the first coded frame that has a presentation timestamp greater than or equal to
appendWindowStart
even if that frame is not a random access point. Supporting this requires multiple decoders or faster than real-time decoding so for now this behavior will not be a normative requirement. - If frame end timestamp is greater than
appendWindowEnd
, then set the need random access point flag to true, drop the coded frame, and jump to the top of the loop to start processing the next coded frame. - If the need random access point flag on track buffer equals true, then run the following steps:
- If the coded frame is not a random access point, then drop the coded frame and jump to the top of the loop to start processing the next coded frame.
- Set the need random access point flag on track buffer to false.
- Let spliced audio frame be an unset variable for holding audio splice information
- Let spliced timed text frame be an unset variable for holding timed text splice information
- If last decode timestamp for track buffer is unset and presentation timestamp falls within the presentation interval of a coded frame in track buffer,then run the following steps:
- Let overlapped frame be the coded frame in track buffer that matches the condition above.
-
- If track buffer contains audio coded frames:
- Run the audio splice frame algorithm and if a splice frame is returned, assign it to spliced audio frame.
- If track buffer contains video coded frames:
-
- Let overlapped frame presentation timestamp equal the presentation timestamp of overlapped frame.
- Let remove window timestamp equal overlapped frame presentation timestamp plus 1 microsecond.
- If the presentation timestamp is less than the remove window timestamp, then remove overlapped frame and any
coded frames that depend on it from track buffer.
Note
This is to compensate for minor errors in frame timestamp computations that can appear when converting back and forth between double precision floating point numbers and rationals. This tolerance allows a frame to replace an existing one as long as it is within 1 microsecond of the existing frame's start time. Frames that come slightly before an existing frame are handled by the removal step below.
- If track buffer contains timed text coded frames:
- Run the text splice frame algorithm and if a splice frame is returned, assign it to spliced timed text frame.
- Remove existing coded frames in track buffer:
- If highest presentation timestamp for track buffer is not set:
- Remove all coded frames from track buffer that have a presentation timestamp greater than or equal to presentation timestamp and less than frame end timestamp.
- If highest presentation timestamp for track buffer is set and less than presentation timestamp
- Remove all coded frames from track buffer that have a presentation timestamp greater than highest presentation timestamp and less than or equal to frame end timestamp.
- Remove decoding dependencies of the coded frames removed in the previous step:
- If detailed information about decoding dependencies is available:
- Remove all coded frames from track buffer that have decoding dependencies on the coded frames removed in
the previous step.
Note
For example if an I-frame is removed in the previous step, then all P-frames & B-frames that depend on that I-frame should be removed from track buffer. This makes sure that decode dependencies are properly maintained during overlaps.
- Otherwise:
- Remove all coded frames between the coded frames removed in the previous step and the next
random access point after those removed frames.
Note
Removing all coded frames until the next random access point is a conservative estimate of the decoding dependencies since it assumes all frames between the removed frames and the next random access point depended on the frames that were removed.
-
- If spliced audio frame is set:
- Add spliced audio frame to the track buffer.
- If spliced timed text frame is set:
- Add spliced timed text frame to the track buffer.
- Otherwise:
- Add the coded frame with the presentation timestamp, decode timestamp, and frame duration to the track buffer.
- Set last decode timestamp for track buffer to decode timestamp.
- Set last frame duration for track buffer to frame duration.
- If highest presentation timestamp for track buffer is unset or frame end timestamp is greater
than highest presentation timestamp, then set highest presentation timestamp for track buffer
to frame end timestamp.
Note
The greater than check is needed because bidirectional prediction between coded frames can cause presentation timestamp to not be monotonically increasing eventhough the decode timestamps are monotonically increasing.
- If highest presentation end timestamp is unset or frame end timestamp is greater than highest presentation end timestamp, then set highest presentation end timestamp equal to frame end timestamp.
- Loop Top: Let presentation timestamp be a double precision floating point representation of the coded frame's presentation timestamp in seconds.
-
If the
HTMLMediaElement.readyState
attribute isHAVE_METADATA
and the new coded frames cause all objects inactiveSourceBuffers
to have media data for the current playback position, then run the following steps:- Set the
HTMLMediaElement.readyState
attribute toHAVE_CURRENT_DATA
. - If this is the first transition to
HAVE_CURRENT_DATA
, then queue a task to fire a simple event namedloadeddata
at the media element.
- Set the
-
If the
HTMLMediaElement.readyState
attribute isHAVE_CURRENT_DATA
and the new coded frames cause all objects inactiveSourceBuffers
to have media data beyond the current playback position, then run the following steps:- Set the
HTMLMediaElement.readyState
attribute toHAVE_FUTURE_DATA
. -
Queue a task to fire a simple event named
canplay
at the media element.
- Set the
-
If the
HTMLMediaElement.readyState
attribute isHAVE_FUTURE_DATA
and the new coded frames cause all objects inactiveSourceBuffers
to have enough data to ensure uninterrupted playback, then run the following steps:- Set the
HTMLMediaElement.readyState
attribute toHAVE_ENOUGH_DATA
. -
Queue a task to fire a simple event named
canplaythrough
at the media element.
- Set the
- If the media segment contains data beyond the current
duration
, then run the duration change algorithm with new duration set to the maximum of the current duration and the highest end timestamp reported byHTMLMediaElement.buffered
.
3.5.9 Coded Frame Removal Algorithm
Follow these steps when coded frames for a specific time range need to be removed from the SourceBuffer:
- Let start be the starting presentation timestamp for the removal range.
- Let end be the end presentation timestamp for the removal range.
For each track buffer in this source buffer, run the following steps:
- Let remove end timestamp be the current value of
duration
-
If this track buffer has a random access point timestamp that is greater than or equal to end, then update remove end timestamp to that random access point timestamp.
NoteRandom access point timestamps can be different across tracks because the dependencies between coded frames within a track are usually different than the dependencies in another track.
- Remove all media data, from this track buffer, that contain starting timestamps greater than or equal to start and less than the remove end timestamp.
-
If this object is in
activeSourceBuffers
, the current playback position is greater than or equal to start and less than the remove end timestamp, andHTMLMediaElement.readyState
is greater thanHAVE_METADATA
, then set theHTMLMediaElement.readyState
attribute toHAVE_METADATA
and stall playback.NoteThis transition occurs because media data for the current position has been removed. Playback cannot progress until media for the current playback position is appended or the selected/enabled tracks change.
- Let remove end timestamp be the current value of
- If buffer full flag equals true and this object is ready to accept more bytes, then set the buffer full flag to false.
3.5.10 Coded Frame Eviction Algorithm
This algorithm is run to free up space in this source buffer when new data is appended.
- Let new data equal the data that is about to be appended to this SourceBuffer.
- If the buffer full flag equals false, then abort these steps.
- Let removal ranges equal a list of presentation time ranges that can be evicted from the presentation to make room for the
new data.
Note
Implementations may use different methods for selecting removal ranges so web applications should not depend on a specific behavior. The web application can use the
buffered
attribute to observe whether portions of the buffered data have been evicted. - For each range in removal ranges, run the coded frame removal algorithm with start and end equal to the removal range start and end timestamp respectively.
3.5.11 Audio Splice Frame Algorithm
Follow these steps when the coded frame processing algorithm needs to generate a splice frame for two overlapping audio coded frames:
- Let track buffer be the track buffer that will contain the splice.
- Let new coded frame be the new coded frame, that is being added to track buffer, which triggered the need for a splice.
- Let presentation timestamp be the presentation timestamp for new coded frame
- Let decode timestamp be the decode timestamp for new coded frame.
- Let frame duration be the coded frame duration of new coded frame.
- Let overlapped frame be the coded frame in track buffer with a presentation interval that contains presentation timestamp.
- Update presentation timestamp and decode timestamp to the nearest audio sample timestamp based on sample rate of the
audio in overlapped frame. If a timestamp is equidistant from both audio sample timestamps, then use the higher timestamp. (eg.
floor(x * sample_rate + 0.5) / sample_rate).
Note
For example, given the following values:
- The presentation timestamp of overlapped frame equals 10.
- The sample rate of overlapped frame equals 8000 Hz
- presentation timestamp equals 10.01255
- decode timestamp equals 10.01255
presentation timestamp and decode timestamp are updated to 10.0125 since 10.01255 is closer to 10 + 100/8000 (10.0125) than 10 + 101/8000 (10.012625)
- If the user agent does not support crossfading then run the following steps:
- Remove overlapped frame from track buffer.
- Add a silence frame to track buffer with the following properties:
- The presentation timestamp set to the overlapped frame presentation timestamp.
- The decode timestamp set to the overlapped frame decode timestamp.
- The coded frame duration set to difference between presentation timestamp and the overlapped frame presentation timestamp.
NoteSome implementations may apply fades to/from silence to coded frames on either side of the inserted silence to make the transition less jarring.
- Return to caller without providing a splice frame.
Note
This is intended to allow new coded frame to be added to the track buffer as if overlapped frame had not been in the track buffer to begin with.
- Let frame end timestamp equal the sum of presentation timestamp and frame duration.
- Let splice end timestamp equal the sum of presentation timestamp and the splice duration of 5 milliseconds.
- Let fade out coded frames equal overlapped frame as well as any additional frames in track buffer that have a presentation timestamp greater than presentation timestamp and less than splice end timestamp.
- Remove all the frames included in fade out coded frames from track buffer.
- Return a splice frame with the following properties:
- The presentation timestamp set to the overlapped frame presentation timestamp.
- The decode timestamp set to the overlapped frame decode timestamp.
- The coded frame duration set to difference between frame end timestamp and the overlapped frame presentation timestamp.
- The fade out coded frames equals fade-out coded frames.
- The fade in coded frame equal new coded frame.
Note
If the new coded frame is less than 5 milliseconds in duration, then coded frames that are appended after the new coded frame will be needed to properly render the splice.
- The splice timestamp equals presentation timestamp.
NoteSee the audio splice rendering algorithm for details on how this splice frame is rendered.
3.5.12 Audio Splice Rendering Algorithm
The following steps are run when a spliced frame, generated by the audio splice frame algorithm, needs to be rendered by the media element:
- Let fade out coded frames be the coded frames that are faded out during the splice.
- Let fade in coded frames be the coded frames that are faded in during the splice.
- Let presentation timestamp be the presentation timestamp of the first coded frame in fade out coded frames.
- Let end timestamp be the sum of the presentation timestamp and the coded frame duration of the last frame in fade in coded frames.
- Let splice timestamp be the presentation timestamp where the splice starts. This corresponds with the presentation timestamp of the first frame in fade in coded frames.
- Let splice end timestamp equal splice timestamp plus five milliseconds.
- Let fade out samples be the samples generated by decoding fade out coded frames.
- Trim fade out samples so that it only contains samples between presentation timestamp and splice end timestamp.
- Let fade in samples be the samples generated by decoding fade in coded frames.
- If fade out samples and fade in samples do not have a common sample rate and channel layout, then convert fade out samples and fade in samples to a common sample rate and channel layout.
- Let output samples be a buffer to hold the output samples.
- Apply a linear gain fade out with a starting gain of 1 and an ending gain of 0 to the samples between splice timestamp and splice end timestamp in fade out samples.
- Apply a linear gain fade in with a starting gain of 0 and an ending gain of 1 to the samples between splice timestamp and splice end timestamp in fade in samples.
- Copy samples between presentation timestamp to splice timestamp from fade out samples into output samples.
- For each sample between splice timestamp and splice end timestamp, compute the sum of a sample from fade out samples and the corresponding sample in fade in samples and store the result in output samples.
- Copy samples between splice end timestamp to end timestamp from fade in samples into output samples.
- Render output samples.
Here is a graphical representation of this algorithm.

3.5.13 Text Splice Frame Algorithm
Follow these steps when the coded frame processing algorithm needs to generate a splice frame for two overlapping timed text coded frames:
- Let track buffer be the track buffer that will contain the splice.
- Let new coded frame be the new coded frame, that is being added to track buffer, which triggered the need for a splice.
- Let presentation timestamp be the presentation timestamp for new coded frame
- Let decode timestamp be the decode timestamp for new coded frame.
- Let frame duration be the coded frame duration of new coded frame.
- Let frame end timestamp equal the sum of presentation timestamp and frame duration.
- Let first overlapped frame be the coded frame in track buffer with a presentation interval that contains presentation timestamp.
- Let overlapped presentation timestamp be the presentation timestamp of the first overlapped frame.
- Let overlapped frames equal first overlapped frame as well as any additional frames in track buffer that have a presentation timestamp greater than presentation timestamp and less than frame end timestamp.
- Remove all the frames included in overlapped frames from track buffer.
- Update the coded frame duration of the first overlapped frame to presentation timestamp - overlapped presentation timestamp.
- Add first overlapped frame to the track buffer.
- Return to caller without providing a splice frame.
Note
This is intended to allow new coded frame to be added to the track buffer as if it hadn't overlapped any frames in track buffer to begin with.
4. SourceBufferList Object
SourceBufferList is a simple container object for SourceBuffer
objects. It provides read-only array access and fires events when the list is modified.
interface SourceBufferList : EventTarget
{
readonly attribute unsigned long
length;
getter SourceBuffer (unsigned long
index);
};
4.1 Attributes
length
of typeunsigned long
, readonly-
Indicates the number of
SourceBuffer
objects in the list.
4.2 Methods
SourceBuffer
-
Allows the SourceBuffer objects in the list to be accessed with an array operator (i.e. []).
Parameter Type Nullable Optional Description index unsigned long
✘ ✘ Return type:getter
When this method is invoked, the user agent must run the following steps:
- If index is greater than or equal to the
length
attribute then return undefined and abort these steps. - Return the index'th
SourceBuffer
object in the list.
- If index is greater than or equal to the
4.3 Event Summary
Event name | Interface | Dispatched when... |
---|---|---|
addsourcebuffer |
Event |
When a SourceBuffer is added to the list. |
removesourcebuffer |
Event |
When a SourceBuffer is removed from the list. |
5. VideoPlaybackQuality Object
interface VideoPlaybackQuality {
readonly attribute DOMHighResTimeStamp
creationTime;
readonly attribute unsigned long
totalVideoFrames;
readonly attribute unsigned long
droppedVideoFrames;
readonly attribute unsigned long
corruptedVideoFrames;
readonly attribute double
totalFrameDelay;
};
5.1 Attributes
corruptedVideoFrames
of typeunsigned long
, readonly-
The total number of corrupted frames that have been detected.
creationTime
of typeDOMHighResTimeStamp
, readonly-
The timestamp returned by Performance.now() when this object was created.
droppedVideoFrames
of typeunsigned long
, readonly-
The total number of frames dropped predecode or dropped because the frame missed its display deadline.
totalFrameDelay
of typedouble
, readonly-
The sum of all displayed frame delays for all displayed frames. (i.e., Frames included in the
totalVideoFrames
count, but not in thedroppedVideoFrames
count. totalVideoFrames
of typeunsigned long
, readonly-
The total number of frames that would have been displayed if no frames are dropped.
6. URL Object Extensions
This section specifies extensions to the URL[FILE-API] object definition.
partial interface URL {
static DOMString
createObjectURL (MediaSource
mediaSource);
};
6.1 Methods
createObjectURL
, static-
Creates URLs for
MediaSource
objects.NoteThis algorithm is intended to mirror the behavior of the createObjectURL()[FILE-API] method with autoRevoke set to true.
Parameter Type Nullable Optional Description mediaSource MediaSource
✘ ✘ Return type:DOMString
When this method is invoked, the user agent must run the following steps:
- If mediaSource is NULL the return null.
- Return a unique MediaSource object URL that can be used to dereference the mediaSource argument, and run the rest of the algorithm asynchronously.
- provide a stable state
- Revoke the MediaSource object URL by calling revokeObjectURL() on it.
7. HTMLMediaElement Extensions
This section specifies what existing attributes on the HTMLMediaElement
must return when a MediaSource
is attached to the element.
The HTMLMediaElement.seekable attribute returns a new static normalized TimeRanges object created based on the following steps:
- If
duration
equals NaN - Return an empty
TimeRanges
object. - If
duration
equals positive Infinity - Return a single range with a start time of 0 and an end time equal to the highest end time reported by the
HTMLMediaElement.buffered
attribute. - Otherwise
- Return a single range with a start time of 0 and an end time equal to
duration
.
The HTMLMediaElement.buffered
attribute returns a new static normalized TimeRanges object created based on the following steps:
- If
activeSourceBuffers
.length equals 0 then return an emptyTimeRanges
object and abort these steps. - Let active ranges be the ranges returned by
buffered
for eachSourceBuffer
object inactiveSourceBuffers
. - Let highest end time be the largest range end time in the active ranges.
- Let intersection ranges equal a
TimeRange
object containing a single range from 0 to highest end time. - For each
SourceBuffer
object inactiveSourceBuffers
run the following steps:- Let source ranges equal the ranges returned by the
buffered
attribute on the currentSourceBuffer
. - If
readyState
is"ended"
, then set the end time on the last range in source ranges to highest end time. - Let new intersection ranges equal the the intersection between the intersection ranges and the source ranges.
- Replace the ranges in intersection ranges with the new intersection ranges.
- Let source ranges equal the ranges returned by the
- Return the intersection ranges.
8. HTMLVideoElement Extensions
This section specifies new attributes and internal state that are being added to the HTMLVideoElement
.
Each HTMLVideoElement
will maintain a total video frame count variable that keeps
track of the total number of frames that have been displayed and dropped. This variable is initialized to 0 when the
element is created and whenever the media element load algorithm is invoked. It is incremented when a video frame is displayed
or when the dropped video frame count is incremented.
Each HTMLVideoElement
will maintain a dropped video frame count variable that keeps
track of the total number of frames that have been dropped. This variable is initialized to 0 when the
element is created and whenever the media element load algorithm is invoked. It is incremented when a video frame is dropped
predecode or when a frame is decoded but dropped because it missed a display deadline.
Each HTMLVideoElement
will maintain a corrupted video frame count variable that keeps
track of the total number of corrupted frames detected. This variable is initialized to 0 when the element is created and whenever the
media element load algorithm is invoked. It is incremented when a corrupted video frame is detected by the decoder. It is up to
the implementation to determine whether to display or drop a corrupted frame. Whichever choice is made, the total video frame count
and dropped video frame count must be updated appropriately.
Each HTMLVideoElement
will maintain a displayed frame delay sum variable that keeps
track of the sum of all displayed frame delays. This variable is initialized to 0 when the element is created and whenever the
media element load algorithm is invoked. When a frame is displayed, its displayed frame delay is computed and added
to this variable.
partial interface HTMLVideoElement {
VideoPlaybackQuality
getVideoPlaybackQuality ();
};
8.1 Methods
getVideoPlaybackQuality
-
Provides the current the playback quality metrics.
No parameters.Return type:VideoPlaybackQuality
When this method is invoked, the user agent must run the following steps:
- Let playbackQuality be a new instance of
VideoPlaybackQuality
. - Set playbackQuality.
creationTime
to the value returned by a call to Performance.now(). - Set playbackQuality.
totalVideoFrames
to the current value of the total video frame count. - Set playbackQuality.
droppedVideoFrames
to the current value of the dropped video frame count. - Set playbackQuality.
corruptedVideoFrames
to the current value of the corrupted video frame count. - Set playbackQuality.
totalFrameDelay
to the current value of the displayed frame delay sum. - Return playbackQuality.
- Let playbackQuality be a new instance of
9. AudioTrack Extensions
This section specifies extensions to the HTML AudioTrack
definition.
partial interface AudioTrack {
attribute DOMString
kind;
attribute DOMString
language;
readonly attribute SourceBuffer
? sourceBuffer;
};
9.1 Attributes
kind
of typeDOMString
,-
Allows the web application to get and update the track
kind
.On getting, return the current value of the attribute. This is either the value provided when this object was created or the value provided on the last successful set operation.
On setting, run the following steps:
- If the value being assigned to this attribute does not match one of the kind categories, then abort these steps.
- Update this attribute to the new value.
- If the
sourceBuffer
attribute on this track is not null, then queue a task to fire a simple event namedchange
atsourceBuffer
.audioTracks
. - Queue a task to fire a simple event named
change
at theAudioTrackList
object referenced by theaudioTracks
attribute on the HTMLMediaElement.
language
of typeDOMString
,-
Allows the web application to get and update the track
language
.On getting, return the current value of the attribute. This is either the value provided when this object was created or the value provided on the last successful set operation.
On setting, run the following steps:
- If the value being assigned to this attribute is not an empty string or a BCP 47 language tag[BCP47], then abort these steps.
- Update this attribute to the new value.
- If the
sourceBuffer
attribute on this track is not null, then queue a task to fire a simple event namedchange
atsourceBuffer
.audioTracks
. - Queue a task to fire a simple event named
change
at theAudioTrackList
object referenced by theaudioTracks
attribute on the HTMLMediaElement.
sourceBuffer
of typeSourceBuffer
, readonly , nullable-
Returns the
SourceBuffer
that created this track. Returns null if this track was not created by aSourceBuffer
or theSourceBuffer
has been removed from thesourceBuffers
attribute of its parent media source.
10. VideoTrack Extensions
This section specifies extensions to the HTML VideoTrack
definition.
partial interface VideoTrack {
attribute DOMString
kind;
attribute DOMString
language;
readonly attribute SourceBuffer
? sourceBuffer;
};
10.1 Attributes
kind
of typeDOMString
,-
Allows the web application to get and update the track
kind
.On getting, return the current value of the attribute. This is either the value provided when this object was created or the value provided on the last successful set operation.
On setting, run the following steps:
- If the value being assigned to this attribute does not match one of the kind categories, then abort these steps.
- Update this attribute to the new value.
- If the
sourceBuffer
attribute on this track is not null, then queue a task to fire a simple event namedchange
atsourceBuffer
.videoTracks
. - Queue a task to fire a simple event named
change
at theVideoTrackList
object referenced by thevideoTracks
attribute on the HTMLMediaElement.
language
of typeDOMString
,-
Allows the web application to get and update the track
language
.On getting, return the current value of the attribute. This is either the value provided when this object was created or the value provided on the last successful set operation.
On setting, run the following steps:
- If the value being assigned to this attribute is not an empty string or a BCP 47 language tag[BCP47], then abort these steps.
- Update this attribute to the new value.
- If the
sourceBuffer
attribute on this track is not null, then queue a task to fire a simple event namedchange
atsourceBuffer
.videoTracks
. - Queue a task to fire a simple event named
change
at theVideoTrackList
object referenced by thevideoTracks
attribute on the HTMLMediaElement.
sourceBuffer
of typeSourceBuffer
, readonly , nullable-
Returns the
SourceBuffer
that created this track. Returns null if this track was not created by aSourceBuffer
or theSourceBuffer
has been removed from thesourceBuffers
attribute of its parent media source.
11. TextTrack Extensions
This section specifies extensions to the HTML TextTrack
definition.
partial interface TextTrack {
attribute DOMString
kind;
attribute DOMString
language;
readonly attribute SourceBuffer
? sourceBuffer;
};
11.1 Attributes
kind
of typeDOMString
,-
Allows the web application to get and update the track
kind
.On getting, return the current value of the attribute. This is either the value provided when this object was created or the value provided on the last successful set operation.
On setting, run the following steps:
- If the value being assigned to this attribute does not match one of the text track kinds, then abort these steps.
- Update this attribute to the new value.
- If the
sourceBuffer
attribute on this track is not null, then queue a task to fire a simple event namedchange
atsourceBuffer
.textTracks
. - Queue a task to fire a simple event named
change
at theTextTrackList
object referenced by thetextTracks
attribute on the HTMLMediaElement.
language
of typeDOMString
,-
Allows the web application to get and update the track
language
.On getting, return the current value of the attribute. This is either the value provided when this object was created or the value provided on the last successful set operation.
On setting, run the following steps:
- If the value being assigned to this attribute is not an valid text track language, then abort these steps.
- Update this attribute to the new value.
- If the
sourceBuffer
attribute on this track is not null, then queue a task to fire a simple event namedchange
atsourceBuffer
.textTracks
. - Queue a task to fire a simple event named
change
at theTextTrackList
object referenced by thetextTracks
attribute on the HTMLMediaElement.
sourceBuffer
of typeSourceBuffer
, readonly , nullable-
Returns the
SourceBuffer
that created this track. Returns null if this track was not created by aSourceBuffer
or theSourceBuffer
has been removed from thesourceBuffers
attribute of its parent media source.
12. Byte Stream Formats
The bytes provided through appendBuffer()
and appendStream()
for a SourceBuffer
form a logical byte stream. The format and
semantics of these byte streams are defined in byte stream format specifications. The byte stream format registry is the
authoritative source for byte stream format specifications that can be accepted by a SourceBuffer
. If a MediaSource
implementation claims
to support any of the MIME types in the registry, then it must implement the corresponding byte stream format specification.
The byte stream format specifications in the registry are not intended to define new storage formats. They simply outline the subset of existing storage format structures that implementations of this specification will accept.
Byte stream format parsing and validation is implemented in the segment parser loop algorithm.
This section provides general requirements for all byte stream format specifications:
- A byte stream format specification must define initialization segments and media segments.
- It must be possible to identify segment boundaries and segment type (initialization or media) by examining the byte stream alone.
- The user agent must run the end of stream algorithm with the error parameter set to
"decode"
when any of the following conditions are met:-
The number and type of tracks are not consistent.
NoteFor example, if the first initialization segment has 2 audio tracks and 1 video track, then all initialization segments that follow it in the byte stream must describe 2 audio tracks and 1 video track.
- Track IDs are not the same across initialization segments, for segments describing multiple tracks of a single type. (e.g. 2 audio tracks).
-
Codecs changes across initialization segments.
NoteFor example, a byte stream that starts with an initialization segment that specifies a single AAC track and later contains an initialization segment that specifies a single AMR-WB track is not allowed. Support for multiple codecs is handled with multiple
SourceBuffer
objects.
-
- The user agent must support the following:
- Track IDs changing across initialization segments if the segments describes only one track of each type.
-
Video frame size changes. The user agent must support seamless playback.
NoteThis will cause the <video> display region to change size if the web application does not use CSS or HTML attributes (width/height) to constrain the element size.
-
Audio channel count changes. The user agent may support this seamlessly and could trigger downmixing.
NoteThis is a quality of implementation issue because changing the channel count may require reinitializing the audio device, resamplers, and channel mixers which tends to be audible.
- The following rules apply to all media segments within a byte stream. A user agent must:
- Map all timestamps to the same media timeline.
- Support seamless playback of media segments having a timestamp gap smaller than the audio frame size. User agent must not reflect these gaps in the
buffered
attribute.NoteThis is intended to simplify switching between audio streams where the frame boundaries don't always line up across encodings (e.g. Vorbis).
- The user agent must run the end of stream algorithm with the error parameter set to
"decode"
when any combination of an initialization segment and any contiguous sequence of media segments satisfies the following conditions:- The number and type (audio, video, text, etc.) of all tracks in the media segments are not identified.
- The decoding capabilities needed to decode each track (i.e. codec and codec parameters) are not provided.
- Encryption parameters necessary to decrypt the content (except the encryption key itself) are not provided for all encrypted tracks.
- All information necessary to decode and render the earliest random access point in the sequence of media segments and all subsequence samples in the sequence
(in presentation time) are not provided. This includes in particular,
- Information that determines the intrinsic width and height of the video (specifically, this requires either the picture or pixel aspect ratio, together with the encoded resolution).
- Information necessary to convert the video decoder output to a format suitable for display
- Information necessary to compute the global presentation timestamp of every sample in the sequence of media segments is not provided.
For example, if I1 is associated with M1, M2, M3 then the above must hold for all the combinations I1+M1, I1+M2, I1+M1+M2, I1+M2+M3, etc.
Byte stream specifications must at a minimum define constraints which ensure that the above requirements hold. Additional constraints may be defined, for example to simplify implementation.
13. Examples
Example use of the Media Source Extensions
<script> function onSourceOpen(videoTag, e) { var mediaSource = e.target; if (mediaSource.sourceBuffers.length > 0) return; var sourceBuffer = mediaSource.addSourceBuffer('video/webm; codecs="vorbis,vp8"'); videoTag.addEventListener('seeking', onSeeking.bind(videoTag, mediaSource)); videoTag.addEventListener('progress', onProgress.bind(videoTag, mediaSource)); var initSegment = GetInitializationSegment(); if (initSegment == null) { // Error fetching the initialization segment. Signal end of stream with an error. mediaSource.endOfStream("network"); return; } // Append the initialization segment. var firstAppendHandler = function(e) { var sourceBuffer = e.target; sourceBuffer.removeEventListener('updateend', firstAppendHandler); // Append some initial media data. appendNextMediaSegment(mediaSource); }; sourceBuffer.addEventListener('updateend', firstAppendHandler); sourceBuffer.appendBuffer(initSegment); } function appendNextMediaSegment(mediaSource) { if (mediaSource.readyState == "closed") return; // If we have run out of stream data, then signal end of stream. if (!HaveMoreMediaSegments()) { mediaSource.endOfStream(); return; } // Make sure the previous append is not still pending. if (mediaSource.sourceBuffers[0].updating) return; var mediaSegment = GetNextMediaSegment(); if (!mediaSegment) { // Error fetching the next media segment. mediaSource.endOfStream("network"); return; } // NOTE: If mediaSource.readyState == “ended”, this appendBuffer() call will // cause mediaSource.readyState to transition to "open". The web application // should be prepared to handle multiple “sourceopen” events. mediaSource.sourceBuffers[0].appendBuffer(mediaSegment); } function onSeeking(mediaSource, e) { var video = e.target; if (mediaSource.readyState == "open") { // Abort current segment append. mediaSource.sourceBuffers[0].abort(); } // Notify the media segment loading code to start fetching data at the // new playback position. SeekToMediaSegmentAt(video.currentTime); // Append a media segment from the new playback position. appendNextMediaSegment(mediaSource); } function onProgress(mediaSource, e) { appendNextMediaSegment(mediaSource); } </script> <video id="v" autoplay> </video> <script> var video = document.getElementById('v'); var mediaSource = new MediaSource(); mediaSource.addEventListener('sourceopen', onSourceOpen.bind(this, video)); video.src = window.URL.createObjectURL(mediaSource); </script>
14. Acknowledgments
The editors would like to thank Alex Giladi, Bob Lund, Chris Poole, Cyril Concolato, David Dorwin, David Singer, Duncan Rowden, Frank Galligan, Glenn Adams, Jerry Smith, Joe Steele, John Simmons, Kevin Streeter, Mark Vickers, Matt Ward, Michael Thornburgh, Philip Jägenstedt, Pierre Lemieux, Ralph Giles, Steven Robertson, and Tatsuya Igarashi for their contributions to this specification.15. Revision History
Version | Comment |
---|---|
14 November 2013 |
|
04 November 2013 |
|
29 October 2013 |
|
28 October 2013 |
|
15 October 2013 |
|
04 October 2013 |
|
26 July 2013 |
|
18 July 2013 |
|
02 July 2013 |
|
05 June 2013 |
|
01 June 2013 |
|
13 May 2013 |
|
06 May 2013 |
|
02 May 2013 |
|
30 April 2013 |
|
24 April 2013 |
|
08 April 2013 |
|
26 March 2013 |
|
12 March 2013 |
|
05 March 2013 |
|
25 February 2013 |
|
19 February 2013 |
|
05 February 2013 |
|
31 January 2013 |
|
30 January 2013 |
|
15 January 2013 | Replace setTrackInfo() and getSourceBuffer() with AudioTrack, VideoTrack, and TextTrack extensions. |
04 January 2013 |
|
14 December 2012 | Pubrules, Link Checker, and Markup Validation fixes. |
13 December 2012 |
|
08 December 2012 |
|
06 December 2012 |
|
28 November 2012 |
|
09 November 2012 | Converted document to ReSpec. |
18 October 2012 | Refactored SourceBuffer.append() & added SourceBuffer.remove(). |
8 October 2012 |
|
1 October 2012 | Fixed various addsourcebuffer & removesourcebuffer bugs and allow append() in ended state. |
13 September 2012 | Updated endOfStream() behavior to change based on the value of HTMLMediaElement.readyState. |
24 August 2012 |
|
22 August 2012 |
|
17 August 2012 | Minor editorial fixes. |
09 August 2012 | Change presentation start time to always be 0 instead of using format specific rules about the first media segment appended. |
30 July 2012 | Added SourceBuffer.timestampOffset and MediaSource.duration. |
17 July 2012 | Replaced SourceBufferList.remove() with MediaSource.removeSourceBuffer(). |
02 July 2012 | Converted to the object-oriented API |
26 June 2012 | Converted to Editor's draft. |
0.5 | Minor updates before proposing to W3C HTML-WG. |
0.4 | Major revision. Adding source IDs, defining buffer model, and clarifying byte stream formats. |
0.3 | Minor text updates. |
0.2 | Updates to reflect initial WebKit implementation. |
0.1 | Initial Proposal |
A. References
A.1 Informative references
- [BCP47]
- A. Phillips; M. Davis. Tags for Identifying Languages. September 2009. IETF Best Current Practice. URL: https://tools.ietf.org/html/bcp47
- [FILE-API]
- Arun Ranganathan; Jonas Sicking. File API. 25 October 2012. W3C Working Draft. URL: https://www.w3.org/TR/2012/WD-FileAPI-20121025
- [HTML5]
- Robin Berjon et al. HTML5. 17 December 2012. W3C Candidate Recommendation. URL: https://www.w3.org/TR/html5/
- [STREAMS-API]
- Feras Moussa. Streams API. 25 October 2012. W3C Editor's Draft. URL: https://dvcs.w3.org/hg/streams-api/raw-file/tip/Overview.htm
- [TYPED-ARRAYS]
- David Herman, Kenneth Russell. Typed Arrays Khronos Working Draft. (Work in progress.) URL: https://www.khronos.org/registry/typedarray/specs/latest/