Copyright © 2012 W3C® (MIT, ERCIM, Keio), All Rights Reserved. W3C liability, trademark and document use rules apply.
This section describes the status of this document at the time of its publication. Other documents may supersede this document. A list of current W3C publications and the latest revision of this technical report can be found in the W3C technical reports index at http://www.w3.org/TR/.
This document was published by the HTML working group as an Editor's Draft. Please submit comments regarding this document by using the W3C's (public bug database) with the product set to HTML WG and the component set to Media Source Extensions. If you cannot access the bug database, submit comments to public-html-media@w3.org (subscribe, archives) and arrangements will be made to transpose the comments to the bug database. All feedback is welcome.
Publication as a Editor's Draft does not imply endorsement by the W3C Membership. This is a draft document and may be updated, replaced or obsoleted by other documents at any time. It is inappropriate to cite this document as other than work in progress.
This document was produced by a group operating under the 5 February 2004 W3C Patent Policy. W3C maintains a public list of any patent disclosures made in connection with the deliverables of the group; that page also includes instructions for disclosing a patent. An individual who has actual knowledge of a patent which the individual believes contains Essential Claim(s) must disclose the information in accordance with section 6 of the W3C Patent Policy.
This proposal extends HTMLMediaElement to allow JavaScript to generate media streams for playback. Allowing JavaScript to generate streams facilitates a variety of use cases like adaptive streaming and time shifting live streams.
This proposal allows JavaScript to dynamically construct media streams for <audio> and <video>.
It defines objects that allow JavaScript to pass media segments to an HTMLMediaElement
.
A buffering model is also included to describe how the user agent should act when different media segments are
appended at different times. Byte stream specifications for WebM & ISO Base Media File Format are given to specify the
expected format of media segments used with these extensions.
This proposal was designed with the following goals in mind:
A sequence of bytes that contains all of the initialization information required to decode a sequence of media segments. This includes codec initialization data, trackID mappings for multiplexed segments, and timestamp offsets (e.g. edit lists).
Container specific examples of initialization segments:
A sequence of bytes that contain packetized & timestamped media data for a portion of the presentation timeline. Media segments are always associated with the most recently appended initialization segment.
Container specific examples of media segments:
A hypothetical buffer that contains a distinct sequence of initialization segments & media segments. When media segments are passed to append()
they update the state of this buffer. The source buffer only allows a single media segment to cover a specific point in the presentation timeline of each track. If a media segment gets appended that contains media data overlapping (in presentation time) with media data from an existing segment, then the new media data will override the old media data. Since media segments depend on initialization segments the source buffer is also responsible for maintaining these associations. During playback, the media element pulls segment data out of the source buffers, demultiplexes it if necessary, and enqueues it into track buffers so it will get decoded and displayed. buffered
describes the time ranges that are covered by media segments in the source buffer.
The set of source buffers that are providing the selected video track
, the enabled audio tracks
, and the "showing"
or "hidden"
text tracks. This is a subset of all the source buffers associated with a specific MediaSource
object. Details about how this set is managed are given here.
A hypothetical buffer that represents initialization and media data for a single AudioTrack
, VideoTrack
, or TextTrack
that has been queued for playback. This buffer may not exist in actual implementations, but it is intended to represent media data that will be decoded no matter what media segments are appended to update the source buffer. This distinction is important when considering appends that happen close to the current playback position. Details about transfers between the source buffer and track buffers are given here.
A position in a media segment where decoding and continuous playback can begin without relying on any previous data in the segment. For video this tends to be the location of I-frames. In the case of audio, most audio frames can be treated as a random access point. Since video tracks tend to have a more sparse distribution of random access points, the location of these points are usually considered the random access points for multiplexed streams.
The subsections below outline the buffering model for this proposal. It describes how to add and remove source buffers from the presentation and describes the various rules and behaviors associated with appending data to an individual source buffer. At the highest level, the web application simply creates source buffers and appends a sequence of initialization segments and media segments to update the buffer's state. The media element pulls media data out of the source buffers, plays it, and fires events just like it would if a normal URL was passed to the src
attribute. The web application is expected to monitor media element events to determine when it needs to append more media segments.
SourceBuffer
objects can be created once a MediaSource
object enters the "open"
state. The application calls addSourceBuffer()
with a type string that indicates the format of the data it intends to append to the new SourceBuffer. If the user agent supports the format and has sufficent resources, a new SourceBuffer
object is created, added to sourceBuffers
, and returned by the method. If the user agent doesn't support the specified format or can't support another SourceBuffer
then it will throw an appropriate exception to signal why the request couldn't be satisfied.
Updating the state of a source buffer requires appending at least one initialization segment and one or more media segments via append()
. The following list outlines some of the basic rules for appending segments.
abort()
is called.append()
).To simplify the implementation and facilitate interoperability, a few constraints are placed on the initialization segments that are appended to a specific SourceBuffer
:
SourceBuffer
MUST describe 2 audio tracks and 1 video track.SourceBuffer
objects.To simplify the implementation and facilitate interoperability, a few constraints are placed on the media segments that are appended to a specific SourceBuffer
:
buffered
.Once a new SourceBuffer
has been created, it expects an initialization segment to be appended first. This first segment indicates the number and type of streams contained in the media segments that follow. This allows the media element to configure the necessary decoders and output devices. This first segment can also cause a HTMLMediaElement.readyState
transition to HAVE_METADATA
if this is the first SourceBuffer
, or if it is the first track of a specific type (i.e. first audio, first video track, or first text track). If neither of the conditions hold then the tracks for this new SourceBuffer
will just appear as disabled tracks and won't affect the current HTMLMediaElement.readyState
until they are selected. The media element will also add the appropriate tracks to the audioTracks
, videoTracks
, & textTracks
collections and fire the necessary change
events. The description for append()
contains all the details.
If a media segment is appended to a time range that is not covered by existing segments in the source buffer, then its data is copied directly into the source buffer. Addition of this data may trigger HTMLMediaElement.readyState
transitions depending on what other data is buffered and whether the media element has determined if it can start playback. Calls to buffered
will always reflect the current TimeRanges
buffered in the SourceBuffer
.
There are several ways that media segments can overlap segments in the source buffer. Behavior for the different overlap situations are described below. If more than one overlap applies, then the start overlap gets resolved first, followed by any complete overlaps, and finally the end overlap. If a segment contains multiple tracks then the overlap is resolved independently for each track.
The figure above shows how the source buffer gets updated when a new media segment completely overlaps a segment in the buffer. In this case, the new segment completely replaces the old segment.
The figure above shows how the source buffer gets updated when the beginning of a new media segment overlaps a segment in the buffer. In this case the new segment replaces all the old media data in the overlapping region. Since media segments are constrained to starting with random access points, this provides a seamless transition between segments.
The one case that requires special attention is where an audio frame overlaps with the start of the new media segment. The base level behavior that MUST be supported requires dropping the old audio frame that overlaps the start of the new segment and inserting silence for the small gap that is created. A higher quality implementation could support outputting a portion of the old segment and all of the new segment or crossfade during the overlapping region. This is a quality of implementation issue. The key property here though is the small silence gap should not be reflected in the ranges reported by buffered
The figure above shows how the source buffer gets updated when the end of a new media segment overlaps a segment in the buffer. In this case, the media element tries to keep as much of the old segment as possible. The amount saved depends on where the closest random access point, in the old segment, is to the end of the new segment. In the case of audio, if the gap is smaller than the size of an audio frame, then the media element should insert silence for this gap and not reflect it in buffered
.
An implementation may keep old segment data before the end of the new segment to avoid creating a gap if it wishes. Doing this though can significantly increase implementation complexity and could cause delays at the splice point. The key property that must be preserved is the entirety of the new segment gets added to the source buffer and it is up to the implementation how much of the old segment data is retained. The web application can use buffered
to determine how much of the old segment was preserved.
The figure above shows how the source buffer gets updated when the new media segment is in the middle of the old segment. This condition is handled by first resolving the start overlap and then resolving the end overlap.
The source buffer represents the media that the web application would like the media element to play. The track buffer contains the data that will actually get decoded and rendered. In most cases the track buffer will simply contain a subset of the source buffer near the current playback position. These two buffers start to diverge though when media segments that overlap or are very close to the current playback position are appended. Depending on the contents of the new media segment it may not be possible to switch to the new data immediately because there isn't a random access point close enough to the current playback position. The quality of the implementation determines how much data is considered "in the track buffer". It should transfer data to the track buffer as late as possible whilst maintaining seamless playback. Some implementations may be able to instantiate multiple decoders or decode the new data significantly faster than real-time to achieve a seamless splice immediately. Other implementations may delay until the next random access point before switching to the newly appended data. Notice that this difference in behavior is only observable when appending close to the current playback position. The track buffer represents a media subsegment, like a group of pictures or something with similar decode dependencies, that the media element commits to playing. This commitment may be influenced by a variety of things like limited decoding resources, hardware decode buffers, a jitter buffer, or the desire to limit implementation complexity.
Here is an example to help clarify the role of the track buffer. Say the current playback position has a timestamp of 8 and the media element pulled frames with timestamp 9 & 10 into the track buffer. The web application then appends a higher quality media segment that starts with a random access point at timestamp 9. The source buffer will get updated with the higher quality data, but the media element won't be able to switch to this higher quality data until the next random access point at timestamp 20. This is because a frame for timestamp 9 is already in the track buffer. As you can see the track buffer represents the "point of no return." for decoding. If a seek occurs the media element may choose to use the higher quality data since a seek might imply flushing the track buffer and the user expects a break in playback.
When a new media segment is appended, memory constraints may cause previously appended segments to get evicted from the source buffer. The eviction algorithm is implementation dependent, but segments that aren't likely to be needed soon are the most likely to get evicted. The buffered
method allows the web application to monitor what time ranges are currently buffered in the source buffer.
Removing a SourceBuffer
with remove()
releases all resources associated with the object. This includes destroying the all the segment data, track buffers, and decoders. The media element will also remove the appropriate tracks from audioTracks
, videoTracks
, & textTracks
and fire the necessary change
events. Playback may become degraded or stop if the currently selected VideoTrack
or the only enabled AudioTracks
are removed.
The MediaSource object represents a source of media data for an HTMLMediaElement. It keeps track of the readyState
for this source as well as a list of SourceBuffer
objects that can be used to add media data to the presentation. MediaSource objects are created by the web application and then attached to an HTMLMediaElement. The application uses the SourceBuffer
objects in sourceBuffers
to add media data to this source. The HTMLMediaElement fetches this media data from the MediaSource
object when it is needed during playback.
[Constructor] interface MediaSource : EventTarget { // All the source buffers created by this object. readonly attribute SourceBufferList sourceBuffers; // Subset of sourceBuffers that provide data for the selected/enabled tracks. readonly attribute SourceBufferList activeSourceBuffers; // Adds another source buffer. SourceBuffer addSourceBuffer(DOMString type); enum State { "closed", "open", "ended" }; readonly attribute State readyState; enum EndOfStreamError { "network", "decode" }; void endOfStream(optional EndOfStreamError error); };
The sourceBuffers
attribute contains the list of SourceBuffer
objects associated with this MediaSource
. When readyState
equals "closed"
this list will be empty. Once readyState
transitions to "open"
SourceBuffer objects can be added to this list by using addSourceBuffer()
.
The activeSourceBuffers
attribute contains the subset of sourceBuffers
that represents the active source buffers.
The addSourceBuffer(type)
method must run the following steps:
INVALID_ACCESS_ERR
exception and abort these steps.SourceBuffer
objects in sourceBuffers
, then throw a NOT_SUPPORTED_ERR
exception and abort these steps.QUOTA_EXCEEDED_ERR
exception and abort these steps.readyState
attribute is not in the "open"
state then throw an INVALID_STATE_ERR
exception and abort these steps.SourceBuffer
object and associated resources.sourceBuffers
and fire a addsourcebuffer
on that object.The readyState
attribute indicates the current state of the MediaSource
object. It can have the following values:
"closed"
"open"
SourceBuffer
objects in sourceBuffers
."ended"
endOfStream()
has been called. Appending data to SourceBuffer
objects in this state is not allowed.When the MediaSource
is created readyState
must be set to "closed"
.
"network"
error
attribute to be set to MediaError.MEDIA_ERR_NETWORK
"decode"
error
attribute to be set to MediaError.MEDIA_ERR_DECODE
The endOfStream(error)
method must run the following steps:
readyState
attribute is not in the "open"
state then throw an INVALID_STATE_ERR
exception and abort these steps.readyState
attribute value to "ended"
.append()
has been played."network"
"decode"
INVALID_ACCESS_ERR
exception.Event name | Interface | Dispatched when... |
---|---|---|
sourceopen |
Event |
When readyState transitions from "closed" to "open" or from "ended" to "open" . |
sourceended |
Event |
When readyState transitions from "open" to "ended" . |
sourceclose |
Event |
When readyState transitions from "open" to "closed" or "ended" to "closed" . |
A MediaSource
object can be attached to a media element by assigning a MediaSource object URL to the media element src
attribute or the src attribute of a <source> inside a media element. MediaSource object URLs are created by passing a MediaSource object to window.URL.createObjectURL().
The following steps are run when a media element attempts the resource fetch algorithm with a MediaSource object URL.
readyState
is NOT set to "closed"
MEDIA_ERR_SRC_NOT_SUPPORTED
error.readyState
attribute to "open"
sourceopen
.append()
The following steps are run in any case where the media element is going to transition to NETWORK_EMPTY
and fire an emptied
event. These steps should be run right before the transition.
readyState
attribute to "closed"
SourceBuffer
objects from sourceBuffers
and fire a removesourcebuffer
event for each one.sourceclose
.seeking algorithm
starts and has reached the stage where it is about to fire the seeking
event.readyState
attribute is set to "ended"
readyState
attribute to "open"
sourceopen
on the MediaSource
object.seeking algorithm
fires the seeking
eventSourceBuffer
object in activeSourceBuffers
activeSourceBuffers
is missing media segments for the desired seek pointHTMLMediaElement.readyState
attribute to HAVE_METADATA
and fire the appropriate event for this transition.append()
. The web application can use buffered
to determine what the media element needs to resume playback.seeking algorithm
and fires the seeked
event indicating that the seek has completed.The following steps are periodically run during playback to make sure that all of the SourceBuffer
objects in activeSourceBuffers
have enough data to ensure uninterrupted playback. Appending new segments and changes to activeSourceBuffers
also cause these steps to run because they affect the conditions that trigger state transitions. The web application can monitor changes in HTMLMediaElement.readyState
to drive media segment appending.
buffered
for all objects in activeSourceBuffers
do not contain TimeRanges
for the current playback position:HTMLMediaElement.readyState
attribute to HAVE_METADATA
and fire the appropriate event for this transition.buffered
for all objects in activeSourceBuffers
contain TimeRanges
that include the current playback position and enough data to ensure uninterrupted playback:HTMLMediaElement.readyState
attribute to HAVE_ENOUGH_DATA
and fire the appropriate event for this transition.HAVE_CURRENT_DATA
.buffered
for at least one object in activeSourceBuffers
contains a TimeRange
that includes the current playback position but not enough data to ensure uninterrupted playback:HTMLMediaElement.readyState
attribute to HAVE_FUTURE_DATA
and fire the appropriate event for this transition.HAVE_CURRENT_DATA
.buffered
for at least one object in activeSourceBuffers
contains a TimeRange
that ends at the current playback position and does not have a range covering the time immediately after the current position:HTMLMediaElement.readyState
attribute to HAVE_CURRENT_DATA
and fire the appropriate event for this transition.During playback activeSourceBuffers
needs to be updated if the selected video track
, the enabled audio tracks
, or a text track mode
changes. When one or more of these changes occur the following steps need to be followed.
SourceBuffer
associated with the previously selected video track is not associated with any other enabled tracks then remove it from activeSourceBuffers
SourceBuffer
associated with the newly selected video track is not already in activeSourceBuffers
then add it.SourceBuffer
associated with this track is not associated with any other enabled or selected trackSourceBuffer
associated with the audio track from activeSourceBuffers
SourceBuffer
associated with this track is not already in activeSourceBuffers
SourceBuffer
associated with the audio track to activeSourceBuffers
mode
becomes "disabled"
and the SourceBuffer
associated with this track is not associated with any other enabled or selected trackSourceBuffer
associated with the text track from activeSourceBuffers
mode
becomes "showing"
or "hidden"
and the SourceBuffer
associated with this track is not already in activeSourceBuffers
SourceBuffer
associated with the text track to activeSourceBuffers
interface SourceBuffer : EventTarget { // Returns the time ranges buffered. readonly attribute TimeRanges buffered; // Append segment data. void append(Uint8Array data); // Abort the current segment append sequence. void abort(); };
The buffered
attribute indicates what TimeRanges
are buffered in the SourceBuffer
. When attribute is read the following steps must occur:
sourceBuffers
attribute of the MediaSource
object that created it then throw an INVALID_STATE_ERR
exception and abort these steps.TimeRanges
for the media segments buffered.The append(data)
method must run the following steps:
INVALID_ACCESS_ERR
exception and abort these steps.sourceBuffers
attribute of the MediaSource
object that created it then throw an INVALID_STATE_ERR
exception and abort these steps.readyState
attribute of the MediaSource
object that created this object is not in the "open"
state then throw an INVALID_STATE_ERR
exception and abort these steps.HTMLMediaElement.readyState
attribute is HAVE_NOTHING
:HTMLMediaElement.readyState
attribute to HAVE_METADATA
and fire the appropriate event for this transition.HTMLMediaElement.readyState
attribute is greater than HAVE_CURRENT_DATA
and the initialization segment contains the first video or first audio track in the presentation:HTMLMediaElement.readyState
attribute to HAVE_METADATA
and fire the appropriate event for this transition.
audioTracks
AudioTrack
and mark it as enabled.SourceBuffer
to activeSourceBuffers
.AudioTrack
for each audio track in the initialization segment.videoTracks
:VideoTrack
and mark it as selected.SourceBuffer
to activeSourceBuffers
.VideoTrack
for each video track in the initialization segment.textTracks
TextTrack
for each text track in the initialization segment.mode
is "showing"
or "hidden"
then add this SourceBuffer
to activeSourceBuffers
.HTMLMediaElement.readyState
attribute is HAVE_METADATA
and data causes all objects in activeSourceBuffers
to have media data for the current playback position.HTMLMediaElement.readyState
attribute to HAVE_CURRENT_DATA
and fire the appropriate event for this transition.HTMLMediaElement.readyState
attribute is HAVE_CURRENT_DATA
and data causes all objects in activeSourceBuffers
to have media data beyond the current playback position.HTMLMediaElement.readyState
attribute to HAVE_FUTURE_DATA
and fire the appropriate event for this transition.HTMLMediaElement.readyState
attribute is HAVE_FUTURE_DATA
and data causes all objects in activeSourceBuffers
to have enough data to start playback.HTMLMediaElement.readyState
attribute to HAVE_ENOUGH_DATA
and fire the appropriate event for this transition.The abort()
method must run the following steps:
sourceBuffers
attribute of the MediaSource
object that created it then throw an INVALID_STATE_ERR
exception and abort these steps.readyState
attribute of the MediaSource
object that created this object is not in the "open"
state then throw an INVALID_STATE_ERR
exception and abort these steps.buffered
will reflect what data, if any, was kept.SourceBufferList is a simple container object for SourceBuffer
objects. It provides a read-only array accessor, fires events when the list is modified, and provides a way to remove SourceBuffer objects from the list.
interface SourceBufferList : EventTarget { readonly attribute unsigned long length; getter SourceBuffer (unsigned long index); void remove(SourceBuffer buffer); };
The length
attribute indicates the number of SourceBuffer
objects in the list.
The getter SourceBuffer (unsigned long index)
method allows the SourceBuffer objects in the list to be accessed with an array operator (i.e. []). This method must run the following steps:
length
attribute then return undefined and abort these steps.SourceBuffer
object in the list.The remove(buffer)
method must run the following steps:
INVALID_ACCESS_ERR
exception and abort these steps.INVALID_STATE_ERR
exception and abort these steps.NOT_FOUND_ERR
exception and abort these steps.audioTracks
, videoTracks
, and textTracks
for all tracks associated with buffer and fire a simple event named change
on the modified lists.removesourcebuffer
event.Event name | Interface | Dispatched when... |
---|---|---|
addsourcebuffer |
Event |
When a SourceBuffer is added to the list. |
removesourcebuffer |
Event |
When a SourceBuffer is removed from the list. |
The bytes provided through append()
for a SourceBuffer
form a logical byte stream. The format of this byte stream depends on the media container format in use and is defined in a byte stream format specification. Byte stream format specifications based on WebM and the ISO Base Media File Format are provided below. If these formats are supported then the byte stream formats described below MUST be supported.
This section provides general requirements for all byte stream formats:
Byte stream specifications must at a minimum define constraints which ensure that the above requirements hold. Additional constraints may be defined, for example to simplify implementation.
Initialization segments are an optimization. They allow a byte stream format to avoid duplication of information in Media Segments that is the same for many Media Segments. Byte stream format specifications need not specify Initialization Segment formats, however. They may instead require that such information is duplicated in every Media Segment.
This section defines segment formats for implementations that choose to support WebM.
A WebM initialization segment must contain a subset of the elements at the start of a typical WebM file.
The following rules apply to WebM initialization segments:
A WebM media segment is a single Cluster element.
The following rules apply to WebM media segments:
The timestamp in the first block of the first media segment appended establishes the starting timestamp for the presentation timeline. All media segments appended after this first segment are expected to have timestamps greater than or equal to this timestamp.
If for some reason a web application doesn't want to append data at the beginning of the timeline, it can establish the starting timestamp by appending a Cluster element that only contains a Timecode element with the presentation start time. This must be done before any other media segments are appended.
A SimpleBlock element with its Keyframe flag set signals the location of a random access point for that track. Media segments containing multiple tracks are only considered a random access point if the first SimpleBlock for each track has its Keyframe flag set. The order of the multiplexed blocks should conform to the WebM Muxer Guidelines.
This section defines segment formats for implementations that choose to support the ISO Base Media File Format ISO/IEC 14496-12 (ISO BMFF).
An ISO BMFF initialization segment shall contain a single Movie Header Box (moov). The tracks in the Movie Header Box shall not contain any samples (i.e. the entry_count in the stts, stsc and stco boxes shall be set to zero). A Movie Extends (mvex) box shall be contained in the Movie Header Box to indicate that Movie Fragments are to be expected.
The initialization segment may contain Edit Boxes (edts) which provide a mapping of composition times for each track to the global presentation time.
An ISO BMFF media segment shall contain a single Movie Fragment Box (moof) followed by one or more Media Data Boxes (mdat).
The following rules shall apply to ISO BMFF media segments:
The Track Fragment Decode Time Box is defined in ISO/IEC 14496-12 Amendment 3.
The earliest presentation timestamp of any sample of the first media segment appended establishes the starting timestamp for the presentation timeline. All media segments appended after this first segment are expected to have presentation timestamps greater than or equal to this timestamp.
If for some reason a web application doesn't want to append data at the beginning of the timeline, it can establish the starting timestamp by appending a Movie Fragment Box containing a Track Fragment Box containing a Track Fragment Decode Time Box. The start time of the presentation is then the presentation time of a hypothetical sample with zero composition offset. This must be done before any other media segments are appended.
A random access point as defined in this specification corresponds to a Stream Access Point of type 1 or 2 as defined in Annex I of ISO/IEC 14496-12 Amendment 3.
Example use of the Media Source Extensions
<script> function onSourceOpen(videoTag, e) { var mediaSource = e.target; var sourceBuffer = mediaSource.addSourceBuffer('video/webm; codecs="vorbis,vp8"'); videoTag.addEventListener('seeking', onSeeking.bind(videoTag, mediaSource)); videoTag.addEventListener('progress', onProgress.bind(videoTag, mediaSource)); var initSegment = GetInitializationSegment(); if (initSegment == null) { // Error fetching the initialization segment. Signal end of stream with an error. mediaSource.endOfStream("network"); return; } // Append the initialization segment. sourceBuffer.append(initSegment); // Append some initial media data. appendNextMediaSegment(mediaSource); } function appendNextMediaSegment(mediaSource) { if (mediaSource.readyState == "ended") return; // If we have run out of stream data, then signal end of stream. if (!HaveMoreMediaSegments()) { mediaSource.endOfStream(); return; } var mediaSegment = GetNextMediaSegment(); if (!mediaSegment) { // Error fetching the next media segment. mediaSource.endOfStream("network"); return; } mediaSource.sourceBuffers[0].append(mediaSegment); } function onSeeking(mediaSource, e) { var video = e.target; // Abort current segment append. mediaSource.sourceBuffers[0].abort(); // Notify the media segment loading code to start fetching data at the // new playback position. SeekToMediaSegmentAt(video.currentTime); // Append media segments from the new playback position. appendNextMediaSegment(mediaSource); appendNextMediaSegment(mediaSource); } function onProgress(mediaSource, e) { appendNextMediaSegment(mediaSource); } </script> <video id="v" autoplay> </video> <script> var video = document.getElementById('v'); var mediaSource = new MediaSource(); mediaSource.addEventListener('sourceopen', onSourceOpen.bind(this, video)); video.src = window.URL.createObjectURL(mediaSource); </script>
Version | Comment |
---|---|
02 July 2012 | Converted to the object-oriented API |
26 June 2012 | Converted to Editor's draft. |
0.5 | Minor updates before proposing to W3C HTML-WG. |
0.4 | Major revision. Adding source IDs, defining buffer model, and clarifying byte stream formats. |
0.3 | Minor text updates. |
0.2 | Updates to reflect initial WebKit implementation. |
0.1 | Initial Proposal |