--- title: 音声と動画の配信 slug: Web/Guide/Audio_and_video_delivery tags: - Audio - HTML5 - Media - NeedsTranslation - Video translation_of: Web/Guide/Audio_and_video_delivery ---
「静的」メディアファイルからアダプティブライブストリームまで、さまざまな方法で Web 上にオーディオとビデオを配信できます。この記事は、Web ベースのメディアのさまざまな配信メカニズムおよび一般的なブラウザとの互換性を探るための出発点として意図されています。
Whether we are dealing with pre-recorded audio files or live streams, the mechanism for making them available through the browser's {{ htmlelement("audio")}} and {{ htmlelement("video")}} elements remains pretty much the same. Currently, to support all browsers we need to specify two formats, although with the adoption of MP3 and MP4 formats in Firefox and Opera, this is changing fast. You can find compatibility information in the following places:
To deliver video and audio, the general workflow is usually something like this:
document.createElement('video')
perhaps?)<audio controls preload="auto"> <source src="audiofile.mp3" type="audio/mpeg"> <!-- fallback for browsers that don't suppport mp3 --> <source src="audiofile.ogg" type="audio/ogg"> <!-- fallback for browsers that don't support audio tag --> <a href="audiofile.mp3">download audio</a> </audio>
The code above will create an audio player that attempts to preload as much audio as possible for smooth playback.
Note: The preload attribute may be ignored by some mobile browsers.
For further info see Cross Browser Audio Basics (HTML5 Audio In Detail)
<video controls width="640" height="480" poster="initialimage.png" autoplay muted> <source src="videofile.mp4" type="video/mp4"> <!-- fallback for browsers that don't suppport mp4 --> <source src="videofile.webm" type="video/webm"> <!-- specifying subtitle files --> <track src="subtitles_en.vtt" kind="subtitles" srclang="en" label="English"> <track src="subtitles_no.vtt" kind="subtitles" srclang="no" label="Norwegian"> <!-- fallback for browsers that don't support video tag --> <a href="videofile.mp4">download video</a> </video>
The code above creates a video player of dimensions 640x480 pixels, displaying a poster image until the video is played. We instruct the video to autoplay but to be muted by default.
Note: The autoplay attribute may be ignored by some mobile browsers.
For further info see <video> element and Creating a cross-browser video player.
You can create a more comprehensive Fallback using Flash. Using flashmediaelement.swf is one way.
<audio controls> <source src="audiofile.mp3" type="audio/mpeg"> <source src="audiofile.ogg" type="audio/ogg"> <!-- fallback for non supporting browsers goes here --> <a href="audiofile.mp3">download audio</a> <object width="320" height="30" type="application/x-shockwave-flash" data="flashmediaelement.swf"> <param name="movie" value="flashmediaelement.swf" /> <param name="flashvars" value="controls=true&isvideo=false&file=audiofile.mp3" /> </object> </audio>
The process is very similar with video — just remember to set isvideo=true
in the flashvars value
parameters.
More options for Fallbacks.
var myAudio = document.createElement('audio'); if (myAudio.canPlayType('audio/mpeg')) { myAudio.setAttribute('src','audiofile.mp3'); } else if (myAudio.canPlayType('audio/ogg')) { myAudio.setAttribute('src','audiofile.ogg'); } myAudio.currentTime = 5; myAudio.play();
We set the source of the audio depending on the type of audio file the browser supports, then set the play-head 5 seconds in and attempt to play it.
Note: Play will be ignored by some mobile browsers unless issued by a user-initiated event.
It's also possible to feed an {{ htmlelement("audio") }} element a base64 encoded WAV file, allowing to generate audio on the fly:
<audio id="player" src="data:audio/x-wav;base64,UklGRvC..."></audio>
Speak.js employs this technique. Try the demo.
var myVideo = document.createElement('video'); if (myVideo.canPlayType('video/mp4')) { myVideo.setAttribute('src','videofile.mp4'); } else if (myVideo.canPlayType('video/webm')) { myVideo.setAttribute('src','videofile.webm'); } myVideo.width = 480; myVideo.height = 320;
We set the source of the video depending on the type of video file the browser supports we then set the width and height of the video.
var context; var request; var source; try { context = new AudioContext(); request = new XMLHttpRequest(); request.open("GET","http://jplayer.org/audio/mp3/RioMez-01-Sleep_together.mp3",true); request.responseType = "arraybuffer"; request.onload = function() { context.decodeAudioData(request.response, function(buffer) { source = context.createBufferSource(); source.buffer = buffer; source.connect(context.destination); // auto play source.start(0); // start was previously noteOn }); }; request.send(); } catch(e) { alert('web audio api not supported'); }
In this example we retrieve an MP3 file via XHR, load it into a source and play it (Try it for yourself). Find more about Web Audio API basics in Using the Web Audio API.
It's also possible to retrieve a live stream from a webcam and/or microphone using getUserMedia
and the Stream API. This makes up part of a wider technology known as WebRTC (Web Real-Time Communications) and is compatible with the latest versions of Chrome, Firefox and Opera.
To grab the stream from your webcam, first set up a {{htmlelement("video")}} element:
<video id="webcam" width="480" height="360"></video>
Next, if supported connect the webcam source to the video element:
if (navigator.mediaDevices) { navigator.mediaDevices.getUserMedia({ video: true, audio: false }) .then(function onSuccess(stream) { var video = document.getElementById('webcam'); video.autoplay = true; video.srcObject = stream; }) .catch(function onError() { alert('There has been a problem retreiving the streams - are you running on file:/// or did you disallow access?'); }); } else { alert('getUserMedia is not supported in this browser.'); }
To find out more, read our {{domxref("MediaDevices.getUserMedia")}} page.
New standards are being rolled out to allow your browser to grab media from your mic or camera using getUserMedia
and record it instantly using the new MediaRecorder API. You take the stream you receive from getUserMedia
, pass it to a MediaRecorder
object, take the resulting output and feed it to your audio or video source*.
The main mechanism is outlined below:
navigator.mediaDevices.getUserMedia({audio:true}) .then(function onSuccess(stream) { var recorder = new MediaRecorder(stream); var data = []; recorder.ondataavailable = function(e) { data.push(e.data); }; recorder.start(); recorder.onerror = function(e) { throw e.error || new Error(e.name); // e.name is FF non-spec } recorder.onstop = function(e) { var audio = document.createElement('audio'); audio.src = window.URL.createObjectURL(new Blob(data)); } setTimeout(function() { rec.stop(); }, 5000); }) .catch(function onError(error) { console.log(error.message); });
See MediaRecorder API for more details.
Media Source Extensions is a W3C working draft that plans to extend {{domxref("HTMLMediaElement")}} to allow JavaScript to generate media streams for playback. Allowing JavaScript to generate streams facilitates a variety of use cases like adaptive streaming and time shifting live streams.
Encrypted Media Extensions is a W3C proposal to extend HTMLMediaElement
, providing APIs to control playback of protected content.
The API supports use cases ranging from simple clear key decryption to high value video (given an appropriate user agent implementation). License/key exchange is controlled by the application, facilitating the development of robust playback applications supporting a range of content decryption and protection technologies.
One of the principle uses of EME is to allow browsers to implement DRM (Digital Rights Management), which helps to prevent web-based content (especially video) from being copied.
New formats and protocols are being rolled out to facilitate adaptive streaming. Adaptive streaming media means that the bandwidth and typically quality of the stream can change in real-time in reaction to the user's available bandwidth. Adaptive streaming is often used in conjunction with live streaming where smooth delivery of audio or video is paramount.
The main formats used for adaptive streaming are HLS and MPEG-DASH. MSE has been designed with DASH in mind. MSE defines byte streams according to ISOBMFF and M2TS (both supported in DASH, the latter supported in HLS). Generally speaking, if you are interested in standards, are looking for flexibility, or wish to support most modern browsers, you are probably better off with DASH.
Note: Currently Safari does not support DASH although dash.js will work on newer versions of Safari scheduled for release with OSX Yosemite.
DASH also provides a number of profiles including simple onDemand profiles that no preprocessing and splitting up of media files. There are also a number of cloud based services that will convert your media to both HLS and DASH.
For further information see Live streaming web audio and video.
You may decide that you want your audio or video player to have a consistent look across browsers, or just wish to tweak it to match your site. The general technique for achieving this is to omit the controls
attribute so that the default browser controls are not displayed, create custom controls using HTML and CSS, then use JavaScript to link your controls to the audio/video API.
If you need something extra, it's possible to add features that are not currently present in default players, such as playback rate, quality stream switches or even audio spectrums. You can also choose how to make your player responsive — for example you might remove the progress bar under certain conditions.
You may detect click, touch and/or keyboard events to trigger actions such as play, pause and scrubbing. It's often important to remember keyboard controls for user convenience and accessibility.
A quick example — first set up your audio and custom controls in HTML:
<audio id="my-audio" src="http://jPlayer.org/audio/mp3/Miaow-01-Tempered-song.mp3"></audio> <button id="my-control">play</button>
add a bit of JavaScript to detect events to play and pause the audio:
window.onload = function() { var myAudio = document.getElementById('my-audio'); var myControl = document.getElementById('my-control'); function switchState() { if (myAudio.paused == true) { myAudio.play(); myControl.innerHTML = "pause"; } else { myAudio.pause(); myControl.innerHTML = "play"; } } function checkKey(e) { if (e.keycode == 32 ) { //spacebar switchState(); } } myControl.addEventListener('click', function() { switchState(); }, false); window.addEventListener( "keypress", checkKey, false ); }
You can try this example out here. For more information, see Creating your own custom audio player.
While stopping the playback of media is as easy as calling the element's pause()
method, the browser keeps downloading the media until the media element is disposed of through garbage collection.
Here's a trick that stops the download at once:
var mediaElement = document.querySelector("#myMediaElementID"); mediaElement.removeAttribute("src");
mediaElement.load();
By removing the media element's src
attribute and invoking the load() method, you release the resources associated with the video, which stops the network download. You must call load()
after removing the attribute, because just removing the src
attribute does not invoke the load algorithm. If the <video>
element also has <source>
element descendants, those should also be removed before calling load()
.
Note that just setting the src
attribute to an empty string will actually cause the browser to treat it as though you're setting a video source to a relative path. This causes the browser to attempt another download to something that is unlikely to be a valid video.
Media elements provide support for moving the current playback position to specific points in the media's content. This is done by setting the value of the currentTime
property on the element; see {{ domxref("HTMLMediaElement") }} for further details on the element's properties. Simply set the value to the time, in seconds, at which you want playback to continue.
You can use the element's seekable
property to determine the ranges of the media that are currently available for seeking to. This returns a {{ domxref("TimeRanges") }} object listing the ranges of times that you can seek to.
var mediaElement = document.querySelector('#mediaElementID'); mediaElement.seekable.start(0); // Returns the starting time (in seconds) mediaElement.seekable.end(0); // Returns the ending time (in seconds) mediaElement.currentTime = 122; // Seek to 122 seconds mediaElement.played.end(0); // Returns the number of seconds the browser has played
When specifying the URI of media for an {{ HTMLElement("audio") }} or {{ HTMLElement("video") }} element, you can optionally include additional information to specify the portion of the media to play. To do this, append a hash mark ("#") followed by the media fragment description.
A time range is specified using the syntax:
#t=[starttime][,endtime]
The time can be specified as a number of seconds (as a floating-point value) or as an hours/minutes/seconds time separated with colons (such as 2:05:01 for 2 hours, 5 minutes, and 1 second).
A few examples:
Note: The playback range portion of the media element URI specification was added to Gecko 9.0 {{ geckoRelease("9.0") }}. At this time, this is the only part of the Media Fragments URI specification implemented by Gecko, and it can only be used when specifying the source for media elements, and not in the address bar.
Starting in Gecko 2.0 {{ geckoRelease("2.0") }}, error handling has been revised to match the latest version of the HTML5 specification. Instead of the error
event being dispatched to the media element itself, it now gets delivered to the child {{ HTMLElement("source") }} elements corresponding to the sources resulting in the error.
This lets you detect which sources failed to load, which may be useful. Consider this HTML:
<video> <source id="mp4_src" src="video.mp4" type='video/mp4; codecs="avc1.42E01E, mp4a.40.2"'> </source> <source id="3gp_src" src="video.3gp" type='video/3gpp; codecs="mp4v.20.8, samr"'> </source> <source id="ogg_src" src="video.ogv" type='video/ogv; codecs="theora, vorbis"'> </source> </video>
Since Firefox doesn't support MP4 and 3GP on some platforms due to their patent-encumbered nature, the {{ HTMLElement("source") }} elements with the IDs "mp4_src" and "3gp_src" will receive error
events before the Ogg resource is loaded. The sources are tried in the order in which they appear, and once one loads successfully, the remaining sources aren't tried at all.
Use the following verified sources within your audio and video elements to check support.
type="audio/mpeg"
): http://jPlayer.org/audio/mp3/Miaow-01-Tempered-song.mp3 (play the MP3 audio live.)type="audio/mp4"
): http://jPlayer.org/audio/m4a/Miaow-01-Tempered-song.m4a (play the MP4 audio live.)type="audio/ogg"
): http://jPlayer.org/audio/ogg/Miaow-01-Tempered-song.ogg (play the OGG audio live.)type="video/mp4"
): http://jPlayer.org/video/m4v/Big_Buck_Bunny_Trailer.m4v (play the MP4 video live.)type="video/webm"
): http://jPlayer.org/video/webm/Big_Buck_Bunny_Trailer.webm (play the WebM video live.)type="video/ogg"
): http://jPlayer.org/video/ogv/Big_Buck_Bunny_Trailer.ogv (play the OGG video live.)If these don't play then the browser you are testing doesn't support the given format. Consider using a different format or using a fallback.
If these work but the files you are supplying don't, there are two possible issues:
Although this is usually supported, you may need to add the following to your media server's .htaccess
file.
# AddType TYPE/SUBTYPE EXTENSION AddType audio/mpeg mp3 AddType audio/mp4 m4a AddType audio/ogg ogg AddType audio/ogg oga AddType video/mp4 mp4 AddType video/mp4 m4v AddType video/ogg ogv AddType video/webm webm AddType video/webm webmv
Your files may have been encoded incorrectly — try encoding using one of the following tools, which are proven to be pretty reliable:
To detect that all child {{ HTMLElement("source") }} elements have failed to load, check the value of the media element's networkState
attribute. If this is HTMLMediaElement.NETWORK_NO_SOURCE
, you know that all the sources failed to load.
If at that point you add another source, by inserting a new {{ HTMLElement("source") }} element as a child of the media element, Gecko attempts to load the specified resource.
Another way to show the fallback content of a video, when none of the sources could be decoded in the current browser, is to add an error handler on the last source element. Then you can replace the video with its fallback content:
<video controls> <source src="dynamicsearch.mp4" type="video/mp4"></source> <a href="dynamicsearch.mp4"> <img src="dynamicsearch.jpg" alt="Dynamic app search in Firefox OS"> </a> <p>Click image to play a video demo of dynamic app search</p> </video>
var v = document.querySelector('video'), sources = v.querySelectorAll('source'), lastsource = sources[sources.length-1]; lastsource.addEventListener('error', function(ev) { var d = document.createElement('div'); d.innerHTML = v.innerHTML; v.parentNode.replaceChild(d, v); }, false);
A number of audio and video JavaScript libaries exist. The most popular libraries allow you to choose a consistent player design over all browsers and provide a fallback for browsers that don't support audio and video natively. Fallbacks often use Adobe Flash or Microsoft Silverlight plugins. Other functionality such as the track element for subtitles can also be provided through media libraries.
This article provides a basic guide to creating an HTML5 audio player that works cross browser, with all the associated attributes, properties and events explained, and a quick guide to custom controls created using the Media API.
playbackRate
property allows us to change the speed or rate at which a piece of web audio or video is playing. This article explains it in detail.Note: Firefox OS versions 1.3 and above support the RTSP protocol for streaming video delivery. A fallback solution for older versions would be to use <video>
along with a suitable format for Gecko (such as WebM) to serve fallback content. More information will be published on this in good time.