Liking cljdoc? Tell your friends :D

web.audio.AudioContext

The AudioContext interface represents an audio-processing graph from audio modules linked together, each represented by an web.audio.AudioNode.

The AudioContext interface represents an audio-processing graph
from audio modules linked together, each represented by an `web.audio.AudioNode`.
raw docstring

base-latencycljs

(base-latency this)

Property.

[Read Only] [Experimental]

The baseLatency read-only property of the web.audio.AudioContext returns a double that represents the number of seconds of processing incurred by the AudioContext passing the audio from the web.audio.AudioDestinationNode the audio subsystem.

var baseLatency = audioCtx.baseLatency;

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/baseLatency

Property.

[Read Only]
[Experimental]

The baseLatency read-only property of the `web.audio.AudioContext`
returns a double that represents the number of seconds of processing
incurred by the AudioContext passing the audio from the `web.audio.AudioDestinationNode`
the audio subsystem.

`var baseLatency = audioCtx.baseLatency;`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/baseLatency`
sourceraw docstring

closecljs

(close this & args)

Method.

The close() method of the web.audio.AudioContext Interface the audio context, releasing any system audio resources that uses.

var audioCtx = new AudioContext(); audioCtx.close().then(function() { ... }); await audioCtx.close();

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/close

Method.

The close() method of the `web.audio.AudioContext` Interface
the audio context, releasing any system audio resources that
uses.

`var audioCtx = new AudioContext();
audioCtx.close().then(function() { ... });
await audioCtx.close();`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/close`
sourceraw docstring

create-java-script-nodecljs

(create-java-script-node this
                         buffer-size
                         num-input-channels
                         num-output-channels)

Method.

[Obsolute]

The following script illustrates the use of createJavaScriptNode():

var jsNode = audioCtx.createJavaScriptNode(bufferSize, numInputChannels, numOutputChannels);

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createJavaScriptNode

Method.

[Obsolute]

The following script illustrates the use of createJavaScriptNode():

`var jsNode = audioCtx.createJavaScriptNode(bufferSize, numInputChannels, numOutputChannels);`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createJavaScriptNode`
sourceraw docstring

create-media-element-sourcecljs

(create-media-element-source this & args)

Method.

For more details about media element audio source nodes, check the web.audio.MediaElementAudioSourceNode reference page.

var audioCtx = new AudioContext(); var source = audioCtx.createMediaElementSource(myMediaElement);

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaElementSource

Method.

For more details about media element audio source nodes, check
the `web.audio.MediaElementAudioSourceNode` reference page.

`var audioCtx = new AudioContext();
var source = audioCtx.createMediaElementSource(myMediaElement);`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaElementSource`
sourceraw docstring

create-media-stream-destinationcljs

(create-media-stream-destination this & args)

Method.

The web.streams.MediaStream is created when the node is created is accessible via the web.audio.MediaStreamAudioDestinationNode's attribute. This stream can be used in a similar way as a MediaStream via navigator.getUserMedia — it can, for example, be sent to remote peer using the RTCPeerConnection addStream() method.

var audioCtx = new AudioContext(); var destination = audioCtx.createMediaStreamDestination();

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaStreamDestination

Method.

The `web.streams.MediaStream` is created when the node is created
is accessible via the `web.audio.MediaStreamAudioDestinationNode`'s
attribute. This stream can be used in a similar way as a MediaStream
via `navigator.getUserMedia` — it can, for example, be sent to
remote peer using the RTCPeerConnection addStream() method.

`var audioCtx = new AudioContext();
var destination = audioCtx.createMediaStreamDestination();`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaStreamDestination`
sourceraw docstring

create-media-stream-sourcecljs

(create-media-stream-source this stream)

Method.

For more details about media stream audio source nodes, check the web.audio.MediaStreamAudioSourceNode reference page.

audioSourceNode = audioContext.createMediaStreamSource(stream);

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaStreamSource

Method.

For more details about media stream audio source nodes, check
the `web.audio.MediaStreamAudioSourceNode` reference page.

`audioSourceNode = audioContext.createMediaStreamSource(stream);`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaStreamSource`
sourceraw docstring

create-media-stream-track-sourcecljs

(create-media-stream-track-source this & args)

Method.

[Draft]

A web.media.MediaStreamTrackAudioSourceNode object which acts a source for audio data found in the specified audio track.

var audioCtx = new AudioContext(); var track = audioCtx.createMediaStreamTrackSource(track);

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaStreamTrackSource

Method.

[Draft]

A `web.media.MediaStreamTrackAudioSourceNode` object which acts
a source for audio data found in the specified audio track.

`var audioCtx = new AudioContext();
var track = audioCtx.createMediaStreamTrackSource(track);`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaStreamTrackSource`
sourceraw docstring

create-wave-tablecljs

(create-wave-table this & args)

Method.

[Obsolute]

The web.audio.AudioContext method createWaveTable() is now you should instead use the method createPeriodicWave().

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createWaveTable

Method.

[Obsolute]

The `web.audio.AudioContext` method createWaveTable() is now
you should instead use the method `createPeriodicWave()`.

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createWaveTable`
sourceraw docstring

get-output-timestampcljs

(get-output-timestamp this)

Method.

[Experimental]

The getOutputTimestamp() property of the web.audio.AudioContext returns a new AudioTimestamp object containing two correlated audio stream position values.

var timestamp = AudioContext.getOutputTimestamp()

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/getOutputTimestamp

Method.

[Experimental]

The getOutputTimestamp() property of the `web.audio.AudioContext`
returns a new AudioTimestamp object containing two correlated
audio stream position values.

`var timestamp = AudioContext.getOutputTimestamp()`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/getOutputTimestamp`
sourceraw docstring

output-latencycljs

(output-latency this)

Property.

[Read Only] [Experimental]

The outputLatency read-only property of the web.audio.AudioContext provides an estimation of the output latency of the current audio

var latency = audioCtx.outputLatency;

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/outputLatency

Property.

[Read Only]
[Experimental]

The outputLatency read-only property of the `web.audio.AudioContext`
provides an estimation of the output latency of the current audio

`var latency = audioCtx.outputLatency;`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/outputLatency`
sourceraw docstring

resumecljs

(resume this)

Method.

The resume() method of the web.audio.AudioContext interface the progression of time in an audio context that has previously suspended.

completePromise = audioContext.resume();

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/resume

Method.

The resume() method of the `web.audio.AudioContext` interface
the progression of time in an audio context that has previously
suspended.

`completePromise = audioContext.resume();`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/resume`
sourceraw docstring

suspendcljs

(suspend this & args)

Method.

The suspend() method of the web.audio.AudioContext Interface the progression of time in the audio context, temporarily halting hardware access and reducing CPU/battery usage in the process this is useful if you want an application to power down the audio when it will not be using an audio context for a while.

var audioCtx = new AudioContext(); audioCtx.suspend().then(function() { ... });

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/suspend

Method.

The suspend() method of the `web.audio.AudioContext` Interface
the progression of time in the audio context, temporarily halting
hardware access and reducing CPU/battery usage in the process
this is useful if you want an application to power down the audio
when it will not be using an audio context for a while.

`var audioCtx = new AudioContext();
audioCtx.suspend().then(function() { ... });`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/suspend`
sourceraw docstring

cljdoc is a website building & hosting documentation for Clojure/Script libraries

× close