Liking cljdoc? Tell your friends :D

web.audio.AudioContext

The AudioContext interface represents an audio-processing graph from audio modules linked together, each represented by an web.audio.AudioNode.

The AudioContext interface represents an audio-processing graph
from audio modules linked together, each represented by an `web.audio.AudioNode`.
raw docstring

base-latencycljs

(base-latency this)

Property.

[Read Only] [Experimental]

The baseLatency read-only property of the web.audio.AudioContext returns a double that represents the number of seconds of processing incurred by the AudioContext passing the audio from the web.audio.AudioDestinationNode the audio subsystem.

var baseLatency = audioCtx.baseLatency;

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/baseLatency

Property.

[Read Only]
[Experimental]

The baseLatency read-only property of the `web.audio.AudioContext`
returns a double that represents the number of seconds of processing
incurred by the AudioContext passing the audio from the `web.audio.AudioDestinationNode`
the audio subsystem.

`var baseLatency = audioCtx.baseLatency;`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/baseLatency`
sourceraw docstring

closecljs

(close this & args)

Method.

The close() method of the web.audio.AudioContext Interface the audio context, releasing any system audio resources that uses.

var audioCtx = new AudioContext(); audioCtx.close().then(function() { ... }); await audioCtx.close();

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/close

Method.

The close() method of the `web.audio.AudioContext` Interface
the audio context, releasing any system audio resources that
uses.

`var audioCtx = new AudioContext();
audioCtx.close().then(function() { ... });
await audioCtx.close();`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/close`
sourceraw docstring

constructorcljs

(constructor & args)

Constructor.

The AudioContext() constructor creates a new web.audio.AudioContext object which represents an audio-processing graph, built from audio modules linked together, each represented by an web.audio.AudioNode.

options Optional An object based on the web.audio.AudioContextOptions dictionary that contains zero or more optional properties to configure the new context. Available properties are as follows: latencyHint Optional The type of playback that the context will be used for, as a value from the web.audio.AudioContextLatencyCategory enum or a double-precision floating-point value indicating the preferred maximum latency of the context in seconds. The user agent may or may not choose to meet this request; check the value of AudioContext.baseLatency to determine the true latency after creating the context. sampleRate Optional The sampleRate to be used by the AudioContext, specified in samples per second. The value may be any value supported by web.audio.AudioBuffer. If not specified, the preferred sample rate for the context's output device is used by default.

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/AudioContext

Constructor.

The AudioContext() constructor creates a new `web.audio.AudioContext` object which represents an audio-processing graph, built from audio modules linked together, each represented by an `web.audio.AudioNode`.

options Optional
An object based on the `web.audio.AudioContextOptions` dictionary that contains zero or more optional properties to configure the new context. Available properties are as follows:
`latencyHint` Optional
The type of playback that the context will be used for, as a value from the `web.audio.AudioContextLatencyCategory` enum or a double-precision floating-point value indicating the preferred maximum latency of the context in seconds. The user agent may or may not choose to meet this request; check the value of `AudioContext.baseLatency` to determine the true latency after creating the context.
`sampleRate` Optional
The `sampleRate` to be used by the AudioContext, specified in samples per second. The value may be any value supported by `web.audio.AudioBuffer`. If not specified, the preferred sample rate for the context's output device is used by default.

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/AudioContext`
sourceraw docstring

create-java-script-nodecljs

(create-java-script-node this
                         buffer-size
                         num-input-channels
                         num-output-channels)

Method.

[Obsolute]

The following script illustrates the use of createJavaScriptNode():

var jsNode = audioCtx.createJavaScriptNode(bufferSize, numInputChannels, numOutputChannels);

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createJavaScriptNode

Method.

[Obsolute]

The following script illustrates the use of createJavaScriptNode():

`var jsNode = audioCtx.createJavaScriptNode(bufferSize, numInputChannels, numOutputChannels);`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createJavaScriptNode`
sourceraw docstring

create-media-element-sourcecljs

(create-media-element-source this & args)

Method.

For more details about media element audio source nodes, check the web.audio.MediaElementAudioSourceNode reference page.

var audioCtx = new AudioContext(); var source = audioCtx.createMediaElementSource(myMediaElement);

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaElementSource

Method.

For more details about media element audio source nodes, check
the `web.audio.MediaElementAudioSourceNode` reference page.

`var audioCtx = new AudioContext();
var source = audioCtx.createMediaElementSource(myMediaElement);`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaElementSource`
sourceraw docstring

create-media-stream-destinationcljs

(create-media-stream-destination this & args)

Method.

The web.streams.MediaStream is created when the node is created is accessible via the web.audio.MediaStreamAudioDestinationNode's attribute. This stream can be used in a similar way as a MediaStream via navigator.getUserMedia — it can, for example, be sent to remote peer using the RTCPeerConnection addStream() method.

var audioCtx = new AudioContext(); var destination = audioCtx.createMediaStreamDestination();

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaStreamDestination

Method.

The `web.streams.MediaStream` is created when the node is created
is accessible via the `web.audio.MediaStreamAudioDestinationNode`'s
attribute. This stream can be used in a similar way as a MediaStream
via `navigator.getUserMedia` — it can, for example, be sent to
remote peer using the RTCPeerConnection addStream() method.

`var audioCtx = new AudioContext();
var destination = audioCtx.createMediaStreamDestination();`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaStreamDestination`
sourceraw docstring

create-media-stream-sourcecljs

(create-media-stream-source this stream)

Method.

For more details about media stream audio source nodes, check the web.audio.MediaStreamAudioSourceNode reference page.

audioSourceNode = audioContext.createMediaStreamSource(stream);

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaStreamSource

Method.

For more details about media stream audio source nodes, check
the `web.audio.MediaStreamAudioSourceNode` reference page.

`audioSourceNode = audioContext.createMediaStreamSource(stream);`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaStreamSource`
sourceraw docstring

create-media-stream-track-sourcecljs

(create-media-stream-track-source this & args)

Method.

[Draft]

A web.media.MediaStreamTrackAudioSourceNode object which acts a source for audio data found in the specified audio track.

var audioCtx = new AudioContext(); var track = audioCtx.createMediaStreamTrackSource(track);

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaStreamTrackSource

Method.

[Draft]

A `web.media.MediaStreamTrackAudioSourceNode` object which acts
a source for audio data found in the specified audio track.

`var audioCtx = new AudioContext();
var track = audioCtx.createMediaStreamTrackSource(track);`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createMediaStreamTrackSource`
sourceraw docstring

create-wave-tablecljs

(create-wave-table this & args)

Method.

[Obsolute]

The web.audio.AudioContext method createWaveTable() is now you should instead use the method createPeriodicWave().

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createWaveTable

Method.

[Obsolute]

The `web.audio.AudioContext` method createWaveTable() is now
you should instead use the method `createPeriodicWave()`.

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/createWaveTable`
sourceraw docstring

get-output-timestampcljs

(get-output-timestamp this)

Method.

[Experimental]

The getOutputTimestamp() property of the web.audio.AudioContext returns a new AudioTimestamp object containing two correlated audio stream position values.

var timestamp = AudioContext.getOutputTimestamp()

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/getOutputTimestamp

Method.

[Experimental]

The getOutputTimestamp() property of the `web.audio.AudioContext`
returns a new AudioTimestamp object containing two correlated
audio stream position values.

`var timestamp = AudioContext.getOutputTimestamp()`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/getOutputTimestamp`
sourceraw docstring

output-latencycljs

(output-latency this)

Property.

[Read Only] [Experimental]

The outputLatency read-only property of the web.audio.AudioContext provides an estimation of the output latency of the current audio

var latency = audioCtx.outputLatency;

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/outputLatency

Property.

[Read Only]
[Experimental]

The outputLatency read-only property of the `web.audio.AudioContext`
provides an estimation of the output latency of the current audio

`var latency = audioCtx.outputLatency;`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/outputLatency`
sourceraw docstring

resumecljs

(resume this)

Method.

The resume() method of the web.audio.AudioContext interface the progression of time in an audio context that has previously suspended.

completePromise = audioContext.resume();

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/resume

Method.

The resume() method of the `web.audio.AudioContext` interface
the progression of time in an audio context that has previously
suspended.

`completePromise = audioContext.resume();`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/resume`
sourceraw docstring

suspendcljs

(suspend this & args)

Method.

The suspend() method of the web.audio.AudioContext Interface the progression of time in the audio context, temporarily halting hardware access and reducing CPU/battery usage in the process this is useful if you want an application to power down the audio when it will not be using an audio context for a while.

var audioCtx = new AudioContext(); audioCtx.suspend().then(function() { ... });

See also: https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/suspend

Method.

The suspend() method of the `web.audio.AudioContext` Interface
the progression of time in the audio context, temporarily halting
hardware access and reducing CPU/battery usage in the process
this is useful if you want an application to power down the audio
when it will not be using an audio context for a while.

`var audioCtx = new AudioContext();
audioCtx.suspend().then(function() { ... });`

See also: `https://developer.mozilla.org/en-US/docs/Web/API/AudioContext/suspend`
sourceraw docstring

cljdoc is a website building & hosting documentation for Clojure/Script libraries

× close