Liking cljdoc? Tell your friends :D

audio.AudioBuffer

The AudioBuffer interface represents a short audio asset residing memory, created from an audio file using the audio.AudioContext.decodeAudioData() or from raw data using audio.AudioContext.createBuffer(). Once into an AudioBuffer, the audio can then be played by being passed an audio.AudioBufferSourceNode.

The AudioBuffer interface represents a short audio asset residing
memory, created from an audio file using the `audio.AudioContext.decodeAudioData()`
or from raw data using `audio.AudioContext.createBuffer()`. Once
into an AudioBuffer, the audio can then be played by being passed
an `audio.AudioBufferSourceNode`.
raw docstring

audio.AudioBufferSourceNode

The AudioBufferSourceNode interface is an audio.AudioScheduledSourceNode represents an audio source consisting of in-memory audio data, in an audio.AudioBuffer. It's especially useful for playing audio which has particularly stringent timing accuracy requirements, as for sounds that must match a specific rhythm and can be kept memory rather than being played from disk or the network.

The AudioBufferSourceNode interface is an `audio.AudioScheduledSourceNode`
represents an audio source consisting of in-memory audio data,
in an `audio.AudioBuffer`. It's especially useful for playing
audio which has particularly stringent timing accuracy requirements,
as for sounds that must match a specific rhythm and can be kept
memory rather than being played from disk or the network.
raw docstring

audio.AudioContext

The AudioContext interface represents an audio-processing graph from audio modules linked together, each represented by an audio.AudioNode.

The AudioContext interface represents an audio-processing graph
from audio modules linked together, each represented by an `audio.AudioNode`.
raw docstring

audio.AudioContextOptions

The AudioContextOptions dictionary is used to specify configuration when constructing a new audio.AudioContext object to represent graph of web audio nodes.

The AudioContextOptions dictionary is used to specify configuration
when constructing a new `audio.AudioContext` object to represent
graph of web audio nodes.
raw docstring

audio.AudioDestinationNode

AudioDestinationNode has no output (as it is the output, no more can be linked after it in the audio graph) and one input. The of channels in the input must be between 0 and the maxChannelCount or an exception is raised.

AudioDestinationNode has no output (as it is the output, no more
can be linked after it in the audio graph) and one input. The
of channels in the input must be between 0 and the maxChannelCount
or an exception is raised.
raw docstring

audio.AudioListener

The AudioListener interface represents the position and orientation the unique person listening to the audio scene, and is used in spatialization. All web.PannerNodes spatialize in relation the AudioListener stored in the web.BaseAudioContext.listener

The AudioListener interface represents the position and orientation
the unique person listening to the audio scene, and is used in
spatialization. All `web.PannerNode`s spatialize in relation
the AudioListener stored in the `web.BaseAudioContext.listener`
raw docstring

audio.AudioNode

The AudioNode interface is a generic interface for representing audio processing module. Examples include:

The AudioNode interface is a generic interface for representing
audio processing module. Examples include:
raw docstring

audio.AudioParam

The Web Audio API's AudioParam interface represents an audio-related usually a parameter of an audio.AudioNode (such as web.GainNode.gain).

The Web Audio API's AudioParam interface represents an audio-related
usually a parameter of an `audio.AudioNode` (such as `web.GainNode.gain`).
raw docstring

audio.AudioScheduledSourceNode

The AudioScheduledSourceNode interface—part of the Web Audio a parent interface for several types of audio source node interfaces share the ability to be started and stopped, optionally at specified Specifically, this interface defines the web.start() and web.stop() as well as the web.onended event handler.

The AudioScheduledSourceNode interface—part of the Web Audio
a parent interface for several types of audio source node interfaces
share the ability to be started and stopped, optionally at specified
Specifically, this interface defines the `web.start()` and `web.stop()`
as well as the `web.onended` event handler.
raw docstring

audio.AudioTrack

The AudioTrack interface represents a single audio track from of the HTML media elements, web.<audio> or web.<video>.

The AudioTrack interface represents a single audio track from
of the HTML media elements, `web.<audio>` or `web.<video>`.
raw docstring

audio.AudioTrackList

The AudioTrackList interface is used to represent a list of the tracks contained within a given HTML media element, with each represented by a separate audio.AudioTrack object in the list.

The AudioTrackList interface is used to represent a list of the
tracks contained within a given HTML media element, with each
represented by a separate `audio.AudioTrack` object in the list.
raw docstring

audio.AudioTrackList.ev

AudioTrackList Events.

AudioTrackList Events.
raw docstring

audio.AudioWorkletGlobalScope

The AudioWorkletGlobalScope interface of the Web Audio API represents global execution context for user-supplied code, which defines audio.AudioWorkletProcessor-derived classes. Each web.BaseAudioContext a single audio.AudioWorklet available under the web.audioWorklet which runs its code in a single AudioWorkletGlobalScope.

The AudioWorkletGlobalScope interface of the Web Audio API represents
global execution context for user-supplied code, which defines
`audio.AudioWorkletProcessor`-derived classes. Each `web.BaseAudioContext`
a single `audio.AudioWorklet` available under the `web.audioWorklet`
which runs its code in a single AudioWorkletGlobalScope.
raw docstring

audio.AudioWorkletNode

The AudioWorkletNode interface of the Web Audio API represents base class for a user-defined audio.AudioNode, which can be to an audio routing graph along with other nodes. It has an associated which does the actual audio processing in a Web Audio rendering

The AudioWorkletNode interface of the Web Audio API represents
base class for a user-defined `audio.AudioNode`, which can be
to an audio routing graph along with other nodes. It has an associated
which does the actual audio processing in a Web Audio rendering
raw docstring

audio.AudioWorkletProcessor

The AudioWorkletProcessor interface of the Web Audio API represents audio processing code behind a custom audio.AudioWorkletNode. lives in the audio.AudioWorkletGlobalScope and runs on the Audio rendering thread. In turn, an audio.AudioWorkletNode on it runs on the main thread.

The AudioWorkletProcessor interface of the Web Audio API represents
audio processing code behind a custom `audio.AudioWorkletNode`.
lives in the `audio.AudioWorkletGlobalScope` and runs on the
Audio rendering thread. In turn, an `audio.AudioWorkletNode`
on it runs on the main thread.
raw docstring

cljdoc is a website building & hosting documentation for Clojure/Script libraries

× close