summaryrefslogtreecommitdiff
path: root/dom/webidl/AudioContext.webidl
diff options
context:
space:
mode:
Diffstat (limited to 'dom/webidl/AudioContext.webidl')
-rw-r--r--dom/webidl/AudioContext.webidl122
1 files changed, 122 insertions, 0 deletions
diff --git a/dom/webidl/AudioContext.webidl b/dom/webidl/AudioContext.webidl
new file mode 100644
index 0000000000..c2f65abaf0
--- /dev/null
+++ b/dom/webidl/AudioContext.webidl
@@ -0,0 +1,122 @@
+/* -*- Mode: IDL; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
+/* This Source Code Form is subject to the terms of the Mozilla Public
+ * License, v. 2.0. If a copy of the MPL was not distributed with this file,
+ * You can obtain one at http://mozilla.org/MPL/2.0/.
+ *
+ * The origin of this IDL file is
+ * https://webaudio.github.io/web-audio-api/
+ *
+ * Copyright © 2012 W3C® (MIT, ERCIM, Keio), All Rights Reserved. W3C
+ * liability, trademark and document use rules apply.
+ */
+
+callback DecodeSuccessCallback = void (AudioBuffer decodedData);
+callback DecodeErrorCallback = void ();
+
+enum AudioContextState {
+ "suspended",
+ "running",
+ "closed"
+};
+
+dictionary PeriodicWaveConstraints {
+ boolean disableNormalization = false;
+};
+
+[Constructor,
+ Constructor(AudioChannel audioChannelType),
+ Pref="dom.webaudio.enabled"]
+interface AudioContext : EventTarget {
+
+ readonly attribute AudioDestinationNode destination;
+ readonly attribute float sampleRate;
+ readonly attribute double currentTime;
+ readonly attribute AudioListener listener;
+ readonly attribute AudioContextState state;
+ [Throws]
+ Promise<void> suspend();
+ [Throws]
+ Promise<void> resume();
+ [Throws]
+ Promise<void> close();
+ attribute EventHandler onstatechange;
+
+ [NewObject, Throws]
+ AudioBuffer createBuffer(unsigned long numberOfChannels, unsigned long length, float sampleRate);
+
+ [Throws]
+ Promise<AudioBuffer> decodeAudioData(ArrayBuffer audioData,
+ optional DecodeSuccessCallback successCallback,
+ optional DecodeErrorCallback errorCallback);
+
+ // AudioNode creation
+ [NewObject, Throws]
+ AudioBufferSourceNode createBufferSource();
+
+ [NewObject, Throws]
+ ConstantSourceNode createConstantSource();
+
+ [NewObject, Throws]
+ MediaStreamAudioDestinationNode createMediaStreamDestination();
+
+ [NewObject, Throws]
+ ScriptProcessorNode createScriptProcessor(optional unsigned long bufferSize = 0,
+ optional unsigned long numberOfInputChannels = 2,
+ optional unsigned long numberOfOutputChannels = 2);
+
+ [NewObject, Throws]
+ StereoPannerNode createStereoPanner();
+ [NewObject, Throws]
+ AnalyserNode createAnalyser();
+ [NewObject, Throws, UnsafeInPrerendering]
+ MediaElementAudioSourceNode createMediaElementSource(HTMLMediaElement mediaElement);
+ [NewObject, Throws, UnsafeInPrerendering]
+ MediaStreamAudioSourceNode createMediaStreamSource(MediaStream mediaStream);
+ [NewObject, Throws]
+ GainNode createGain();
+ [NewObject, Throws]
+ DelayNode createDelay(optional double maxDelayTime = 1);
+ [NewObject, Throws]
+ BiquadFilterNode createBiquadFilter();
+ [NewObject, Throws]
+ IIRFilterNode createIIRFilter(sequence<double> feedforward, sequence<double> feedback);
+ [NewObject, Throws]
+ WaveShaperNode createWaveShaper();
+ [NewObject, Throws]
+ PannerNode createPanner();
+ [NewObject, Throws]
+ ConvolverNode createConvolver();
+
+ [NewObject, Throws]
+ ChannelSplitterNode createChannelSplitter(optional unsigned long numberOfOutputs = 6);
+ [NewObject, Throws]
+ ChannelMergerNode createChannelMerger(optional unsigned long numberOfInputs = 6);
+
+ [NewObject, Throws]
+ DynamicsCompressorNode createDynamicsCompressor();
+
+ [NewObject, Throws]
+ OscillatorNode createOscillator();
+ [NewObject, Throws]
+ PeriodicWave createPeriodicWave(Float32Array real, Float32Array imag, optional PeriodicWaveConstraints constraints);
+
+};
+
+// Mozilla extensions
+partial interface AudioContext {
+ // Read AudioChannel.webidl for more information about this attribute.
+ [Pref="media.useAudioChannelAPI"]
+ readonly attribute AudioChannel mozAudioChannelType;
+
+ // These 2 events are dispatched when the AudioContext object is muted by
+ // the AudioChannelService. It's call 'interrupt' because when this event is
+ // dispatched on a HTMLMediaElement, the audio stream is paused.
+ [Pref="media.useAudioChannelAPI"]
+ attribute EventHandler onmozinterruptbegin;
+
+ [Pref="media.useAudioChannelAPI"]
+ attribute EventHandler onmozinterruptend;
+
+ // This method is for test only.
+ [ChromeOnly] AudioChannel testAudioChannelInAudioNodeStream();
+};