зеркало из https://github.com/mozilla/gecko-dev.git
254 строки
7.1 KiB
C++
254 строки
7.1 KiB
C++
/* -*- Mode: C++; tab-width: 2; indent-tabs-mode: nil; c-basic-offset: 2 -*- */
|
|
/* vim:set ts=2 sw=2 sts=2 et cindent: */
|
|
/* This Source Code Form is subject to the terms of the Mozilla Public
|
|
* License, v. 2.0. If a copy of the MPL was not distributed with this
|
|
* file, You can obtain one at http://mozilla.org/MPL/2.0/. */
|
|
|
|
#ifndef AudioParam_h_
|
|
#define AudioParam_h_
|
|
|
|
#include "AudioParamTimeline.h"
|
|
#include "nsWrapperCache.h"
|
|
#include "nsCycleCollectionParticipant.h"
|
|
#include "AudioNode.h"
|
|
#include "mozilla/dom/TypedArray.h"
|
|
#include "WebAudioUtils.h"
|
|
#include "js/TypeDecls.h"
|
|
|
|
namespace mozilla {
|
|
|
|
namespace dom {
|
|
|
|
class AudioParam final : public nsWrapperCache,
|
|
public AudioParamTimeline
|
|
{
|
|
virtual ~AudioParam();
|
|
|
|
public:
|
|
AudioParam(AudioNode* aNode,
|
|
uint32_t aIndex,
|
|
float aDefaultValue,
|
|
const char* aName);
|
|
|
|
NS_IMETHOD_(MozExternalRefCountType) AddRef(void);
|
|
NS_IMETHOD_(MozExternalRefCountType) Release(void);
|
|
NS_DECL_CYCLE_COLLECTION_SCRIPT_HOLDER_NATIVE_CLASS(AudioParam)
|
|
|
|
AudioContext* GetParentObject() const
|
|
{
|
|
return mNode->Context();
|
|
}
|
|
|
|
JSObject* WrapObject(JSContext* aCx, JS::Handle<JSObject*> aGivenProto) override;
|
|
|
|
// We override SetValueCurveAtTime to convert the Float32Array to the wrapper
|
|
// object.
|
|
AudioParam* SetValueCurveAtTime(const Float32Array& aValues,
|
|
double aStartTime,
|
|
double aDuration,
|
|
ErrorResult& aRv)
|
|
{
|
|
if (!WebAudioUtils::IsTimeValid(aStartTime)) {
|
|
aRv.Throw(NS_ERROR_DOM_NOT_SUPPORTED_ERR);
|
|
return this;
|
|
}
|
|
aValues.ComputeLengthAndData();
|
|
|
|
aStartTime = std::max(aStartTime, GetParentObject()->CurrentTime());
|
|
EventInsertionHelper(aRv, AudioTimelineEvent::SetValueCurve,
|
|
aStartTime, 0.0f, 0.0f, aDuration, aValues.Data(),
|
|
aValues.Length());
|
|
return this;
|
|
}
|
|
|
|
void SetValue(float aValue)
|
|
{
|
|
AudioTimelineEvent event(AudioTimelineEvent::SetValue, 0.0f, aValue);
|
|
|
|
ErrorResult rv;
|
|
if (!ValidateEvent(event, rv)) {
|
|
MOZ_ASSERT(false, "This should not happen, "
|
|
"setting the value should always work");
|
|
return;
|
|
}
|
|
|
|
AudioParamTimeline::SetValue(aValue);
|
|
|
|
SendEventToEngine(event);
|
|
}
|
|
|
|
AudioParam* SetValueAtTime(float aValue, double aStartTime, ErrorResult& aRv)
|
|
{
|
|
if (!WebAudioUtils::IsTimeValid(aStartTime)) {
|
|
aRv.Throw(NS_ERROR_DOM_NOT_SUPPORTED_ERR);
|
|
return this;
|
|
}
|
|
aStartTime = std::max(aStartTime, GetParentObject()->CurrentTime());
|
|
EventInsertionHelper(aRv, AudioTimelineEvent::SetValueAtTime,
|
|
aStartTime, aValue);
|
|
|
|
return this;
|
|
}
|
|
|
|
AudioParam* LinearRampToValueAtTime(float aValue, double aEndTime,
|
|
ErrorResult& aRv)
|
|
{
|
|
if (!WebAudioUtils::IsTimeValid(aEndTime)) {
|
|
aRv.Throw(NS_ERROR_DOM_NOT_SUPPORTED_ERR);
|
|
return this;
|
|
}
|
|
aEndTime = std::max(aEndTime, GetParentObject()->CurrentTime());
|
|
EventInsertionHelper(aRv, AudioTimelineEvent::LinearRamp, aEndTime, aValue);
|
|
return this;
|
|
}
|
|
|
|
AudioParam* ExponentialRampToValueAtTime(float aValue, double aEndTime,
|
|
ErrorResult& aRv)
|
|
{
|
|
if (!WebAudioUtils::IsTimeValid(aEndTime)) {
|
|
aRv.Throw(NS_ERROR_DOM_NOT_SUPPORTED_ERR);
|
|
return this;
|
|
}
|
|
aEndTime = std::max(aEndTime, GetParentObject()->CurrentTime());
|
|
EventInsertionHelper(aRv, AudioTimelineEvent::ExponentialRamp,
|
|
aEndTime, aValue);
|
|
return this;
|
|
}
|
|
|
|
AudioParam* SetTargetAtTime(float aTarget, double aStartTime,
|
|
double aTimeConstant, ErrorResult& aRv)
|
|
{
|
|
if (!WebAudioUtils::IsTimeValid(aStartTime) ||
|
|
!WebAudioUtils::IsTimeValid(aTimeConstant)) {
|
|
aRv.Throw(NS_ERROR_DOM_NOT_SUPPORTED_ERR);
|
|
return this;
|
|
}
|
|
aStartTime = std::max(aStartTime, GetParentObject()->CurrentTime());
|
|
EventInsertionHelper(aRv, AudioTimelineEvent::SetTarget,
|
|
aStartTime, aTarget,
|
|
aTimeConstant);
|
|
|
|
return this;
|
|
}
|
|
|
|
AudioParam* CancelScheduledValues(double aStartTime, ErrorResult& aRv)
|
|
{
|
|
if (!WebAudioUtils::IsTimeValid(aStartTime)) {
|
|
aRv.Throw(NS_ERROR_DOM_NOT_SUPPORTED_ERR);
|
|
return this;
|
|
}
|
|
|
|
aStartTime = std::max(aStartTime, GetParentObject()->CurrentTime());
|
|
|
|
// Remove some events on the main thread copy.
|
|
AudioEventTimeline::CancelScheduledValues(aStartTime);
|
|
|
|
AudioTimelineEvent event(AudioTimelineEvent::Cancel, aStartTime, 0.0f);
|
|
|
|
SendEventToEngine(event);
|
|
|
|
return this;
|
|
}
|
|
|
|
uint32_t ParentNodeId()
|
|
{
|
|
return mNode->Id();
|
|
}
|
|
|
|
void GetName(nsAString& aName)
|
|
{
|
|
aName.AssignASCII(mName);
|
|
}
|
|
|
|
float DefaultValue() const
|
|
{
|
|
return mDefaultValue;
|
|
}
|
|
|
|
const nsTArray<AudioNode::InputNode>& InputNodes() const
|
|
{
|
|
return mInputNodes;
|
|
}
|
|
|
|
void RemoveInputNode(uint32_t aIndex)
|
|
{
|
|
mInputNodes.RemoveElementAt(aIndex);
|
|
}
|
|
|
|
AudioNode::InputNode* AppendInputNode()
|
|
{
|
|
return mInputNodes.AppendElement();
|
|
}
|
|
|
|
// May create the stream if it doesn't exist
|
|
MediaStream* Stream();
|
|
|
|
size_t SizeOfExcludingThis(MallocSizeOf aMallocSizeOf) const override
|
|
{
|
|
size_t amount = AudioParamTimeline::SizeOfExcludingThis(aMallocSizeOf);
|
|
// Not owned:
|
|
// - mNode
|
|
|
|
// Just count the array, actual nodes are counted in mNode.
|
|
amount += mInputNodes.ShallowSizeOfExcludingThis(aMallocSizeOf);
|
|
|
|
if (mNodeStreamPort) {
|
|
amount += mNodeStreamPort->SizeOfIncludingThis(aMallocSizeOf);
|
|
}
|
|
|
|
return amount;
|
|
}
|
|
|
|
size_t SizeOfIncludingThis(MallocSizeOf aMallocSizeOf) const override
|
|
{
|
|
return aMallocSizeOf(this) + SizeOfExcludingThis(aMallocSizeOf);
|
|
}
|
|
|
|
private:
|
|
void EventInsertionHelper(ErrorResult& aRv,
|
|
AudioTimelineEvent::Type aType,
|
|
double aTime, float aValue,
|
|
double aTimeConstant = 0.0,
|
|
float aDuration = 0.0,
|
|
const float* aCurve = nullptr,
|
|
uint32_t aCurveLength = 0)
|
|
{
|
|
AudioTimelineEvent event(aType, aTime, aValue,
|
|
aTimeConstant, aDuration, aCurve, aCurveLength);
|
|
|
|
if (!ValidateEvent(event, aRv)) {
|
|
return;
|
|
}
|
|
|
|
AudioEventTimeline::InsertEvent<double>(event);
|
|
|
|
SendEventToEngine(event);
|
|
|
|
CleanupOldEvents();
|
|
}
|
|
|
|
void CleanupOldEvents();
|
|
|
|
void SendEventToEngine(const AudioTimelineEvent& aEvent);
|
|
|
|
void DisconnectFromGraphAndDestroyStream();
|
|
|
|
nsCycleCollectingAutoRefCnt mRefCnt;
|
|
NS_DECL_OWNINGTHREAD
|
|
RefPtr<AudioNode> mNode;
|
|
// For every InputNode, there is a corresponding entry in mOutputParams of the
|
|
// InputNode's mInputNode.
|
|
nsTArray<AudioNode::InputNode> mInputNodes;
|
|
const char* mName;
|
|
// The input port used to connect the AudioParam's stream to its node's stream
|
|
RefPtr<MediaInputPort> mNodeStreamPort;
|
|
const uint32_t mIndex;
|
|
const float mDefaultValue;
|
|
};
|
|
|
|
} // namespace dom
|
|
} // namespace mozilla
|
|
|
|
#endif
|
|
|