react-native-audio-api 0.10.0-nightly-d3a7f65-20251028 → 0.10.0-nightly-e16d7ff-20251030
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/RNAudioAPI.podspec +7 -6
- package/android/src/main/cpp/audioapi/CMakeLists.txt +6 -0
- package/android/src/main/java/com/swmansion/audioapi/system/MediaNotificationManager.kt +11 -0
- package/common/cpp/audioapi/HostObjects/BaseAudioContextHostObject.cpp +18 -0
- package/common/cpp/audioapi/HostObjects/BaseAudioContextHostObject.h +1 -0
- package/common/cpp/audioapi/HostObjects/effects/ConvolverNodeHostObject.cpp +47 -0
- package/common/cpp/audioapi/HostObjects/effects/ConvolverNodeHostObject.h +20 -0
- package/common/cpp/audioapi/core/AudioNode.h +3 -2
- package/common/cpp/audioapi/core/BaseAudioContext.cpp +10 -0
- package/common/cpp/audioapi/core/BaseAudioContext.h +2 -0
- package/common/cpp/audioapi/core/effects/ConvolverNode.cpp +210 -0
- package/common/cpp/audioapi/core/effects/ConvolverNode.h +55 -0
- package/common/cpp/audioapi/core/sources/StreamerNode.cpp +59 -58
- package/common/cpp/audioapi/core/sources/StreamerNode.h +37 -8
- package/common/cpp/audioapi/core/utils/AudioNodeManager.cpp +5 -0
- package/common/cpp/audioapi/dsp/AudioUtils.cpp +1 -1
- package/common/cpp/audioapi/dsp/Convolver.cpp +213 -0
- package/common/cpp/audioapi/dsp/Convolver.h +45 -0
- package/common/cpp/audioapi/dsp/FFT.cpp +0 -26
- package/common/cpp/audioapi/dsp/FFT.h +26 -2
- package/common/cpp/audioapi/utils/AlignedAllocator.hpp +50 -0
- package/common/cpp/audioapi/utils/AudioBus.cpp +28 -0
- package/common/cpp/audioapi/utils/AudioBus.h +3 -0
- package/common/cpp/test/CMakeLists.txt +16 -14
- package/lib/commonjs/api.js +7 -0
- package/lib/commonjs/api.js.map +1 -1
- package/lib/commonjs/api.web.js +8 -0
- package/lib/commonjs/api.web.js.map +1 -1
- package/lib/commonjs/core/BaseAudioContext.js +12 -0
- package/lib/commonjs/core/BaseAudioContext.js.map +1 -1
- package/lib/commonjs/core/ConvolverNode.js +37 -0
- package/lib/commonjs/core/ConvolverNode.js.map +1 -0
- package/lib/commonjs/types.js +4 -0
- package/lib/commonjs/web-core/AudioContext.js +12 -0
- package/lib/commonjs/web-core/AudioContext.js.map +1 -1
- package/lib/commonjs/web-core/ConvolverNode.js +40 -0
- package/lib/commonjs/web-core/ConvolverNode.js.map +1 -0
- package/lib/commonjs/web-core/ConvolverNodeOptions.js +6 -0
- package/lib/commonjs/web-core/ConvolverNodeOptions.js.map +1 -0
- package/lib/commonjs/web-core/OfflineAudioContext.js +12 -0
- package/lib/commonjs/web-core/OfflineAudioContext.js.map +1 -1
- package/lib/module/api.js +1 -0
- package/lib/module/api.js.map +1 -1
- package/lib/module/api.web.js +1 -0
- package/lib/module/api.web.js.map +1 -1
- package/lib/module/core/BaseAudioContext.js +12 -0
- package/lib/module/core/BaseAudioContext.js.map +1 -1
- package/lib/module/core/ConvolverNode.js +31 -0
- package/lib/module/core/ConvolverNode.js.map +1 -0
- package/lib/module/types.js +2 -0
- package/lib/module/web-core/AudioContext.js +12 -0
- package/lib/module/web-core/AudioContext.js.map +1 -1
- package/lib/module/web-core/ConvolverNode.js +34 -0
- package/lib/module/web-core/ConvolverNode.js.map +1 -0
- package/lib/module/web-core/ConvolverNodeOptions.js +4 -0
- package/lib/module/web-core/ConvolverNodeOptions.js.map +1 -0
- package/lib/module/web-core/OfflineAudioContext.js +12 -0
- package/lib/module/web-core/OfflineAudioContext.js.map +1 -1
- package/lib/typescript/api.d.ts +1 -0
- package/lib/typescript/api.d.ts.map +1 -1
- package/lib/typescript/api.web.d.ts +1 -0
- package/lib/typescript/api.web.d.ts.map +1 -1
- package/lib/typescript/core/BaseAudioContext.d.ts +3 -1
- package/lib/typescript/core/BaseAudioContext.d.ts.map +1 -1
- package/lib/typescript/core/ConvolverNode.d.ts +12 -0
- package/lib/typescript/core/ConvolverNode.d.ts.map +1 -0
- package/lib/typescript/interfaces.d.ts +5 -0
- package/lib/typescript/interfaces.d.ts.map +1 -1
- package/lib/typescript/types.d.ts +5 -0
- package/lib/typescript/types.d.ts.map +1 -1
- package/lib/typescript/web-core/AudioContext.d.ts +3 -0
- package/lib/typescript/web-core/AudioContext.d.ts.map +1 -1
- package/lib/typescript/web-core/BaseAudioContext.d.ts +2 -0
- package/lib/typescript/web-core/BaseAudioContext.d.ts.map +1 -1
- package/lib/typescript/web-core/ConvolverNode.d.ts +11 -0
- package/lib/typescript/web-core/ConvolverNode.d.ts.map +1 -0
- package/lib/typescript/web-core/ConvolverNodeOptions.d.ts +6 -0
- package/lib/typescript/web-core/ConvolverNodeOptions.d.ts.map +1 -0
- package/lib/typescript/web-core/OfflineAudioContext.d.ts +3 -0
- package/lib/typescript/web-core/OfflineAudioContext.d.ts.map +1 -1
- package/package.json +1 -1
- package/src/api.ts +1 -0
- package/src/api.web.ts +1 -0
- package/src/core/BaseAudioContext.ts +23 -0
- package/src/core/ConvolverNode.ts +35 -0
- package/src/interfaces.ts +11 -0
- package/src/types.ts +7 -0
- package/src/web-core/AudioContext.tsx +25 -0
- package/src/web-core/BaseAudioContext.tsx +2 -0
- package/src/web-core/ConvolverNode.tsx +43 -0
- package/src/web-core/ConvolverNodeOptions.tsx +6 -0
- package/src/web-core/OfflineAudioContext.tsx +25 -0
package/RNAudioAPI.podspec
CHANGED
|
@@ -32,6 +32,13 @@ Pod::Spec.new do |s|
|
|
|
32
32
|
sss.header_dir = "audioapi"
|
|
33
33
|
sss.header_mappings_dir = "ios/audioapi"
|
|
34
34
|
end
|
|
35
|
+
|
|
36
|
+
ss.subspec "audioapi_dsp" do |sss|
|
|
37
|
+
sss.source_files = "common/cpp/audioapi/dsp/**/*.{cpp}"
|
|
38
|
+
sss.header_dir = "audioapi/dsp"
|
|
39
|
+
sss.header_mappings_dir = "common/cpp/audioapi/dsp"
|
|
40
|
+
sss.compiler_flags = "-O3"
|
|
41
|
+
end
|
|
35
42
|
end
|
|
36
43
|
|
|
37
44
|
s.ios.frameworks = 'CoreFoundation', 'CoreAudio', 'AudioToolbox', 'Accelerate', 'MediaPlayer', 'AVFoundation'
|
|
@@ -50,8 +57,6 @@ Pod::Spec.new do |s|
|
|
|
50
57
|
external_dir_relative = "common/cpp/audioapi/external"
|
|
51
58
|
lib_dir = "$(PROJECT_DIR)/#{rn_audio_dir_relative}/#{external_dir_relative}/$(PLATFORM_NAME)"
|
|
52
59
|
|
|
53
|
-
external_dir = File.join(__dir__, "common/cpp/audioapi/external")
|
|
54
|
-
|
|
55
60
|
s.ios.vendored_frameworks = [
|
|
56
61
|
'common/cpp/audioapi/external/libavcodec.xcframework',
|
|
57
62
|
'common/cpp/audioapi/external/libavformat.xcframework',
|
|
@@ -92,10 +97,6 @@ s.user_target_xcconfig = {
|
|
|
92
97
|
$(inherited)
|
|
93
98
|
$(PODS_ROOT)/Headers/Public/RNAudioAPI
|
|
94
99
|
$(PODS_TARGET_SRCROOT)/common/cpp
|
|
95
|
-
#{external_dir}/include
|
|
96
|
-
#{external_dir}/include/opus
|
|
97
|
-
#{external_dir}/include/vorbis
|
|
98
|
-
#{external_dir}/ffmpeg_include
|
|
99
100
|
].join(' ')
|
|
100
101
|
}
|
|
101
102
|
# Use install_modules_dependencies helper to install the dependencies if React Native version >=0.71.0.
|
|
@@ -3,6 +3,12 @@ cmake_minimum_required(VERSION 3.12.0)
|
|
|
3
3
|
file(GLOB_RECURSE ANDROID_CPP_SOURCES CONFIGURE_DEPENDS "${ANDROID_CPP_DIR}/audioapi/*.cpp")
|
|
4
4
|
file(GLOB_RECURSE COMMON_CPP_SOURCES CONFIGURE_DEPENDS "${COMMON_CPP_DIR}/audioapi/*.cpp" "${COMMON_CPP_DIR}/audioapi/*.c")
|
|
5
5
|
|
|
6
|
+
set_source_files_properties(
|
|
7
|
+
${COMMON_CPP_SOURCES}/dsp/*.cpp
|
|
8
|
+
PROPERTIES
|
|
9
|
+
COMPILE_FLAGS "-O3"
|
|
10
|
+
)
|
|
11
|
+
|
|
6
12
|
set(INCLUDE_DIR ${COMMON_CPP_DIR}/audioapi/external/include)
|
|
7
13
|
set(FFMPEG_INCLUDE_DIR ${COMMON_CPP_DIR}/audioapi/external/ffmpeg_include)
|
|
8
14
|
set(EXTERNAL_DIR ${COMMON_CPP_DIR}/audioapi/external)
|
|
@@ -258,5 +258,16 @@ class MediaNotificationManager(
|
|
|
258
258
|
}
|
|
259
259
|
super.onDestroy()
|
|
260
260
|
}
|
|
261
|
+
|
|
262
|
+
override fun onTimeout(startId: Int) {
|
|
263
|
+
stopForegroundService()
|
|
264
|
+
}
|
|
265
|
+
|
|
266
|
+
override fun onTimeout(
|
|
267
|
+
startId: Int,
|
|
268
|
+
fgsType: Int,
|
|
269
|
+
) {
|
|
270
|
+
stopForegroundService()
|
|
271
|
+
}
|
|
261
272
|
}
|
|
262
273
|
}
|
|
@@ -5,6 +5,7 @@
|
|
|
5
5
|
#include <audioapi/HostObjects/analysis/AnalyserNodeHostObject.h>
|
|
6
6
|
#include <audioapi/HostObjects/destinations/AudioDestinationNodeHostObject.h>
|
|
7
7
|
#include <audioapi/HostObjects/effects/BiquadFilterNodeHostObject.h>
|
|
8
|
+
#include <audioapi/HostObjects/effects/ConvolverNodeHostObject.h>
|
|
8
9
|
#include <audioapi/HostObjects/effects/GainNodeHostObject.h>
|
|
9
10
|
#include <audioapi/HostObjects/effects/PeriodicWaveHostObject.h>
|
|
10
11
|
#include <audioapi/HostObjects/effects/StereoPannerNodeHostObject.h>
|
|
@@ -49,6 +50,7 @@ BaseAudioContextHostObject::BaseAudioContextHostObject(
|
|
|
49
50
|
JSI_EXPORT_FUNCTION(BaseAudioContextHostObject, createBufferQueueSource),
|
|
50
51
|
JSI_EXPORT_FUNCTION(BaseAudioContextHostObject, createBuffer),
|
|
51
52
|
JSI_EXPORT_FUNCTION(BaseAudioContextHostObject, createPeriodicWave),
|
|
53
|
+
JSI_EXPORT_FUNCTION(BaseAudioContextHostObject, createConvolver),
|
|
52
54
|
JSI_EXPORT_FUNCTION(BaseAudioContextHostObject, createAnalyser));
|
|
53
55
|
}
|
|
54
56
|
|
|
@@ -269,4 +271,20 @@ JSI_HOST_FUNCTION_IMPL(BaseAudioContextHostObject, createAnalyser) {
|
|
|
269
271
|
auto analyserHostObject = std::make_shared<AnalyserNodeHostObject>(analyser);
|
|
270
272
|
return jsi::Object::createFromHostObject(runtime, analyserHostObject);
|
|
271
273
|
}
|
|
274
|
+
|
|
275
|
+
JSI_HOST_FUNCTION_IMPL(BaseAudioContextHostObject, createConvolver) {
|
|
276
|
+
auto disableNormalization = args[1].getBool();
|
|
277
|
+
std::shared_ptr<ConvolverNode> convolver;
|
|
278
|
+
if (args[0].isUndefined()) {
|
|
279
|
+
convolver = context_->createConvolver(nullptr, disableNormalization);
|
|
280
|
+
} else {
|
|
281
|
+
auto bufferHostObject =
|
|
282
|
+
args[0].getObject(runtime).asHostObject<AudioBufferHostObject>(runtime);
|
|
283
|
+
convolver = context_->createConvolver(
|
|
284
|
+
bufferHostObject->audioBuffer_, disableNormalization);
|
|
285
|
+
}
|
|
286
|
+
auto convolverHostObject =
|
|
287
|
+
std::make_shared<ConvolverNodeHostObject>(convolver);
|
|
288
|
+
return jsi::Object::createFromHostObject(runtime, convolverHostObject);
|
|
289
|
+
}
|
|
272
290
|
} // namespace audioapi
|
|
@@ -41,6 +41,7 @@ class BaseAudioContextHostObject : public JsiHostObject {
|
|
|
41
41
|
JSI_HOST_FUNCTION_DECL(createBuffer);
|
|
42
42
|
JSI_HOST_FUNCTION_DECL(createPeriodicWave);
|
|
43
43
|
JSI_HOST_FUNCTION_DECL(createAnalyser);
|
|
44
|
+
JSI_HOST_FUNCTION_DECL(createConvolver);
|
|
44
45
|
|
|
45
46
|
std::shared_ptr<BaseAudioContext> context_;
|
|
46
47
|
|
|
@@ -0,0 +1,47 @@
|
|
|
1
|
+
#include <audioapi/HostObjects/effects/ConvolverNodeHostObject.h>
|
|
2
|
+
|
|
3
|
+
#include <audioapi/HostObjects/sources/AudioBufferHostObject.h>
|
|
4
|
+
#include <audioapi/core/effects/ConvolverNode.h>
|
|
5
|
+
|
|
6
|
+
namespace audioapi {
|
|
7
|
+
|
|
8
|
+
ConvolverNodeHostObject::ConvolverNodeHostObject(
|
|
9
|
+
const std::shared_ptr<ConvolverNode> &node)
|
|
10
|
+
: AudioNodeHostObject(node) {
|
|
11
|
+
addGetters(
|
|
12
|
+
JSI_EXPORT_PROPERTY_GETTER(ConvolverNodeHostObject, normalize),
|
|
13
|
+
JSI_EXPORT_PROPERTY_GETTER(ConvolverNodeHostObject, buffer));
|
|
14
|
+
addSetters(
|
|
15
|
+
JSI_EXPORT_PROPERTY_SETTER(ConvolverNodeHostObject, normalize),
|
|
16
|
+
JSI_EXPORT_PROPERTY_SETTER(ConvolverNodeHostObject, buffer));
|
|
17
|
+
}
|
|
18
|
+
|
|
19
|
+
JSI_PROPERTY_GETTER_IMPL(ConvolverNodeHostObject, normalize) {
|
|
20
|
+
auto convolverNode = std::static_pointer_cast<ConvolverNode>(node_);
|
|
21
|
+
return {convolverNode->getNormalize_()};
|
|
22
|
+
}
|
|
23
|
+
|
|
24
|
+
JSI_PROPERTY_GETTER_IMPL(ConvolverNodeHostObject, buffer) {
|
|
25
|
+
auto convolverNode = std::static_pointer_cast<ConvolverNode>(node_);
|
|
26
|
+
auto buffer = convolverNode->getBuffer();
|
|
27
|
+
auto bufferHostObject = std::make_shared<AudioBufferHostObject>(buffer);
|
|
28
|
+
return jsi::Object::createFromHostObject(runtime, bufferHostObject);
|
|
29
|
+
}
|
|
30
|
+
|
|
31
|
+
JSI_PROPERTY_SETTER_IMPL(ConvolverNodeHostObject, normalize) {
|
|
32
|
+
auto convolverNode = std::static_pointer_cast<ConvolverNode>(node_);
|
|
33
|
+
convolverNode->setNormalize(value.getBool());
|
|
34
|
+
}
|
|
35
|
+
|
|
36
|
+
JSI_PROPERTY_SETTER_IMPL(ConvolverNodeHostObject, buffer) {
|
|
37
|
+
auto convolverNode = std::static_pointer_cast<ConvolverNode>(node_);
|
|
38
|
+
if (value.isNull()) {
|
|
39
|
+
convolverNode->setBuffer(nullptr);
|
|
40
|
+
return;
|
|
41
|
+
}
|
|
42
|
+
|
|
43
|
+
auto bufferHostObject =
|
|
44
|
+
value.getObject(runtime).asHostObject<AudioBufferHostObject>(runtime);
|
|
45
|
+
convolverNode->setBuffer(bufferHostObject->audioBuffer_);
|
|
46
|
+
}
|
|
47
|
+
} // namespace audioapi
|
|
@@ -0,0 +1,20 @@
|
|
|
1
|
+
#pragma once
|
|
2
|
+
|
|
3
|
+
#include <audioapi/HostObjects/AudioNodeHostObject.h>
|
|
4
|
+
|
|
5
|
+
#include <memory>
|
|
6
|
+
|
|
7
|
+
namespace audioapi {
|
|
8
|
+
using namespace facebook;
|
|
9
|
+
|
|
10
|
+
class ConvolverNode;
|
|
11
|
+
|
|
12
|
+
class ConvolverNodeHostObject : public AudioNodeHostObject {
|
|
13
|
+
public:
|
|
14
|
+
explicit ConvolverNodeHostObject(const std::shared_ptr<ConvolverNode> &node);
|
|
15
|
+
JSI_PROPERTY_GETTER_DECL(normalize);
|
|
16
|
+
JSI_PROPERTY_GETTER_DECL(buffer);
|
|
17
|
+
JSI_PROPERTY_SETTER_DECL(normalize);
|
|
18
|
+
JSI_PROPERTY_SETTER_DECL(buffer);
|
|
19
|
+
};
|
|
20
|
+
} // namespace audioapi
|
|
@@ -41,6 +41,7 @@ class AudioNode : public std::enable_shared_from_this<AudioNode> {
|
|
|
41
41
|
protected:
|
|
42
42
|
friend class AudioNodeManager;
|
|
43
43
|
friend class AudioDestinationNode;
|
|
44
|
+
friend class ConvolverNode;
|
|
44
45
|
|
|
45
46
|
BaseAudioContext *context_;
|
|
46
47
|
std::shared_ptr<AudioBus> audioBus_;
|
|
@@ -68,10 +69,10 @@ class AudioNode : public std::enable_shared_from_this<AudioNode> {
|
|
|
68
69
|
static std::string toString(ChannelCountMode mode);
|
|
69
70
|
static std::string toString(ChannelInterpretation interpretation);
|
|
70
71
|
|
|
72
|
+
virtual std::shared_ptr<AudioBus> processInputs(const std::shared_ptr<AudioBus>& outputBus, int framesToProcess, bool checkIsAlreadyProcessed);
|
|
71
73
|
virtual std::shared_ptr<AudioBus> processNode(const std::shared_ptr<AudioBus>&, int) = 0;
|
|
72
74
|
|
|
73
75
|
bool isAlreadyProcessed();
|
|
74
|
-
std::shared_ptr<AudioBus> processInputs(const std::shared_ptr<AudioBus>& outputBus, int framesToProcess, bool checkIsAlreadyProcessed);
|
|
75
76
|
std::shared_ptr<AudioBus> applyChannelCountMode(const std::shared_ptr<AudioBus> &processingBus);
|
|
76
77
|
void mixInputsBuses(const std::shared_ptr<AudioBus>& processingBus);
|
|
77
78
|
|
|
@@ -81,7 +82,7 @@ class AudioNode : public std::enable_shared_from_this<AudioNode> {
|
|
|
81
82
|
void disconnectParam(const std::shared_ptr<AudioParam> ¶m);
|
|
82
83
|
|
|
83
84
|
void onInputEnabled();
|
|
84
|
-
void onInputDisabled();
|
|
85
|
+
virtual void onInputDisabled();
|
|
85
86
|
void onInputConnected(AudioNode *node);
|
|
86
87
|
void onInputDisconnected(AudioNode *node);
|
|
87
88
|
|
|
@@ -2,6 +2,7 @@
|
|
|
2
2
|
#include <audioapi/core/analysis/AnalyserNode.h>
|
|
3
3
|
#include <audioapi/core/destinations/AudioDestinationNode.h>
|
|
4
4
|
#include <audioapi/core/effects/BiquadFilterNode.h>
|
|
5
|
+
#include <audioapi/core/effects/ConvolverNode.h>
|
|
5
6
|
#include <audioapi/core/effects/GainNode.h>
|
|
6
7
|
#include <audioapi/core/effects/StereoPannerNode.h>
|
|
7
8
|
#include <audioapi/core/effects/WorkletNode.h>
|
|
@@ -182,6 +183,15 @@ std::shared_ptr<AnalyserNode> BaseAudioContext::createAnalyser() {
|
|
|
182
183
|
return analyser;
|
|
183
184
|
}
|
|
184
185
|
|
|
186
|
+
std::shared_ptr<ConvolverNode> BaseAudioContext::createConvolver(
|
|
187
|
+
std::shared_ptr<AudioBuffer> buffer,
|
|
188
|
+
bool disableNormalization) {
|
|
189
|
+
auto convolver =
|
|
190
|
+
std::make_shared<ConvolverNode>(this, buffer, disableNormalization);
|
|
191
|
+
nodeManager_->addProcessingNode(convolver);
|
|
192
|
+
return convolver;
|
|
193
|
+
}
|
|
194
|
+
|
|
185
195
|
AudioNodeManager *BaseAudioContext::getNodeManager() {
|
|
186
196
|
return nodeManager_.get();
|
|
187
197
|
}
|
|
@@ -28,6 +28,7 @@ class AudioBufferSourceNode;
|
|
|
28
28
|
class AudioBufferQueueSourceNode;
|
|
29
29
|
class AnalyserNode;
|
|
30
30
|
class AudioEventHandlerRegistry;
|
|
31
|
+
class ConvolverNode;
|
|
31
32
|
class IAudioEventHandlerRegistry;
|
|
32
33
|
class RecorderAdapterNode;
|
|
33
34
|
class WorkletSourceNode;
|
|
@@ -76,6 +77,7 @@ class BaseAudioContext {
|
|
|
76
77
|
bool disableNormalization,
|
|
77
78
|
int length);
|
|
78
79
|
std::shared_ptr<AnalyserNode> createAnalyser();
|
|
80
|
+
std::shared_ptr<ConvolverNode> createConvolver(std::shared_ptr<AudioBuffer> buffer, bool disableNormalization);
|
|
79
81
|
|
|
80
82
|
std::shared_ptr<PeriodicWave> getBasicWaveForm(OscillatorType type);
|
|
81
83
|
[[nodiscard]] float getNyquistFrequency() const;
|
|
@@ -0,0 +1,210 @@
|
|
|
1
|
+
#include <audioapi/core/BaseAudioContext.h>
|
|
2
|
+
#include <audioapi/core/effects/ConvolverNode.h>
|
|
3
|
+
#include <audioapi/core/sources/AudioBuffer.h>
|
|
4
|
+
#include <audioapi/core/utils/Constants.h>
|
|
5
|
+
#include <audioapi/dsp/AudioUtils.h>
|
|
6
|
+
#include <audioapi/dsp/FFT.h>
|
|
7
|
+
#include <audioapi/utils/AudioArray.h>
|
|
8
|
+
#include <iostream>
|
|
9
|
+
#include <thread>
|
|
10
|
+
|
|
11
|
+
namespace audioapi {
|
|
12
|
+
ConvolverNode::ConvolverNode(
|
|
13
|
+
BaseAudioContext *context,
|
|
14
|
+
std::shared_ptr<AudioBuffer> buffer,
|
|
15
|
+
bool disableNormalization)
|
|
16
|
+
: AudioNode(context),
|
|
17
|
+
buffer_(nullptr),
|
|
18
|
+
internalBuffer_(nullptr),
|
|
19
|
+
signalledToStop_(false),
|
|
20
|
+
remainingSegments_(0),
|
|
21
|
+
internalBufferIndex_(0),
|
|
22
|
+
scaleFactor_(1.0f),
|
|
23
|
+
intermediateBus_(nullptr) {
|
|
24
|
+
channelCount_ = 2;
|
|
25
|
+
channelCountMode_ = ChannelCountMode::CLAMPED_MAX;
|
|
26
|
+
normalize_ = !disableNormalization;
|
|
27
|
+
gainCalibrationSampleRate_ = context->getSampleRate();
|
|
28
|
+
setBuffer(buffer);
|
|
29
|
+
audioBus_ = std::make_shared<AudioBus>(
|
|
30
|
+
RENDER_QUANTUM_SIZE, channelCount_, context->getSampleRate());
|
|
31
|
+
isInitialized_ = true;
|
|
32
|
+
}
|
|
33
|
+
|
|
34
|
+
bool ConvolverNode::getNormalize_() const {
|
|
35
|
+
return normalize_;
|
|
36
|
+
}
|
|
37
|
+
|
|
38
|
+
const std::shared_ptr<AudioBuffer> &ConvolverNode::getBuffer() const {
|
|
39
|
+
return buffer_;
|
|
40
|
+
}
|
|
41
|
+
|
|
42
|
+
void ConvolverNode::setNormalize(bool normalize) {
|
|
43
|
+
if (normalize_ != normalize) {
|
|
44
|
+
normalize_ = normalize;
|
|
45
|
+
if (normalize_ && buffer_)
|
|
46
|
+
calculateNormalizationScale();
|
|
47
|
+
}
|
|
48
|
+
if (!normalize_) {
|
|
49
|
+
scaleFactor_ = 1.0f;
|
|
50
|
+
}
|
|
51
|
+
}
|
|
52
|
+
|
|
53
|
+
void ConvolverNode::setBuffer(const std::shared_ptr<AudioBuffer> &buffer) {
|
|
54
|
+
if (buffer_ != buffer && buffer != nullptr) {
|
|
55
|
+
buffer_ = buffer;
|
|
56
|
+
if (normalize_)
|
|
57
|
+
calculateNormalizationScale();
|
|
58
|
+
threadPool_ = std::make_shared<ThreadPool>(4);
|
|
59
|
+
convolvers_.clear();
|
|
60
|
+
for (int i = 0; i < buffer->getNumberOfChannels(); ++i) {
|
|
61
|
+
convolvers_.emplace_back();
|
|
62
|
+
AudioArray channelData(buffer->getLength());
|
|
63
|
+
memcpy(
|
|
64
|
+
channelData.getData(),
|
|
65
|
+
buffer->getChannelData(i),
|
|
66
|
+
buffer->getLength() * sizeof(float));
|
|
67
|
+
convolvers_.back().init(
|
|
68
|
+
RENDER_QUANTUM_SIZE, channelData, buffer->getLength());
|
|
69
|
+
}
|
|
70
|
+
if (buffer->getNumberOfChannels() == 1) {
|
|
71
|
+
// add one more convolver, because right now input is always stereo
|
|
72
|
+
convolvers_.emplace_back();
|
|
73
|
+
AudioArray channelData(buffer->getLength());
|
|
74
|
+
memcpy(
|
|
75
|
+
channelData.getData(),
|
|
76
|
+
buffer->getChannelData(0),
|
|
77
|
+
buffer->getLength() * sizeof(float));
|
|
78
|
+
convolvers_.back().init(
|
|
79
|
+
RENDER_QUANTUM_SIZE, channelData, buffer->getLength());
|
|
80
|
+
}
|
|
81
|
+
internalBuffer_ = std::make_shared<AudioBus>(
|
|
82
|
+
RENDER_QUANTUM_SIZE * 2, channelCount_, buffer->getSampleRate());
|
|
83
|
+
intermediateBus_ = std::make_shared<AudioBus>(
|
|
84
|
+
RENDER_QUANTUM_SIZE, convolvers_.size(), buffer->getSampleRate());
|
|
85
|
+
internalBufferIndex_ = 0;
|
|
86
|
+
}
|
|
87
|
+
}
|
|
88
|
+
|
|
89
|
+
void ConvolverNode::onInputDisabled() {
|
|
90
|
+
numberOfEnabledInputNodes_ -= 1;
|
|
91
|
+
if (isEnabled() && numberOfEnabledInputNodes_ == 0) {
|
|
92
|
+
signalledToStop_ = true;
|
|
93
|
+
remainingSegments_ = convolvers_.at(0).getSegCount();
|
|
94
|
+
}
|
|
95
|
+
}
|
|
96
|
+
|
|
97
|
+
std::shared_ptr<AudioBus> ConvolverNode::processInputs(
|
|
98
|
+
const std::shared_ptr<AudioBus> &outputBus,
|
|
99
|
+
int framesToProcess,
|
|
100
|
+
bool checkIsAlreadyProcessed) {
|
|
101
|
+
if (internalBufferIndex_ < framesToProcess) {
|
|
102
|
+
return AudioNode::processInputs(outputBus, RENDER_QUANTUM_SIZE, false);
|
|
103
|
+
}
|
|
104
|
+
return AudioNode::processInputs(outputBus, 0, false);
|
|
105
|
+
}
|
|
106
|
+
|
|
107
|
+
// processing pipeline: processingBus -> intermediateBus_ -> audioBus_ (mixing
|
|
108
|
+
// with intermediateBus_)
|
|
109
|
+
std::shared_ptr<AudioBus> ConvolverNode::processNode(
|
|
110
|
+
const std::shared_ptr<AudioBus> &processingBus,
|
|
111
|
+
int framesToProcess) {
|
|
112
|
+
if (signalledToStop_) {
|
|
113
|
+
if (remainingSegments_ > 0) {
|
|
114
|
+
remainingSegments_--;
|
|
115
|
+
} else {
|
|
116
|
+
disable();
|
|
117
|
+
signalledToStop_ = false;
|
|
118
|
+
internalBufferIndex_ = 0;
|
|
119
|
+
return processingBus;
|
|
120
|
+
}
|
|
121
|
+
}
|
|
122
|
+
if (internalBufferIndex_ < framesToProcess) {
|
|
123
|
+
performConvolution(processingBus); // result returned to intermediateBus_
|
|
124
|
+
audioBus_->sum(intermediateBus_.get());
|
|
125
|
+
|
|
126
|
+
internalBuffer_->copy(
|
|
127
|
+
audioBus_.get(), 0, internalBufferIndex_, RENDER_QUANTUM_SIZE);
|
|
128
|
+
internalBufferIndex_ += RENDER_QUANTUM_SIZE;
|
|
129
|
+
}
|
|
130
|
+
audioBus_->zero();
|
|
131
|
+
audioBus_->copy(internalBuffer_.get(), 0, 0, framesToProcess);
|
|
132
|
+
int remainingFrames = internalBufferIndex_ - framesToProcess;
|
|
133
|
+
if (remainingFrames > 0) {
|
|
134
|
+
for (int i = 0; i < internalBuffer_->getNumberOfChannels(); ++i) {
|
|
135
|
+
memmove(
|
|
136
|
+
internalBuffer_->getChannel(i)->getData(),
|
|
137
|
+
internalBuffer_->getChannel(i)->getData() + framesToProcess,
|
|
138
|
+
remainingFrames * sizeof(float));
|
|
139
|
+
}
|
|
140
|
+
}
|
|
141
|
+
internalBufferIndex_ -= framesToProcess;
|
|
142
|
+
|
|
143
|
+
for (int i = 0; i < audioBus_->getNumberOfChannels(); ++i) {
|
|
144
|
+
dsp::multiplyByScalar(
|
|
145
|
+
audioBus_->getChannel(i)->getData(),
|
|
146
|
+
scaleFactor_,
|
|
147
|
+
audioBus_->getChannel(i)->getData(),
|
|
148
|
+
framesToProcess);
|
|
149
|
+
}
|
|
150
|
+
|
|
151
|
+
return audioBus_;
|
|
152
|
+
}
|
|
153
|
+
|
|
154
|
+
void ConvolverNode::calculateNormalizationScale() {
|
|
155
|
+
int numberOfChannels = buffer_->getNumberOfChannels();
|
|
156
|
+
int length = buffer_->getLength();
|
|
157
|
+
|
|
158
|
+
float power = 0;
|
|
159
|
+
|
|
160
|
+
for (int channel = 0; channel < numberOfChannels; ++channel) {
|
|
161
|
+
float channelPower = 0;
|
|
162
|
+
auto channelData = buffer_->getChannelData(channel);
|
|
163
|
+
for (int i = 0; i < length; ++i) {
|
|
164
|
+
float sample = channelData[i];
|
|
165
|
+
channelPower += sample * sample;
|
|
166
|
+
}
|
|
167
|
+
power += channelPower;
|
|
168
|
+
}
|
|
169
|
+
|
|
170
|
+
power = std::sqrt(power / (numberOfChannels * length));
|
|
171
|
+
if (power < MIN_IR_POWER) {
|
|
172
|
+
power = MIN_IR_POWER;
|
|
173
|
+
}
|
|
174
|
+
scaleFactor_ = 1 / power;
|
|
175
|
+
scaleFactor_ *= std::pow(10, GAIN_CALIBRATION * 0.05f);
|
|
176
|
+
scaleFactor_ *= gainCalibrationSampleRate_ / buffer_->getSampleRate();
|
|
177
|
+
}
|
|
178
|
+
|
|
179
|
+
void ConvolverNode::performConvolution(
|
|
180
|
+
const std::shared_ptr<AudioBus> &processingBus) {
|
|
181
|
+
if (processingBus->getNumberOfChannels() == 1) {
|
|
182
|
+
for (int i = 0; i < convolvers_.size(); ++i) {
|
|
183
|
+
threadPool_->schedule([&, i] {
|
|
184
|
+
convolvers_[i].process(
|
|
185
|
+
processingBus->getChannel(0)->getData(),
|
|
186
|
+
intermediateBus_->getChannel(i)->getData());
|
|
187
|
+
});
|
|
188
|
+
}
|
|
189
|
+
} else if (processingBus->getNumberOfChannels() == 2) {
|
|
190
|
+
std::vector<int> inputChannelMap;
|
|
191
|
+
std::vector<int> outputChannelMap;
|
|
192
|
+
if (convolvers_.size() == 2) {
|
|
193
|
+
inputChannelMap = {0, 1};
|
|
194
|
+
outputChannelMap = {0, 1};
|
|
195
|
+
} else { // 4 channel IR
|
|
196
|
+
inputChannelMap = {0, 0, 1, 1};
|
|
197
|
+
outputChannelMap = {0, 3, 2, 1};
|
|
198
|
+
}
|
|
199
|
+
for (int i = 0; i < convolvers_.size(); ++i) {
|
|
200
|
+
threadPool_->schedule(
|
|
201
|
+
[this, i, inputChannelMap, outputChannelMap, &processingBus] {
|
|
202
|
+
convolvers_[i].process(
|
|
203
|
+
processingBus->getChannel(inputChannelMap[i])->getData(),
|
|
204
|
+
intermediateBus_->getChannel(outputChannelMap[i])->getData());
|
|
205
|
+
});
|
|
206
|
+
}
|
|
207
|
+
}
|
|
208
|
+
threadPool_->wait();
|
|
209
|
+
}
|
|
210
|
+
} // namespace audioapi
|
|
@@ -0,0 +1,55 @@
|
|
|
1
|
+
#pragma once
|
|
2
|
+
|
|
3
|
+
#include <audioapi/core/AudioNode.h>
|
|
4
|
+
#include <audioapi/core/AudioParam.h>
|
|
5
|
+
#include <audioapi/dsp/Convolver.h>
|
|
6
|
+
|
|
7
|
+
#include <memory>
|
|
8
|
+
#include <vector>
|
|
9
|
+
|
|
10
|
+
#include <audioapi/utils/ThreadPool.hpp>
|
|
11
|
+
|
|
12
|
+
static constexpr int GAIN_CALIBRATION = -58; // magic number so that processed signal and dry signal have roughly the same volume
|
|
13
|
+
static constexpr double MIN_IR_POWER = 0.000125;
|
|
14
|
+
|
|
15
|
+
namespace audioapi {
|
|
16
|
+
|
|
17
|
+
class AudioBus;
|
|
18
|
+
class AudioBuffer;
|
|
19
|
+
|
|
20
|
+
class ConvolverNode : public AudioNode {
|
|
21
|
+
public:
|
|
22
|
+
explicit ConvolverNode(BaseAudioContext *context, std::shared_ptr<AudioBuffer> buffer, bool disableNormalization);
|
|
23
|
+
|
|
24
|
+
[[nodiscard]] bool getNormalize_() const;
|
|
25
|
+
[[nodiscard]] const std::shared_ptr<AudioBuffer> &getBuffer() const;
|
|
26
|
+
void setNormalize(bool normalize);
|
|
27
|
+
void setBuffer(const std::shared_ptr<AudioBuffer> &buffer);
|
|
28
|
+
|
|
29
|
+
protected:
|
|
30
|
+
std::shared_ptr<AudioBus> processNode(const std::shared_ptr<AudioBus>& processingBus, int framesToProcess) override;
|
|
31
|
+
|
|
32
|
+
private:
|
|
33
|
+
std::shared_ptr<AudioBus> processInputs(const std::shared_ptr<AudioBus>& outputBus, int framesToProcess, bool checkIsAlreadyProcessed) override;
|
|
34
|
+
void onInputDisabled() override;
|
|
35
|
+
float gainCalibrationSampleRate_;
|
|
36
|
+
size_t remainingSegments_;
|
|
37
|
+
size_t internalBufferIndex_;
|
|
38
|
+
bool normalize_;
|
|
39
|
+
bool signalledToStop_;
|
|
40
|
+
float scaleFactor_;
|
|
41
|
+
std::shared_ptr<AudioBus>intermediateBus_;
|
|
42
|
+
|
|
43
|
+
// impulse response buffer
|
|
44
|
+
std::shared_ptr<AudioBuffer> buffer_;
|
|
45
|
+
// buffer to hold internal processed data
|
|
46
|
+
std::shared_ptr<AudioBus> internalBuffer_;
|
|
47
|
+
// vectors of convolvers, one per channel
|
|
48
|
+
std::vector<Convolver> convolvers_;
|
|
49
|
+
std::shared_ptr<ThreadPool> threadPool_;
|
|
50
|
+
|
|
51
|
+
void calculateNormalizationScale();
|
|
52
|
+
void performConvolution(const std::shared_ptr<AudioBus>& processingBus);
|
|
53
|
+
};
|
|
54
|
+
|
|
55
|
+
} // namespace audioapi
|