react-native-davoice 1.0.10 → 1.0.11
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/TTSRNBridge.podspec +1 -1
- package/ios/SpeechBridge/SpeechBridge.m +134 -28
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64/DavoiceTTS.framework/DavoiceTTS +0 -0
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64/DavoiceTTS.framework/Headers/DavoiceTTS-Swift.h +2 -0
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64/DavoiceTTS.framework/Modules/DavoiceTTS.swiftmodule/arm64-apple-ios.abi.json +5201 -5055
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64/DavoiceTTS.framework/Modules/DavoiceTTS.swiftmodule/arm64-apple-ios.private.swiftinterface +14 -12
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64/DavoiceTTS.framework/Modules/DavoiceTTS.swiftmodule/arm64-apple-ios.swiftinterface +14 -12
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64_x86_64-simulator/DavoiceTTS.framework/DavoiceTTS +0 -0
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64_x86_64-simulator/DavoiceTTS.framework/Headers/DavoiceTTS-Swift.h +4 -0
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64_x86_64-simulator/DavoiceTTS.framework/Modules/DavoiceTTS.swiftmodule/arm64-apple-ios-simulator.abi.json +8439 -8293
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64_x86_64-simulator/DavoiceTTS.framework/Modules/DavoiceTTS.swiftmodule/arm64-apple-ios-simulator.private.swiftinterface +82 -80
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64_x86_64-simulator/DavoiceTTS.framework/Modules/DavoiceTTS.swiftmodule/arm64-apple-ios-simulator.swiftinterface +82 -80
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64_x86_64-simulator/DavoiceTTS.framework/Modules/DavoiceTTS.swiftmodule/x86_64-apple-ios-simulator.abi.json +8439 -8293
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64_x86_64-simulator/DavoiceTTS.framework/Modules/DavoiceTTS.swiftmodule/x86_64-apple-ios-simulator.private.swiftinterface +82 -80
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64_x86_64-simulator/DavoiceTTS.framework/Modules/DavoiceTTS.swiftmodule/x86_64-apple-ios-simulator.swiftinterface +82 -80
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64_x86_64-simulator/DavoiceTTS.framework/_CodeSignature/CodeDirectory +0 -0
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64_x86_64-simulator/DavoiceTTS.framework/_CodeSignature/CodeRequirements-1 +0 -0
- package/ios/TTSRNBridge/DavoiceTTS.xcframework/ios-arm64_x86_64-simulator/DavoiceTTS.framework/_CodeSignature/CodeResources +27 -27
- package/package.json +1 -1
- package/speech/index.ts +31 -8
|
@@ -15,81 +15,6 @@ import _StringProcessing
|
|
|
15
15
|
import _SwiftConcurrencyShims
|
|
16
16
|
import onnxruntime_objc
|
|
17
17
|
import phonemes
|
|
18
|
-
@objc public protocol STTDelegate {
|
|
19
|
-
@objc func stt(_ stt: DavoiceTTS.STT, didEmitEvent name: Swift.String, body: [Swift.String : Any]?)
|
|
20
|
-
}
|
|
21
|
-
@objc @_inheritsConvenienceInitializers @objcMembers final public class STT : ObjectiveC.NSObject, Speech.SFSpeechRecognizerDelegate {
|
|
22
|
-
@objc weak final public var delegate: (any DavoiceTTS.STTDelegate)?
|
|
23
|
-
@objc final public var continuous: Swift.Bool
|
|
24
|
-
@objc final public var aecEnabled: Swift.Bool
|
|
25
|
-
@objc final public var force16kMicSampleRate: Swift.Bool
|
|
26
|
-
@objc final public var useLegacySpeakerGateBehavior: Swift.Bool
|
|
27
|
-
@objc final public var useSpeakerGateHangover: Swift.Bool
|
|
28
|
-
@objc final public var speakerGateHangoverSeconds: Swift.Double
|
|
29
|
-
@objc final public var useShortSpeakerVerificationTailWindow: Swift.Bool
|
|
30
|
-
@objc final public var shortSpeakerVerificationTailSeconds: Swift.Float
|
|
31
|
-
@objc final public var speakerPreRollFlushMaxSeconds: Swift.Double
|
|
32
|
-
@objc public static let supportedEvents: [Swift.String]
|
|
33
|
-
@objc final public func setLicense(licenseKey: Swift.String) -> Swift.Bool
|
|
34
|
-
@objc final public func pauseSpeechRecognitionLite()
|
|
35
|
-
@objc final public func unPauseSpeechRecognitionLite(_ times: Foundation.NSNumber)
|
|
36
|
-
@objc(unPauseSpeechRecognitionLite:preFetch:) final public func unPauseSpeechRecognitionLite(_ times: Foundation.NSNumber, preFetch: Foundation.NSNumber)
|
|
37
|
-
@objc final public func pauseMicrophoneAndWait(_ timeoutMs: Foundation.NSNumber, completion: @escaping (Swift.Bool, Swift.String?) -> Swift.Void)
|
|
38
|
-
@objc final public func unPauseMicrophoneAndWait(_ timeoutMs: Foundation.NSNumber, completion: @escaping (Swift.Bool, Swift.String?) -> Swift.Void)
|
|
39
|
-
@objc final public func pauseMicrophone()
|
|
40
|
-
@objc final public func unPauseMicrophone()
|
|
41
|
-
@objc final public func setAECEnabled(_ enabled: Swift.Bool)
|
|
42
|
-
@objc final public func isAECEnabled() -> Swift.Bool
|
|
43
|
-
@objc final public func isSpeechAvailable(_ completion: @escaping (Swift.Bool) -> Swift.Void)
|
|
44
|
-
@objc final public func isRecognizing() -> Swift.Bool
|
|
45
|
-
@objc final public func startSpeech(localeStr: Swift.String?)
|
|
46
|
-
@objc final public func startSpeech(localeStr: Swift.String?, onboardingJsonPath: Swift.String)
|
|
47
|
-
@objc final public func stopSpeech(_ completion: ((Swift.Bool) -> Swift.Void)? = nil)
|
|
48
|
-
@objc final public func cancelSpeech(_ completion: ((Swift.Bool) -> Swift.Void)? = nil)
|
|
49
|
-
@objc final public func destroySpeech(_ completion: ((Swift.Bool) -> Swift.Void)? = nil)
|
|
50
|
-
@objc final public func teardown()
|
|
51
|
-
@objc final public func speechRecognizer(_ speechRecognizer: Speech.SFSpeechRecognizer, availabilityDidChange available: Swift.Bool)
|
|
52
|
-
@objc override dynamic public init()
|
|
53
|
-
@objc deinit
|
|
54
|
-
}
|
|
55
|
-
@objc @_inheritsConvenienceInitializers @_hasMissingDesignatedInitializers final public class SwiftSoundQueue : ObjectiveC.NSObject, AVFAudio.AVAudioPlayerDelegate {
|
|
56
|
-
@objc deinit
|
|
57
|
-
public static let shared: DavoiceTTS.SwiftSoundQueue
|
|
58
|
-
final public func notifyWhenURLFinishes(_ url: Foundation.URL, _ cb: @escaping () -> Swift.Void)
|
|
59
|
-
final public var onItemDone: ((Foundation.URL) -> Swift.Void)?
|
|
60
|
-
final public var onQueueEmpty: (() -> Swift.Void)?
|
|
61
|
-
final public func enqueue(_ source: DavoiceTTS.SwiftSoundQueue.Source)
|
|
62
|
-
final public func enqueueMany(_ sources: [DavoiceTTS.SwiftSoundQueue.Source])
|
|
63
|
-
final public func stop()
|
|
64
|
-
public enum Source {
|
|
65
|
-
case fileURL(Foundation.URL)
|
|
66
|
-
case named(Swift.String)
|
|
67
|
-
}
|
|
68
|
-
@objc final public func audioPlayerDidFinishPlaying(_ p: AVFAudio.AVAudioPlayer, successfully ok: Swift.Bool)
|
|
69
|
-
final public func activatePlaybackOnlySession()
|
|
70
|
-
final public func activateSpeaker()
|
|
71
|
-
}
|
|
72
|
-
@objc @objcMembers final public class DaVoiceTTS : ObjectiveC.NSObject {
|
|
73
|
-
public struct Config {
|
|
74
|
-
}
|
|
75
|
-
@objc final public var onLastUtteranceFinished: (() -> Swift.Void)?
|
|
76
|
-
@objc deinit
|
|
77
|
-
@objc public static func activateLicense(licenseKey: Swift.String) -> Swift.Bool
|
|
78
|
-
@objc final public func setLicense(licenseKey: Swift.String) -> Swift.Bool
|
|
79
|
-
public struct Biquad {
|
|
80
|
-
public var b0: Swift.Float, b1: Swift.Float, b2: Swift.Float, a1: Swift.Float, a2: Swift.Float
|
|
81
|
-
public init(b0: Swift.Float, b1: Swift.Float, b2: Swift.Float, a1: Swift.Float, a2: Swift.Float)
|
|
82
|
-
public mutating func process(_ x: Swift.UnsafeMutablePointer<Swift.Float>, _ n: Swift.Int)
|
|
83
|
-
}
|
|
84
|
-
@objc public init(model: Foundation.URL) throws
|
|
85
|
-
@objc final public func stopSpeaking()
|
|
86
|
-
@objc final public func destroy()
|
|
87
|
-
@objc(speak:sid:) final public func speak(_ txt: Swift.String, sid: Swift.Int32 = 0)
|
|
88
|
-
@objc(speak:sid:speed:) final public func speak(_ txt: Swift.String, sid: Swift.Int32 = 0, speed: Swift.Float)
|
|
89
|
-
@objc final public func synthesize_top(_ _text: Swift.String, speakerId: Swift.Int32 = 0, token: Foundation.UUID, speed_adjuster: Swift.Float) throws -> AVFAudio.AVAudioPCMBuffer
|
|
90
|
-
@objc final public func playWav(_ url: Foundation.URL, markAsLastUtterance: Swift.Bool = true)
|
|
91
|
-
@objc final public func playBuffer(_ buffer: AVFAudio.AVAudioPCMBuffer, markAsLastUtterance: Swift.Bool = true)
|
|
92
|
-
}
|
|
93
18
|
public typealias EngineSchedule = (_ url: Foundation.URL, _ onDone: @escaping () -> Swift.Void) -> Swift.Bool
|
|
94
19
|
public typealias IsEngineReady = () -> Swift.Bool
|
|
95
20
|
public typealias useOnlyEnginePlayback = () -> Swift.Bool
|
|
@@ -102,11 +27,6 @@ public enum AudioPlaybackHook {
|
|
|
102
27
|
public static var stopEnginePlayback: DavoiceTTS.StopEnginePlayback?
|
|
103
28
|
public static var currentEngine: DavoiceTTS.CurrentEngineProvider?
|
|
104
29
|
}
|
|
105
|
-
@_inheritsConvenienceInitializers @objc final public class LicenseManager : ObjectiveC.NSObject {
|
|
106
|
-
@objc public static func isLicenseValid(licenseKey: Swift.String) -> Swift.Bool
|
|
107
|
-
@objc override dynamic public init()
|
|
108
|
-
@objc deinit
|
|
109
|
-
}
|
|
110
30
|
public enum SVLogLevel : Swift.Int, Swift.Codable {
|
|
111
31
|
case off, error, warn, info, debug, trace
|
|
112
32
|
public init?(rawValue: Swift.Int)
|
|
@@ -209,6 +129,88 @@ public enum SpeakerVerificationError : Swift.Error, Swift.CustomStringConvertibl
|
|
|
209
129
|
@objc override dynamic public init()
|
|
210
130
|
@objc deinit
|
|
211
131
|
}
|
|
132
|
+
@objc public protocol STTDelegate {
|
|
133
|
+
@objc func stt(_ stt: DavoiceTTS.STT, didEmitEvent name: Swift.String, body: [Swift.String : Any]?)
|
|
134
|
+
}
|
|
135
|
+
@objc @_inheritsConvenienceInitializers @objcMembers final public class STT : ObjectiveC.NSObject, Speech.SFSpeechRecognizerDelegate {
|
|
136
|
+
@objc weak final public var delegate: (any DavoiceTTS.STTDelegate)?
|
|
137
|
+
@objc final public var continuous: Swift.Bool
|
|
138
|
+
@objc final public var aecEnabled: Swift.Bool
|
|
139
|
+
@objc final public var force16kMicSampleRate: Swift.Bool
|
|
140
|
+
@objc final public var useLegacySpeakerGateBehavior: Swift.Bool
|
|
141
|
+
@objc final public var useSpeakerGateHangover: Swift.Bool
|
|
142
|
+
@objc final public var speakerGateHangoverSeconds: Swift.Double
|
|
143
|
+
@objc final public var useShortSpeakerVerificationTailWindow: Swift.Bool
|
|
144
|
+
@objc final public var shortSpeakerVerificationTailSeconds: Swift.Float
|
|
145
|
+
@objc final public var speakerPreRollFlushMaxSeconds: Swift.Double
|
|
146
|
+
@objc public static let supportedEvents: [Swift.String]
|
|
147
|
+
@objc final public func setLicense(licenseKey: Swift.String) -> Swift.Bool
|
|
148
|
+
@objc final public func pauseSpeechRecognitionLite()
|
|
149
|
+
@objc final public func pauseSpeechRecognitionLiteAndWait(_ timeoutMs: Foundation.NSNumber, completion: @escaping (Swift.Bool, Swift.String?) -> Swift.Void)
|
|
150
|
+
@objc final public func unPauseSpeechRecognitionLite(_ times: Foundation.NSNumber)
|
|
151
|
+
@objc(unPauseSpeechRecognitionLite:preFetch:) final public func unPauseSpeechRecognitionLite(_ times: Foundation.NSNumber, preFetch: Foundation.NSNumber)
|
|
152
|
+
@objc final public func unPauseSpeechRecognitionLiteAndWait(_ times: Foundation.NSNumber, preFetch: Foundation.NSNumber, timeoutMs: Foundation.NSNumber, completion: @escaping (Swift.Bool, Swift.String?) -> Swift.Void)
|
|
153
|
+
@objc final public func pauseMicrophoneAndWait(_ timeoutMs: Foundation.NSNumber, completion: @escaping (Swift.Bool, Swift.String?) -> Swift.Void)
|
|
154
|
+
@objc final public func unPauseMicrophoneAndWait(_ timeoutMs: Foundation.NSNumber, completion: @escaping (Swift.Bool, Swift.String?) -> Swift.Void)
|
|
155
|
+
@objc final public func pauseMicrophone()
|
|
156
|
+
@objc final public func unPauseMicrophone()
|
|
157
|
+
@objc final public func setAECEnabled(_ enabled: Swift.Bool)
|
|
158
|
+
@objc final public func isAECEnabled() -> Swift.Bool
|
|
159
|
+
@objc final public func isSpeechAvailable(_ completion: @escaping (Swift.Bool) -> Swift.Void)
|
|
160
|
+
@objc final public func isRecognizing() -> Swift.Bool
|
|
161
|
+
@objc final public func startSpeech(localeStr: Swift.String?)
|
|
162
|
+
@objc final public func startSpeech(localeStr: Swift.String?, onboardingJsonPath: Swift.String)
|
|
163
|
+
@objc final public func stopSpeech(_ completion: ((Swift.Bool) -> Swift.Void)? = nil)
|
|
164
|
+
@objc final public func cancelSpeech(_ completion: ((Swift.Bool) -> Swift.Void)? = nil)
|
|
165
|
+
@objc final public func destroySpeech(_ completion: ((Swift.Bool) -> Swift.Void)? = nil)
|
|
166
|
+
@objc final public func teardown()
|
|
167
|
+
@objc final public func speechRecognizer(_ speechRecognizer: Speech.SFSpeechRecognizer, availabilityDidChange available: Swift.Bool)
|
|
168
|
+
@objc override dynamic public init()
|
|
169
|
+
@objc deinit
|
|
170
|
+
}
|
|
171
|
+
@objc @_inheritsConvenienceInitializers @_hasMissingDesignatedInitializers final public class SwiftSoundQueue : ObjectiveC.NSObject, AVFAudio.AVAudioPlayerDelegate {
|
|
172
|
+
@objc deinit
|
|
173
|
+
public static let shared: DavoiceTTS.SwiftSoundQueue
|
|
174
|
+
final public func notifyWhenURLFinishes(_ url: Foundation.URL, _ cb: @escaping () -> Swift.Void)
|
|
175
|
+
final public var onItemDone: ((Foundation.URL) -> Swift.Void)?
|
|
176
|
+
final public var onQueueEmpty: (() -> Swift.Void)?
|
|
177
|
+
final public func enqueue(_ source: DavoiceTTS.SwiftSoundQueue.Source)
|
|
178
|
+
final public func enqueueMany(_ sources: [DavoiceTTS.SwiftSoundQueue.Source])
|
|
179
|
+
final public func stop()
|
|
180
|
+
public enum Source {
|
|
181
|
+
case fileURL(Foundation.URL)
|
|
182
|
+
case named(Swift.String)
|
|
183
|
+
}
|
|
184
|
+
@objc final public func audioPlayerDidFinishPlaying(_ p: AVFAudio.AVAudioPlayer, successfully ok: Swift.Bool)
|
|
185
|
+
final public func activatePlaybackOnlySession()
|
|
186
|
+
final public func activateSpeaker()
|
|
187
|
+
}
|
|
188
|
+
@objc @objcMembers final public class DaVoiceTTS : ObjectiveC.NSObject {
|
|
189
|
+
public struct Config {
|
|
190
|
+
}
|
|
191
|
+
@objc final public var onLastUtteranceFinished: (() -> Swift.Void)?
|
|
192
|
+
@objc deinit
|
|
193
|
+
@objc public static func activateLicense(licenseKey: Swift.String) -> Swift.Bool
|
|
194
|
+
@objc final public func setLicense(licenseKey: Swift.String) -> Swift.Bool
|
|
195
|
+
public struct Biquad {
|
|
196
|
+
public var b0: Swift.Float, b1: Swift.Float, b2: Swift.Float, a1: Swift.Float, a2: Swift.Float
|
|
197
|
+
public init(b0: Swift.Float, b1: Swift.Float, b2: Swift.Float, a1: Swift.Float, a2: Swift.Float)
|
|
198
|
+
public mutating func process(_ x: Swift.UnsafeMutablePointer<Swift.Float>, _ n: Swift.Int)
|
|
199
|
+
}
|
|
200
|
+
@objc public init(model: Foundation.URL) throws
|
|
201
|
+
@objc final public func stopSpeaking()
|
|
202
|
+
@objc final public func destroy()
|
|
203
|
+
@objc(speak:sid:) final public func speak(_ txt: Swift.String, sid: Swift.Int32 = 0)
|
|
204
|
+
@objc(speak:sid:speed:) final public func speak(_ txt: Swift.String, sid: Swift.Int32 = 0, speed: Swift.Float)
|
|
205
|
+
@objc final public func synthesize_top(_ _text: Swift.String, speakerId: Swift.Int32 = 0, token: Foundation.UUID, speed_adjuster: Swift.Float) throws -> AVFAudio.AVAudioPCMBuffer
|
|
206
|
+
@objc final public func playWav(_ url: Foundation.URL, markAsLastUtterance: Swift.Bool = true)
|
|
207
|
+
@objc final public func playBuffer(_ buffer: AVFAudio.AVAudioPCMBuffer, markAsLastUtterance: Swift.Bool = true)
|
|
208
|
+
}
|
|
209
|
+
@_inheritsConvenienceInitializers @objc final public class LicenseManager : ObjectiveC.NSObject {
|
|
210
|
+
@objc public static func isLicenseValid(licenseKey: Swift.String) -> Swift.Bool
|
|
211
|
+
@objc override dynamic public init()
|
|
212
|
+
@objc deinit
|
|
213
|
+
}
|
|
212
214
|
extension DavoiceTTS.SVLogLevel : Swift.Equatable {}
|
|
213
215
|
extension DavoiceTTS.SVLogLevel : Swift.Hashable {}
|
|
214
216
|
extension DavoiceTTS.SVLogLevel : Swift.RawRepresentable {}
|
|
@@ -15,81 +15,6 @@ import _StringProcessing
|
|
|
15
15
|
import _SwiftConcurrencyShims
|
|
16
16
|
import onnxruntime_objc
|
|
17
17
|
import phonemes
|
|
18
|
-
@objc public protocol STTDelegate {
|
|
19
|
-
@objc func stt(_ stt: DavoiceTTS.STT, didEmitEvent name: Swift.String, body: [Swift.String : Any]?)
|
|
20
|
-
}
|
|
21
|
-
@objc @_inheritsConvenienceInitializers @objcMembers final public class STT : ObjectiveC.NSObject, Speech.SFSpeechRecognizerDelegate {
|
|
22
|
-
@objc weak final public var delegate: (any DavoiceTTS.STTDelegate)?
|
|
23
|
-
@objc final public var continuous: Swift.Bool
|
|
24
|
-
@objc final public var aecEnabled: Swift.Bool
|
|
25
|
-
@objc final public var force16kMicSampleRate: Swift.Bool
|
|
26
|
-
@objc final public var useLegacySpeakerGateBehavior: Swift.Bool
|
|
27
|
-
@objc final public var useSpeakerGateHangover: Swift.Bool
|
|
28
|
-
@objc final public var speakerGateHangoverSeconds: Swift.Double
|
|
29
|
-
@objc final public var useShortSpeakerVerificationTailWindow: Swift.Bool
|
|
30
|
-
@objc final public var shortSpeakerVerificationTailSeconds: Swift.Float
|
|
31
|
-
@objc final public var speakerPreRollFlushMaxSeconds: Swift.Double
|
|
32
|
-
@objc public static let supportedEvents: [Swift.String]
|
|
33
|
-
@objc final public func setLicense(licenseKey: Swift.String) -> Swift.Bool
|
|
34
|
-
@objc final public func pauseSpeechRecognitionLite()
|
|
35
|
-
@objc final public func unPauseSpeechRecognitionLite(_ times: Foundation.NSNumber)
|
|
36
|
-
@objc(unPauseSpeechRecognitionLite:preFetch:) final public func unPauseSpeechRecognitionLite(_ times: Foundation.NSNumber, preFetch: Foundation.NSNumber)
|
|
37
|
-
@objc final public func pauseMicrophoneAndWait(_ timeoutMs: Foundation.NSNumber, completion: @escaping (Swift.Bool, Swift.String?) -> Swift.Void)
|
|
38
|
-
@objc final public func unPauseMicrophoneAndWait(_ timeoutMs: Foundation.NSNumber, completion: @escaping (Swift.Bool, Swift.String?) -> Swift.Void)
|
|
39
|
-
@objc final public func pauseMicrophone()
|
|
40
|
-
@objc final public func unPauseMicrophone()
|
|
41
|
-
@objc final public func setAECEnabled(_ enabled: Swift.Bool)
|
|
42
|
-
@objc final public func isAECEnabled() -> Swift.Bool
|
|
43
|
-
@objc final public func isSpeechAvailable(_ completion: @escaping (Swift.Bool) -> Swift.Void)
|
|
44
|
-
@objc final public func isRecognizing() -> Swift.Bool
|
|
45
|
-
@objc final public func startSpeech(localeStr: Swift.String?)
|
|
46
|
-
@objc final public func startSpeech(localeStr: Swift.String?, onboardingJsonPath: Swift.String)
|
|
47
|
-
@objc final public func stopSpeech(_ completion: ((Swift.Bool) -> Swift.Void)? = nil)
|
|
48
|
-
@objc final public func cancelSpeech(_ completion: ((Swift.Bool) -> Swift.Void)? = nil)
|
|
49
|
-
@objc final public func destroySpeech(_ completion: ((Swift.Bool) -> Swift.Void)? = nil)
|
|
50
|
-
@objc final public func teardown()
|
|
51
|
-
@objc final public func speechRecognizer(_ speechRecognizer: Speech.SFSpeechRecognizer, availabilityDidChange available: Swift.Bool)
|
|
52
|
-
@objc override dynamic public init()
|
|
53
|
-
@objc deinit
|
|
54
|
-
}
|
|
55
|
-
@objc @_inheritsConvenienceInitializers @_hasMissingDesignatedInitializers final public class SwiftSoundQueue : ObjectiveC.NSObject, AVFAudio.AVAudioPlayerDelegate {
|
|
56
|
-
@objc deinit
|
|
57
|
-
public static let shared: DavoiceTTS.SwiftSoundQueue
|
|
58
|
-
final public func notifyWhenURLFinishes(_ url: Foundation.URL, _ cb: @escaping () -> Swift.Void)
|
|
59
|
-
final public var onItemDone: ((Foundation.URL) -> Swift.Void)?
|
|
60
|
-
final public var onQueueEmpty: (() -> Swift.Void)?
|
|
61
|
-
final public func enqueue(_ source: DavoiceTTS.SwiftSoundQueue.Source)
|
|
62
|
-
final public func enqueueMany(_ sources: [DavoiceTTS.SwiftSoundQueue.Source])
|
|
63
|
-
final public func stop()
|
|
64
|
-
public enum Source {
|
|
65
|
-
case fileURL(Foundation.URL)
|
|
66
|
-
case named(Swift.String)
|
|
67
|
-
}
|
|
68
|
-
@objc final public func audioPlayerDidFinishPlaying(_ p: AVFAudio.AVAudioPlayer, successfully ok: Swift.Bool)
|
|
69
|
-
final public func activatePlaybackOnlySession()
|
|
70
|
-
final public func activateSpeaker()
|
|
71
|
-
}
|
|
72
|
-
@objc @objcMembers final public class DaVoiceTTS : ObjectiveC.NSObject {
|
|
73
|
-
public struct Config {
|
|
74
|
-
}
|
|
75
|
-
@objc final public var onLastUtteranceFinished: (() -> Swift.Void)?
|
|
76
|
-
@objc deinit
|
|
77
|
-
@objc public static func activateLicense(licenseKey: Swift.String) -> Swift.Bool
|
|
78
|
-
@objc final public func setLicense(licenseKey: Swift.String) -> Swift.Bool
|
|
79
|
-
public struct Biquad {
|
|
80
|
-
public var b0: Swift.Float, b1: Swift.Float, b2: Swift.Float, a1: Swift.Float, a2: Swift.Float
|
|
81
|
-
public init(b0: Swift.Float, b1: Swift.Float, b2: Swift.Float, a1: Swift.Float, a2: Swift.Float)
|
|
82
|
-
public mutating func process(_ x: Swift.UnsafeMutablePointer<Swift.Float>, _ n: Swift.Int)
|
|
83
|
-
}
|
|
84
|
-
@objc public init(model: Foundation.URL) throws
|
|
85
|
-
@objc final public func stopSpeaking()
|
|
86
|
-
@objc final public func destroy()
|
|
87
|
-
@objc(speak:sid:) final public func speak(_ txt: Swift.String, sid: Swift.Int32 = 0)
|
|
88
|
-
@objc(speak:sid:speed:) final public func speak(_ txt: Swift.String, sid: Swift.Int32 = 0, speed: Swift.Float)
|
|
89
|
-
@objc final public func synthesize_top(_ _text: Swift.String, speakerId: Swift.Int32 = 0, token: Foundation.UUID, speed_adjuster: Swift.Float) throws -> AVFAudio.AVAudioPCMBuffer
|
|
90
|
-
@objc final public func playWav(_ url: Foundation.URL, markAsLastUtterance: Swift.Bool = true)
|
|
91
|
-
@objc final public func playBuffer(_ buffer: AVFAudio.AVAudioPCMBuffer, markAsLastUtterance: Swift.Bool = true)
|
|
92
|
-
}
|
|
93
18
|
public typealias EngineSchedule = (_ url: Foundation.URL, _ onDone: @escaping () -> Swift.Void) -> Swift.Bool
|
|
94
19
|
public typealias IsEngineReady = () -> Swift.Bool
|
|
95
20
|
public typealias useOnlyEnginePlayback = () -> Swift.Bool
|
|
@@ -102,11 +27,6 @@ public enum AudioPlaybackHook {
|
|
|
102
27
|
public static var stopEnginePlayback: DavoiceTTS.StopEnginePlayback?
|
|
103
28
|
public static var currentEngine: DavoiceTTS.CurrentEngineProvider?
|
|
104
29
|
}
|
|
105
|
-
@_inheritsConvenienceInitializers @objc final public class LicenseManager : ObjectiveC.NSObject {
|
|
106
|
-
@objc public static func isLicenseValid(licenseKey: Swift.String) -> Swift.Bool
|
|
107
|
-
@objc override dynamic public init()
|
|
108
|
-
@objc deinit
|
|
109
|
-
}
|
|
110
30
|
public enum SVLogLevel : Swift.Int, Swift.Codable {
|
|
111
31
|
case off, error, warn, info, debug, trace
|
|
112
32
|
public init?(rawValue: Swift.Int)
|
|
@@ -209,6 +129,88 @@ public enum SpeakerVerificationError : Swift.Error, Swift.CustomStringConvertibl
|
|
|
209
129
|
@objc override dynamic public init()
|
|
210
130
|
@objc deinit
|
|
211
131
|
}
|
|
132
|
+
@objc public protocol STTDelegate {
|
|
133
|
+
@objc func stt(_ stt: DavoiceTTS.STT, didEmitEvent name: Swift.String, body: [Swift.String : Any]?)
|
|
134
|
+
}
|
|
135
|
+
@objc @_inheritsConvenienceInitializers @objcMembers final public class STT : ObjectiveC.NSObject, Speech.SFSpeechRecognizerDelegate {
|
|
136
|
+
@objc weak final public var delegate: (any DavoiceTTS.STTDelegate)?
|
|
137
|
+
@objc final public var continuous: Swift.Bool
|
|
138
|
+
@objc final public var aecEnabled: Swift.Bool
|
|
139
|
+
@objc final public var force16kMicSampleRate: Swift.Bool
|
|
140
|
+
@objc final public var useLegacySpeakerGateBehavior: Swift.Bool
|
|
141
|
+
@objc final public var useSpeakerGateHangover: Swift.Bool
|
|
142
|
+
@objc final public var speakerGateHangoverSeconds: Swift.Double
|
|
143
|
+
@objc final public var useShortSpeakerVerificationTailWindow: Swift.Bool
|
|
144
|
+
@objc final public var shortSpeakerVerificationTailSeconds: Swift.Float
|
|
145
|
+
@objc final public var speakerPreRollFlushMaxSeconds: Swift.Double
|
|
146
|
+
@objc public static let supportedEvents: [Swift.String]
|
|
147
|
+
@objc final public func setLicense(licenseKey: Swift.String) -> Swift.Bool
|
|
148
|
+
@objc final public func pauseSpeechRecognitionLite()
|
|
149
|
+
@objc final public func pauseSpeechRecognitionLiteAndWait(_ timeoutMs: Foundation.NSNumber, completion: @escaping (Swift.Bool, Swift.String?) -> Swift.Void)
|
|
150
|
+
@objc final public func unPauseSpeechRecognitionLite(_ times: Foundation.NSNumber)
|
|
151
|
+
@objc(unPauseSpeechRecognitionLite:preFetch:) final public func unPauseSpeechRecognitionLite(_ times: Foundation.NSNumber, preFetch: Foundation.NSNumber)
|
|
152
|
+
@objc final public func unPauseSpeechRecognitionLiteAndWait(_ times: Foundation.NSNumber, preFetch: Foundation.NSNumber, timeoutMs: Foundation.NSNumber, completion: @escaping (Swift.Bool, Swift.String?) -> Swift.Void)
|
|
153
|
+
@objc final public func pauseMicrophoneAndWait(_ timeoutMs: Foundation.NSNumber, completion: @escaping (Swift.Bool, Swift.String?) -> Swift.Void)
|
|
154
|
+
@objc final public func unPauseMicrophoneAndWait(_ timeoutMs: Foundation.NSNumber, completion: @escaping (Swift.Bool, Swift.String?) -> Swift.Void)
|
|
155
|
+
@objc final public func pauseMicrophone()
|
|
156
|
+
@objc final public func unPauseMicrophone()
|
|
157
|
+
@objc final public func setAECEnabled(_ enabled: Swift.Bool)
|
|
158
|
+
@objc final public func isAECEnabled() -> Swift.Bool
|
|
159
|
+
@objc final public func isSpeechAvailable(_ completion: @escaping (Swift.Bool) -> Swift.Void)
|
|
160
|
+
@objc final public func isRecognizing() -> Swift.Bool
|
|
161
|
+
@objc final public func startSpeech(localeStr: Swift.String?)
|
|
162
|
+
@objc final public func startSpeech(localeStr: Swift.String?, onboardingJsonPath: Swift.String)
|
|
163
|
+
@objc final public func stopSpeech(_ completion: ((Swift.Bool) -> Swift.Void)? = nil)
|
|
164
|
+
@objc final public func cancelSpeech(_ completion: ((Swift.Bool) -> Swift.Void)? = nil)
|
|
165
|
+
@objc final public func destroySpeech(_ completion: ((Swift.Bool) -> Swift.Void)? = nil)
|
|
166
|
+
@objc final public func teardown()
|
|
167
|
+
@objc final public func speechRecognizer(_ speechRecognizer: Speech.SFSpeechRecognizer, availabilityDidChange available: Swift.Bool)
|
|
168
|
+
@objc override dynamic public init()
|
|
169
|
+
@objc deinit
|
|
170
|
+
}
|
|
171
|
+
@objc @_inheritsConvenienceInitializers @_hasMissingDesignatedInitializers final public class SwiftSoundQueue : ObjectiveC.NSObject, AVFAudio.AVAudioPlayerDelegate {
|
|
172
|
+
@objc deinit
|
|
173
|
+
public static let shared: DavoiceTTS.SwiftSoundQueue
|
|
174
|
+
final public func notifyWhenURLFinishes(_ url: Foundation.URL, _ cb: @escaping () -> Swift.Void)
|
|
175
|
+
final public var onItemDone: ((Foundation.URL) -> Swift.Void)?
|
|
176
|
+
final public var onQueueEmpty: (() -> Swift.Void)?
|
|
177
|
+
final public func enqueue(_ source: DavoiceTTS.SwiftSoundQueue.Source)
|
|
178
|
+
final public func enqueueMany(_ sources: [DavoiceTTS.SwiftSoundQueue.Source])
|
|
179
|
+
final public func stop()
|
|
180
|
+
public enum Source {
|
|
181
|
+
case fileURL(Foundation.URL)
|
|
182
|
+
case named(Swift.String)
|
|
183
|
+
}
|
|
184
|
+
@objc final public func audioPlayerDidFinishPlaying(_ p: AVFAudio.AVAudioPlayer, successfully ok: Swift.Bool)
|
|
185
|
+
final public func activatePlaybackOnlySession()
|
|
186
|
+
final public func activateSpeaker()
|
|
187
|
+
}
|
|
188
|
+
@objc @objcMembers final public class DaVoiceTTS : ObjectiveC.NSObject {
|
|
189
|
+
public struct Config {
|
|
190
|
+
}
|
|
191
|
+
@objc final public var onLastUtteranceFinished: (() -> Swift.Void)?
|
|
192
|
+
@objc deinit
|
|
193
|
+
@objc public static func activateLicense(licenseKey: Swift.String) -> Swift.Bool
|
|
194
|
+
@objc final public func setLicense(licenseKey: Swift.String) -> Swift.Bool
|
|
195
|
+
public struct Biquad {
|
|
196
|
+
public var b0: Swift.Float, b1: Swift.Float, b2: Swift.Float, a1: Swift.Float, a2: Swift.Float
|
|
197
|
+
public init(b0: Swift.Float, b1: Swift.Float, b2: Swift.Float, a1: Swift.Float, a2: Swift.Float)
|
|
198
|
+
public mutating func process(_ x: Swift.UnsafeMutablePointer<Swift.Float>, _ n: Swift.Int)
|
|
199
|
+
}
|
|
200
|
+
@objc public init(model: Foundation.URL) throws
|
|
201
|
+
@objc final public func stopSpeaking()
|
|
202
|
+
@objc final public func destroy()
|
|
203
|
+
@objc(speak:sid:) final public func speak(_ txt: Swift.String, sid: Swift.Int32 = 0)
|
|
204
|
+
@objc(speak:sid:speed:) final public func speak(_ txt: Swift.String, sid: Swift.Int32 = 0, speed: Swift.Float)
|
|
205
|
+
@objc final public func synthesize_top(_ _text: Swift.String, speakerId: Swift.Int32 = 0, token: Foundation.UUID, speed_adjuster: Swift.Float) throws -> AVFAudio.AVAudioPCMBuffer
|
|
206
|
+
@objc final public func playWav(_ url: Foundation.URL, markAsLastUtterance: Swift.Bool = true)
|
|
207
|
+
@objc final public func playBuffer(_ buffer: AVFAudio.AVAudioPCMBuffer, markAsLastUtterance: Swift.Bool = true)
|
|
208
|
+
}
|
|
209
|
+
@_inheritsConvenienceInitializers @objc final public class LicenseManager : ObjectiveC.NSObject {
|
|
210
|
+
@objc public static func isLicenseValid(licenseKey: Swift.String) -> Swift.Bool
|
|
211
|
+
@objc override dynamic public init()
|
|
212
|
+
@objc deinit
|
|
213
|
+
}
|
|
212
214
|
extension DavoiceTTS.SVLogLevel : Swift.Equatable {}
|
|
213
215
|
extension DavoiceTTS.SVLogLevel : Swift.Hashable {}
|
|
214
216
|
extension DavoiceTTS.SVLogLevel : Swift.RawRepresentable {}
|
|
Binary file
|
|
Binary file
|
|
@@ -6,7 +6,7 @@
|
|
|
6
6
|
<dict>
|
|
7
7
|
<key>Headers/DavoiceTTS-Swift.h</key>
|
|
8
8
|
<data>
|
|
9
|
-
|
|
9
|
+
O9fuulSbauw2eQFehC0CQllYmK0=
|
|
10
10
|
</data>
|
|
11
11
|
<key>Info.plist</key>
|
|
12
12
|
<data>
|
|
@@ -14,11 +14,11 @@
|
|
|
14
14
|
</data>
|
|
15
15
|
<key>Modules/DavoiceTTS.swiftmodule/arm64-apple-ios-simulator.abi.json</key>
|
|
16
16
|
<data>
|
|
17
|
-
|
|
17
|
+
OVe8F+zNw7x1MUtcM2pki2hBsTY=
|
|
18
18
|
</data>
|
|
19
19
|
<key>Modules/DavoiceTTS.swiftmodule/arm64-apple-ios-simulator.private.swiftinterface</key>
|
|
20
20
|
<data>
|
|
21
|
-
|
|
21
|
+
EwWDjYnAIYKIdMgvrEi1c/2tGj0=
|
|
22
22
|
</data>
|
|
23
23
|
<key>Modules/DavoiceTTS.swiftmodule/arm64-apple-ios-simulator.swiftdoc</key>
|
|
24
24
|
<data>
|
|
@@ -26,19 +26,19 @@
|
|
|
26
26
|
</data>
|
|
27
27
|
<key>Modules/DavoiceTTS.swiftmodule/arm64-apple-ios-simulator.swiftinterface</key>
|
|
28
28
|
<data>
|
|
29
|
-
|
|
29
|
+
EwWDjYnAIYKIdMgvrEi1c/2tGj0=
|
|
30
30
|
</data>
|
|
31
31
|
<key>Modules/DavoiceTTS.swiftmodule/arm64-apple-ios-simulator.swiftmodule</key>
|
|
32
32
|
<data>
|
|
33
|
-
|
|
33
|
+
/GURfoag4NeuE1qjppNBFUKDrmg=
|
|
34
34
|
</data>
|
|
35
35
|
<key>Modules/DavoiceTTS.swiftmodule/x86_64-apple-ios-simulator.abi.json</key>
|
|
36
36
|
<data>
|
|
37
|
-
|
|
37
|
+
OVe8F+zNw7x1MUtcM2pki2hBsTY=
|
|
38
38
|
</data>
|
|
39
39
|
<key>Modules/DavoiceTTS.swiftmodule/x86_64-apple-ios-simulator.private.swiftinterface</key>
|
|
40
40
|
<data>
|
|
41
|
-
|
|
41
|
+
/ac7kA5Qr0tE1ZWnK54KYSRP6Kg=
|
|
42
42
|
</data>
|
|
43
43
|
<key>Modules/DavoiceTTS.swiftmodule/x86_64-apple-ios-simulator.swiftdoc</key>
|
|
44
44
|
<data>
|
|
@@ -46,11 +46,11 @@
|
|
|
46
46
|
</data>
|
|
47
47
|
<key>Modules/DavoiceTTS.swiftmodule/x86_64-apple-ios-simulator.swiftinterface</key>
|
|
48
48
|
<data>
|
|
49
|
-
|
|
49
|
+
/ac7kA5Qr0tE1ZWnK54KYSRP6Kg=
|
|
50
50
|
</data>
|
|
51
51
|
<key>Modules/DavoiceTTS.swiftmodule/x86_64-apple-ios-simulator.swiftmodule</key>
|
|
52
52
|
<data>
|
|
53
|
-
|
|
53
|
+
sQ4eyOVpPbKHbdw9V5WgekD1X00=
|
|
54
54
|
</data>
|
|
55
55
|
<key>Modules/module.modulemap</key>
|
|
56
56
|
<data>
|
|
@@ -63,33 +63,33 @@
|
|
|
63
63
|
<dict>
|
|
64
64
|
<key>hash</key>
|
|
65
65
|
<data>
|
|
66
|
-
|
|
66
|
+
O9fuulSbauw2eQFehC0CQllYmK0=
|
|
67
67
|
</data>
|
|
68
68
|
<key>hash2</key>
|
|
69
69
|
<data>
|
|
70
|
-
|
|
70
|
+
Y0gWRE1HbAh+Ep+eN90BRhHRY+BHK6RCoEHXGwJT6yU=
|
|
71
71
|
</data>
|
|
72
72
|
</dict>
|
|
73
73
|
<key>Modules/DavoiceTTS.swiftmodule/arm64-apple-ios-simulator.abi.json</key>
|
|
74
74
|
<dict>
|
|
75
75
|
<key>hash</key>
|
|
76
76
|
<data>
|
|
77
|
-
|
|
77
|
+
OVe8F+zNw7x1MUtcM2pki2hBsTY=
|
|
78
78
|
</data>
|
|
79
79
|
<key>hash2</key>
|
|
80
80
|
<data>
|
|
81
|
-
|
|
81
|
+
yvNv0jWONZRKhYwFB4r1ckjMBPtFOTYQFZP1tORGZaU=
|
|
82
82
|
</data>
|
|
83
83
|
</dict>
|
|
84
84
|
<key>Modules/DavoiceTTS.swiftmodule/arm64-apple-ios-simulator.private.swiftinterface</key>
|
|
85
85
|
<dict>
|
|
86
86
|
<key>hash</key>
|
|
87
87
|
<data>
|
|
88
|
-
|
|
88
|
+
EwWDjYnAIYKIdMgvrEi1c/2tGj0=
|
|
89
89
|
</data>
|
|
90
90
|
<key>hash2</key>
|
|
91
91
|
<data>
|
|
92
|
-
|
|
92
|
+
Qh3/z8VPveDXEeenp/KaCkoThRYdoMTkh0YRjieXtyA=
|
|
93
93
|
</data>
|
|
94
94
|
</dict>
|
|
95
95
|
<key>Modules/DavoiceTTS.swiftmodule/arm64-apple-ios-simulator.swiftdoc</key>
|
|
@@ -107,44 +107,44 @@
|
|
|
107
107
|
<dict>
|
|
108
108
|
<key>hash</key>
|
|
109
109
|
<data>
|
|
110
|
-
|
|
110
|
+
EwWDjYnAIYKIdMgvrEi1c/2tGj0=
|
|
111
111
|
</data>
|
|
112
112
|
<key>hash2</key>
|
|
113
113
|
<data>
|
|
114
|
-
|
|
114
|
+
Qh3/z8VPveDXEeenp/KaCkoThRYdoMTkh0YRjieXtyA=
|
|
115
115
|
</data>
|
|
116
116
|
</dict>
|
|
117
117
|
<key>Modules/DavoiceTTS.swiftmodule/arm64-apple-ios-simulator.swiftmodule</key>
|
|
118
118
|
<dict>
|
|
119
119
|
<key>hash</key>
|
|
120
120
|
<data>
|
|
121
|
-
|
|
121
|
+
/GURfoag4NeuE1qjppNBFUKDrmg=
|
|
122
122
|
</data>
|
|
123
123
|
<key>hash2</key>
|
|
124
124
|
<data>
|
|
125
|
-
|
|
125
|
+
MizfQRzexLlGnSj5wkb+LbFdA9zNwYGp/G2SAyYWbtk=
|
|
126
126
|
</data>
|
|
127
127
|
</dict>
|
|
128
128
|
<key>Modules/DavoiceTTS.swiftmodule/x86_64-apple-ios-simulator.abi.json</key>
|
|
129
129
|
<dict>
|
|
130
130
|
<key>hash</key>
|
|
131
131
|
<data>
|
|
132
|
-
|
|
132
|
+
OVe8F+zNw7x1MUtcM2pki2hBsTY=
|
|
133
133
|
</data>
|
|
134
134
|
<key>hash2</key>
|
|
135
135
|
<data>
|
|
136
|
-
|
|
136
|
+
yvNv0jWONZRKhYwFB4r1ckjMBPtFOTYQFZP1tORGZaU=
|
|
137
137
|
</data>
|
|
138
138
|
</dict>
|
|
139
139
|
<key>Modules/DavoiceTTS.swiftmodule/x86_64-apple-ios-simulator.private.swiftinterface</key>
|
|
140
140
|
<dict>
|
|
141
141
|
<key>hash</key>
|
|
142
142
|
<data>
|
|
143
|
-
|
|
143
|
+
/ac7kA5Qr0tE1ZWnK54KYSRP6Kg=
|
|
144
144
|
</data>
|
|
145
145
|
<key>hash2</key>
|
|
146
146
|
<data>
|
|
147
|
-
|
|
147
|
+
uwr4QOXNWm0yUsi8gROp5E2tSPT9U/dzEjDmb4h6E18=
|
|
148
148
|
</data>
|
|
149
149
|
</dict>
|
|
150
150
|
<key>Modules/DavoiceTTS.swiftmodule/x86_64-apple-ios-simulator.swiftdoc</key>
|
|
@@ -162,22 +162,22 @@
|
|
|
162
162
|
<dict>
|
|
163
163
|
<key>hash</key>
|
|
164
164
|
<data>
|
|
165
|
-
|
|
165
|
+
/ac7kA5Qr0tE1ZWnK54KYSRP6Kg=
|
|
166
166
|
</data>
|
|
167
167
|
<key>hash2</key>
|
|
168
168
|
<data>
|
|
169
|
-
|
|
169
|
+
uwr4QOXNWm0yUsi8gROp5E2tSPT9U/dzEjDmb4h6E18=
|
|
170
170
|
</data>
|
|
171
171
|
</dict>
|
|
172
172
|
<key>Modules/DavoiceTTS.swiftmodule/x86_64-apple-ios-simulator.swiftmodule</key>
|
|
173
173
|
<dict>
|
|
174
174
|
<key>hash</key>
|
|
175
175
|
<data>
|
|
176
|
-
|
|
176
|
+
sQ4eyOVpPbKHbdw9V5WgekD1X00=
|
|
177
177
|
</data>
|
|
178
178
|
<key>hash2</key>
|
|
179
179
|
<data>
|
|
180
|
-
|
|
180
|
+
fbn7hMROMn8KBMieplX2hJ7lpb8GQVNVkiZiM84vYOA=
|
|
181
181
|
</data>
|
|
182
182
|
</dict>
|
|
183
183
|
<key>Modules/module.modulemap</key>
|
package/package.json
CHANGED
package/speech/index.ts
CHANGED
|
@@ -471,6 +471,10 @@ class Speech {
|
|
|
471
471
|
async start(locale: string, options: Record<string, any> = {}) {
|
|
472
472
|
this.ensureListeners();
|
|
473
473
|
// Prefer unified on iOS
|
|
474
|
+
if (Platform.OS === 'ios' && (NativeSpeech as any)?.startSpeechAsync) {
|
|
475
|
+
await (NativeSpeech as any).startSpeechAsync(locale, 2500);
|
|
476
|
+
return;
|
|
477
|
+
}
|
|
474
478
|
if (Platform.OS === 'ios' && NativeSpeech?.startSpeech) {
|
|
475
479
|
return new Promise<void>((resolve) => NativeSpeech.startSpeech(locale, () => resolve()));
|
|
476
480
|
}
|
|
@@ -508,6 +512,10 @@ class Speech {
|
|
|
508
512
|
|
|
509
513
|
async startWithSVOnboardingJson(locale: string, onboardingJsonPath: string): Promise<void> {
|
|
510
514
|
this.ensureListeners();
|
|
515
|
+
if (Platform.OS === 'ios' && (NativeSpeech as any)?.startSpeechWithSVOnboardingJsonAsync) {
|
|
516
|
+
await (NativeSpeech as any).startSpeechWithSVOnboardingJsonAsync(locale, onboardingJsonPath, 2500);
|
|
517
|
+
return;
|
|
518
|
+
}
|
|
511
519
|
if (Platform.OS === 'ios' && NativeSpeech?.startSpeechWithSVOnboardingJson) {
|
|
512
520
|
return new Promise<void>((resolve) =>
|
|
513
521
|
NativeSpeech.startSpeechWithSVOnboardingJson(locale, onboardingJsonPath, () => resolve()),
|
|
@@ -516,7 +524,8 @@ class Speech {
|
|
|
516
524
|
return this.start(locale);
|
|
517
525
|
}
|
|
518
526
|
|
|
519
|
-
|
|
527
|
+
// XXX BUG THE NATIVE SIDE DOES NOT REALLY AWAITS
|
|
528
|
+
async pauseSpeechRecognition(): Promise<void> {
|
|
520
529
|
this.logCall('pauseSpeechRecognitionLite');
|
|
521
530
|
|
|
522
531
|
const mod: any = Platform.OS === 'ios' ? NativeSpeech : NativeSTT;
|
|
@@ -524,10 +533,16 @@ class Speech {
|
|
|
524
533
|
|
|
525
534
|
if (!fn) {
|
|
526
535
|
dbg(`pauseSpeechRecognitionLite not available on ${Platform.OS === 'ios' ? 'NativeSpeech' : 'NativeSTT'}`);
|
|
527
|
-
return
|
|
536
|
+
return;
|
|
537
|
+
}
|
|
538
|
+
|
|
539
|
+
if (Platform.OS === 'ios' && typeof mod?.pauseSpeechRecognitionLiteAsync === 'function') {
|
|
540
|
+
const result = await mod.pauseSpeechRecognitionLiteAsync(1500);
|
|
541
|
+
if (result?.ok === false) dbgErr('pauseSpeechRecognitionLiteAsync failed', result?.reason);
|
|
542
|
+
return;
|
|
528
543
|
}
|
|
529
544
|
if (Platform.OS === 'ios') {
|
|
530
|
-
|
|
545
|
+
await new Promise<void>((resolve, reject) => {
|
|
531
546
|
try {
|
|
532
547
|
fn.call(mod, (ok: boolean) => {
|
|
533
548
|
if (!ok) dbgErr('pauseSpeechRecognitionLite returned false');
|
|
@@ -537,9 +552,10 @@ class Speech {
|
|
|
537
552
|
reject(e as any);
|
|
538
553
|
}
|
|
539
554
|
});
|
|
555
|
+
return;
|
|
540
556
|
}
|
|
541
557
|
|
|
542
|
-
|
|
558
|
+
await new Promise<void>((resolve, reject) => {
|
|
543
559
|
try {
|
|
544
560
|
fn.call(mod, async (ok: boolean) => {
|
|
545
561
|
if (!ok) dbgErr('pauseSpeechRecognitionLite returned false');
|
|
@@ -553,7 +569,7 @@ class Speech {
|
|
|
553
569
|
});
|
|
554
570
|
}
|
|
555
571
|
|
|
556
|
-
unPauseSpeechRecognition(times: number, preFetchMs: number = 0): Promise<void> {
|
|
572
|
+
async unPauseSpeechRecognition(times: number, preFetchMs: number = 0): Promise<void> {
|
|
557
573
|
this.logCall('unPauseSpeechRecognitionLite', { times, preFetchMs });
|
|
558
574
|
|
|
559
575
|
const mod: any = Platform.OS === 'ios' ? NativeSpeech : NativeSTT;
|
|
@@ -561,11 +577,17 @@ class Speech {
|
|
|
561
577
|
|
|
562
578
|
if (!fn) {
|
|
563
579
|
dbg(`unPauseSpeechRecognitionLite(times) not available on ${Platform.OS === 'ios' ? 'NativeSpeech' : 'NativeSTT'}`);
|
|
564
|
-
return
|
|
580
|
+
return;
|
|
581
|
+
}
|
|
582
|
+
|
|
583
|
+
if (Platform.OS === 'ios' && typeof mod?.unPauseSpeechRecognitionLiteAsync === 'function') {
|
|
584
|
+
const result = await mod.unPauseSpeechRecognitionLiteAsync(times, preFetchMs, 2500);
|
|
585
|
+
if (result?.ok === false) dbgErr('unPauseSpeechRecognitionLiteAsync failed', result?.reason);
|
|
586
|
+
return;
|
|
565
587
|
}
|
|
566
588
|
|
|
567
589
|
if (Platform.OS === 'ios') {
|
|
568
|
-
|
|
590
|
+
await new Promise<void>((resolve, reject) => {
|
|
569
591
|
try {
|
|
570
592
|
const done = (ok: boolean) => {
|
|
571
593
|
if (!ok) dbgErr('unPauseSpeechRecognitionLite(times) returned false');
|
|
@@ -585,8 +607,9 @@ class Speech {
|
|
|
585
607
|
reject(e as any);
|
|
586
608
|
}
|
|
587
609
|
});
|
|
610
|
+
return;
|
|
588
611
|
}
|
|
589
|
-
|
|
612
|
+
await new Promise<void>((resolve, reject) => {
|
|
590
613
|
try {
|
|
591
614
|
const done = async (ok: boolean) => {
|
|
592
615
|
if (!ok) dbgErr('unPauseSpeechRecognitionLite(times) returned false');
|