whisper.rn 0.4.0-rc.2 → 0.4.0-rc.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/android/src/main/java/com/rnwhisper/WhisperContext.java +26 -12
- package/ios/RNWhisperContext.h +1 -1
- package/ios/RNWhisperContext.mm +21 -16
- package/lib/typescript/index.d.ts +1 -1
- package/package.json +1 -1
- package/src/index.ts +1 -1
- package/ios/RNWhisper.xcodeproj/project.xcworkspace/contents.xcworkspacedata +0 -4
- package/ios/RNWhisper.xcodeproj/project.xcworkspace/xcshareddata/IDEWorkspaceChecks.plist +0 -8
- package/ios/RNWhisper.xcodeproj/project.xcworkspace/xcuserdata/jhen.xcuserdatad/UserInterfaceState.xcuserstate +0 -0
- package/ios/RNWhisper.xcodeproj/xcuserdata/jhen.xcuserdatad/xcschemes/xcschememanagement.plist +0 -19
|
@@ -82,8 +82,9 @@ public class WhisperContext {
|
|
|
82
82
|
private boolean vad(ReadableMap options, short[] shortBuffer, int nSamples, int n) {
|
|
83
83
|
boolean isSpeech = true;
|
|
84
84
|
if (!isTranscribing && options.hasKey("useVad") && options.getBoolean("useVad")) {
|
|
85
|
-
int
|
|
86
|
-
|
|
85
|
+
int vadMs = options.hasKey("vadMs") ? options.getInt("vadMs") : 2000;
|
|
86
|
+
if (vadMs < 2000) vadMs = 2000;
|
|
87
|
+
int sampleSize = (int) (SAMPLE_RATE * vadMs / 1000);
|
|
87
88
|
if (nSamples + n > sampleSize) {
|
|
88
89
|
int start = nSamples + n - sampleSize;
|
|
89
90
|
float[] audioData = new float[sampleSize];
|
|
@@ -100,6 +101,21 @@ public class WhisperContext {
|
|
|
100
101
|
return isSpeech;
|
|
101
102
|
}
|
|
102
103
|
|
|
104
|
+
private void finishRealtimeTranscribe(ReadableMap options, WritableMap result) {
|
|
105
|
+
String audioOutputPath = options.hasKey("audioOutputPath") ? options.getString("audioOutputPath") : null;
|
|
106
|
+
if (audioOutputPath != null) {
|
|
107
|
+
// TODO: Append in real time so we don't need to keep all slices & also reduce memory usage
|
|
108
|
+
Log.d(NAME, "Begin saving wav file to " + audioOutputPath);
|
|
109
|
+
try {
|
|
110
|
+
AudioUtils.saveWavFile(AudioUtils.concatShortBuffers(shortBufferSlices), audioOutputPath);
|
|
111
|
+
} catch (IOException e) {
|
|
112
|
+
Log.e(NAME, "Error saving wav file: " + e.getMessage());
|
|
113
|
+
}
|
|
114
|
+
}
|
|
115
|
+
|
|
116
|
+
emitTranscribeEvent("@RNWhisper_onRealtimeTranscribeEnd", Arguments.createMap());
|
|
117
|
+
}
|
|
118
|
+
|
|
103
119
|
public int startRealtimeTranscribe(int jobId, ReadableMap options) {
|
|
104
120
|
if (isCapturing || isTranscribing) {
|
|
105
121
|
return -100;
|
|
@@ -131,7 +147,7 @@ public class WhisperContext {
|
|
|
131
147
|
shortBufferSlices.add(new short[audioSliceSec * SAMPLE_RATE]);
|
|
132
148
|
sliceNSamples = new ArrayList<Integer>();
|
|
133
149
|
sliceNSamples.add(0);
|
|
134
|
-
|
|
150
|
+
|
|
135
151
|
isCapturing = true;
|
|
136
152
|
recorder.startRecording();
|
|
137
153
|
|
|
@@ -159,12 +175,12 @@ public class WhisperContext {
|
|
|
159
175
|
nSamples == nSamplesTranscribing &&
|
|
160
176
|
sliceIndex == transcribeSliceIndex
|
|
161
177
|
) {
|
|
162
|
-
|
|
178
|
+
finishRealtimeTranscribe(options, Arguments.createMap());
|
|
163
179
|
} else if (!isTranscribing) {
|
|
164
180
|
short[] shortBuffer = shortBufferSlices.get(sliceIndex);
|
|
165
181
|
boolean isSpeech = vad(options, shortBuffer, nSamples, 0);
|
|
166
182
|
if (!isSpeech) {
|
|
167
|
-
|
|
183
|
+
finishRealtimeTranscribe(options, Arguments.createMap());
|
|
168
184
|
break;
|
|
169
185
|
}
|
|
170
186
|
isTranscribing = true;
|
|
@@ -210,11 +226,9 @@ public class WhisperContext {
|
|
|
210
226
|
Log.e(NAME, "Error transcribing realtime: " + e.getMessage());
|
|
211
227
|
}
|
|
212
228
|
}
|
|
213
|
-
|
|
214
|
-
Log.d(NAME, "Begin saving wav file to " + audioOutputPath);
|
|
215
|
-
AudioUtils.saveWavFile(AudioUtils.concatShortBuffers(shortBufferSlices), audioOutputPath);
|
|
229
|
+
|
|
216
230
|
if (!isTranscribing) {
|
|
217
|
-
|
|
231
|
+
finishRealtimeTranscribe(options, Arguments.createMap());
|
|
218
232
|
}
|
|
219
233
|
if (fullHandler != null) {
|
|
220
234
|
fullHandler.join(); // Wait for full transcribe to finish
|
|
@@ -288,7 +302,7 @@ public class WhisperContext {
|
|
|
288
302
|
if (isStopped && !continueNeeded) {
|
|
289
303
|
payload.putBoolean("isCapturing", false);
|
|
290
304
|
payload.putBoolean("isStoppedByAction", isStoppedByAction);
|
|
291
|
-
|
|
305
|
+
finishRealtimeTranscribe(options, payload);
|
|
292
306
|
} else if (code == 0) {
|
|
293
307
|
payload.putBoolean("isCapturing", true);
|
|
294
308
|
emitTranscribeEvent("@RNWhisper_onRealtimeTranscribe", payload);
|
|
@@ -401,7 +415,7 @@ public class WhisperContext {
|
|
|
401
415
|
options.hasKey("maxLen") ? options.getInt("maxLen") : -1,
|
|
402
416
|
// jboolean token_timestamps,
|
|
403
417
|
options.hasKey("tokenTimestamps") ? options.getBoolean("tokenTimestamps") : false,
|
|
404
|
-
|
|
418
|
+
|
|
405
419
|
// jint offset,
|
|
406
420
|
options.hasKey("offset") ? options.getInt("offset") : -1,
|
|
407
421
|
// jint duration,
|
|
@@ -577,4 +591,4 @@ public class WhisperContext {
|
|
|
577
591
|
protected static native int getTextSegmentT0(long context, int index);
|
|
578
592
|
protected static native int getTextSegmentT1(long context, int index);
|
|
579
593
|
protected static native void freeContext(long contextPtr);
|
|
580
|
-
}
|
|
594
|
+
}
|
package/ios/RNWhisperContext.h
CHANGED
package/ios/RNWhisperContext.mm
CHANGED
|
@@ -56,7 +56,9 @@
|
|
|
56
56
|
self->recordState.audioOutputPath = options[@"audioOutputPath"];
|
|
57
57
|
|
|
58
58
|
self->recordState.useVad = options[@"useVad"] != nil ? [options[@"useVad"] boolValue] : false;
|
|
59
|
-
self->recordState.
|
|
59
|
+
self->recordState.vadMs = options[@"vadMs"] != nil ? [options[@"vadMs"] intValue] : 2000;
|
|
60
|
+
if (self->recordState.vadMs < 2000) self->recordState.vadMs = 2000;
|
|
61
|
+
|
|
60
62
|
self->recordState.vadThold = options[@"vadThold"] != nil ? [options[@"vadThold"] floatValue] : 0.6f;
|
|
61
63
|
self->recordState.vadFreqThold = options[@"vadFreqThold"] != nil ? [options[@"vadFreqThold"] floatValue] : 100.0f;
|
|
62
64
|
|
|
@@ -98,7 +100,7 @@ bool vad(RNWhisperContextRecordState *state, int16_t* audioBufferI16, int nSampl
|
|
|
98
100
|
{
|
|
99
101
|
bool isSpeech = true;
|
|
100
102
|
if (!state->isTranscribing && state->useVad) {
|
|
101
|
-
int sampleSize = state->
|
|
103
|
+
int sampleSize = (int) (WHISPER_SAMPLE_RATE * state->vadMs / 1000);
|
|
102
104
|
if (nSamples + n > sampleSize) {
|
|
103
105
|
int start = nSamples + n - sampleSize;
|
|
104
106
|
std::vector<float> audioBufferF32Vec(sampleSize);
|
|
@@ -126,7 +128,7 @@ void AudioInputCallback(void * inUserData,
|
|
|
126
128
|
if (!state->isCapturing) {
|
|
127
129
|
NSLog(@"[RNWhisper] Not capturing, ignoring audio");
|
|
128
130
|
if (!state->isTranscribing) {
|
|
129
|
-
state->
|
|
131
|
+
[state->mSelf finishRealtimeTranscribe:state result:@{}];
|
|
130
132
|
}
|
|
131
133
|
return;
|
|
132
134
|
}
|
|
@@ -149,14 +151,14 @@ void AudioInputCallback(void * inUserData,
|
|
|
149
151
|
nSamples == state->nSamplesTranscribing &&
|
|
150
152
|
state->sliceIndex == state->transcribeSliceIndex
|
|
151
153
|
) {
|
|
152
|
-
state->
|
|
154
|
+
[state->mSelf finishRealtimeTranscribe:state result:@{}];
|
|
153
155
|
} else if (
|
|
154
156
|
!state->isTranscribing &&
|
|
155
157
|
nSamples != state->nSamplesTranscribing
|
|
156
158
|
) {
|
|
157
159
|
int16_t* audioBufferI16 = (int16_t*) [state->shortBufferSlices[state->sliceIndex] pointerValue];
|
|
158
160
|
if (!vad(state, audioBufferI16, nSamples, 0)) {
|
|
159
|
-
state->
|
|
161
|
+
[state->mSelf finishRealtimeTranscribe:state result:@{}];
|
|
160
162
|
return;
|
|
161
163
|
}
|
|
162
164
|
state->isTranscribing = true;
|
|
@@ -201,6 +203,19 @@ void AudioInputCallback(void * inUserData,
|
|
|
201
203
|
}
|
|
202
204
|
}
|
|
203
205
|
|
|
206
|
+
- (void)finishRealtimeTranscribe:(RNWhisperContextRecordState*) state result:(NSDictionary*)result {
|
|
207
|
+
// Save wav if needed
|
|
208
|
+
if (state->audioOutputPath != nil) {
|
|
209
|
+
// TODO: Append in real time so we don't need to keep all slices & also reduce memory usage
|
|
210
|
+
[RNWhisperAudioUtils
|
|
211
|
+
saveWavFile:[RNWhisperAudioUtils concatShortBuffers:state->shortBufferSlices
|
|
212
|
+
sliceNSamples:state->sliceNSamples]
|
|
213
|
+
audioOutputFile:state->audioOutputPath
|
|
214
|
+
];
|
|
215
|
+
}
|
|
216
|
+
state->transcribeHandler(state->jobId, @"end", result);
|
|
217
|
+
}
|
|
218
|
+
|
|
204
219
|
- (void)fullTranscribeSamples:(RNWhisperContextRecordState*) state {
|
|
205
220
|
int nSamplesOfIndex = [[state->sliceNSamples objectAtIndex:state->transcribeSliceIndex] intValue];
|
|
206
221
|
state->nSamplesTranscribing = nSamplesOfIndex;
|
|
@@ -260,17 +275,7 @@ void AudioInputCallback(void * inUserData,
|
|
|
260
275
|
result[@"isStoppedByAction"] = @(state->isStoppedByAction);
|
|
261
276
|
result[@"isCapturing"] = @(false);
|
|
262
277
|
|
|
263
|
-
|
|
264
|
-
if (state->audioOutputPath != nil) {
|
|
265
|
-
// TODO: Append in real time so we don't need to keep all slices & also reduce memory usage
|
|
266
|
-
[RNWhisperAudioUtils
|
|
267
|
-
saveWavFile:[RNWhisperAudioUtils concatShortBuffers:state->shortBufferSlices
|
|
268
|
-
sliceNSamples:state->sliceNSamples]
|
|
269
|
-
audioOutputFile:state->audioOutputPath
|
|
270
|
-
];
|
|
271
|
-
}
|
|
272
|
-
|
|
273
|
-
state->transcribeHandler(state->jobId, @"end", result);
|
|
278
|
+
[state->mSelf finishRealtimeTranscribe:state result:result];
|
|
274
279
|
} else if (code == 0) {
|
|
275
280
|
result[@"isCapturing"] = @(true);
|
|
276
281
|
state->transcribeHandler(state->jobId, @"transcribe", result);
|
|
@@ -59,7 +59,7 @@ export type TranscribeRealtimeOptions = TranscribeOptions & {
|
|
|
59
59
|
*/
|
|
60
60
|
useVad?: boolean;
|
|
61
61
|
/**
|
|
62
|
-
* The length of the collected audio is used for VAD. (ms) (Default: 2000)
|
|
62
|
+
* The length of the collected audio is used for VAD, cannot be less than 2000ms. (ms) (Default: 2000)
|
|
63
63
|
*/
|
|
64
64
|
vadMs?: number;
|
|
65
65
|
/**
|
package/package.json
CHANGED
package/src/index.ts
CHANGED
|
@@ -106,7 +106,7 @@ export type TranscribeRealtimeOptions = TranscribeOptions & {
|
|
|
106
106
|
*/
|
|
107
107
|
useVad?: boolean
|
|
108
108
|
/**
|
|
109
|
-
* The length of the collected audio is used for VAD. (ms) (Default: 2000)
|
|
109
|
+
* The length of the collected audio is used for VAD, cannot be less than 2000ms. (ms) (Default: 2000)
|
|
110
110
|
*/
|
|
111
111
|
vadMs?: number
|
|
112
112
|
/**
|
|
Binary file
|
package/ios/RNWhisper.xcodeproj/xcuserdata/jhen.xcuserdatad/xcschemes/xcschememanagement.plist
DELETED
|
@@ -1,19 +0,0 @@
|
|
|
1
|
-
<?xml version="1.0" encoding="UTF-8"?>
|
|
2
|
-
<!DOCTYPE plist PUBLIC "-//Apple//DTD PLIST 1.0//EN" "http://www.apple.com/DTDs/PropertyList-1.0.dtd">
|
|
3
|
-
<plist version="1.0">
|
|
4
|
-
<dict>
|
|
5
|
-
<key>SchemeUserState</key>
|
|
6
|
-
<dict>
|
|
7
|
-
<key>RNWhisper.xcscheme_^#shared#^_</key>
|
|
8
|
-
<dict>
|
|
9
|
-
<key>orderHint</key>
|
|
10
|
-
<integer>0</integer>
|
|
11
|
-
</dict>
|
|
12
|
-
<key>WhisperCpp.xcscheme_^#shared#^_</key>
|
|
13
|
-
<dict>
|
|
14
|
-
<key>orderHint</key>
|
|
15
|
-
<integer>0</integer>
|
|
16
|
-
</dict>
|
|
17
|
-
</dict>
|
|
18
|
-
</dict>
|
|
19
|
-
</plist>
|