@speakableio/core 0.1.88 → 0.1.90
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.native.d.mts +8 -2
- package/dist/index.native.d.ts +8 -2
- package/dist/index.native.js +8 -4
- package/dist/index.native.js.map +1 -1
- package/dist/index.native.mjs +8 -4
- package/dist/index.native.mjs.map +1 -1
- package/dist/index.web.d.mts +8 -2
- package/dist/index.web.js +8 -4
- package/dist/index.web.js.map +1 -1
- package/package.json +1 -1
package/dist/index.native.d.mts
CHANGED
|
@@ -2368,7 +2368,10 @@ declare const useBaseOpenAI: ({ onTranscriptSuccess, onTranscriptError, onComple
|
|
|
2368
2368
|
url: string;
|
|
2369
2369
|
fileName: string;
|
|
2370
2370
|
}>;
|
|
2371
|
-
uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<
|
|
2371
|
+
uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<{
|
|
2372
|
+
transcript: string;
|
|
2373
|
+
audioUrl: string;
|
|
2374
|
+
}>;
|
|
2372
2375
|
onGetAudioUrlAndTranscript?: (args: {
|
|
2373
2376
|
transcript?: string;
|
|
2374
2377
|
audioUrl?: string;
|
|
@@ -2378,7 +2381,10 @@ declare const useBaseOpenAI: ({ onTranscriptSuccess, onTranscriptError, onComple
|
|
|
2378
2381
|
url: string;
|
|
2379
2382
|
fileName: string;
|
|
2380
2383
|
}>;
|
|
2381
|
-
uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<
|
|
2384
|
+
uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<{
|
|
2385
|
+
transcript: string;
|
|
2386
|
+
audioUrl: string;
|
|
2387
|
+
}>;
|
|
2382
2388
|
getTranscript: (audioUrl: string, language: string) => Promise<string>;
|
|
2383
2389
|
getFreeResponseCompletion: (messages: string[], isFreeResponse: boolean, feedbackLanguage: string, gradingStandard?: string) => Promise<any>;
|
|
2384
2390
|
getFeedback: ({ cardId, language, writtenResponse, audio, autoGrade, file, }: {
|
package/dist/index.native.d.ts
CHANGED
|
@@ -2368,7 +2368,10 @@ declare const useBaseOpenAI: ({ onTranscriptSuccess, onTranscriptError, onComple
|
|
|
2368
2368
|
url: string;
|
|
2369
2369
|
fileName: string;
|
|
2370
2370
|
}>;
|
|
2371
|
-
uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<
|
|
2371
|
+
uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<{
|
|
2372
|
+
transcript: string;
|
|
2373
|
+
audioUrl: string;
|
|
2374
|
+
}>;
|
|
2372
2375
|
onGetAudioUrlAndTranscript?: (args: {
|
|
2373
2376
|
transcript?: string;
|
|
2374
2377
|
audioUrl?: string;
|
|
@@ -2378,7 +2381,10 @@ declare const useBaseOpenAI: ({ onTranscriptSuccess, onTranscriptError, onComple
|
|
|
2378
2381
|
url: string;
|
|
2379
2382
|
fileName: string;
|
|
2380
2383
|
}>;
|
|
2381
|
-
uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<
|
|
2384
|
+
uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<{
|
|
2385
|
+
transcript: string;
|
|
2386
|
+
audioUrl: string;
|
|
2387
|
+
}>;
|
|
2382
2388
|
getTranscript: (audioUrl: string, language: string) => Promise<string>;
|
|
2383
2389
|
getFreeResponseCompletion: (messages: string[], isFreeResponse: boolean, feedbackLanguage: string, gradingStandard?: string) => Promise<any>;
|
|
2384
2390
|
getFeedback: ({ cardId, language, writtenResponse, audio, autoGrade, file, }: {
|
package/dist/index.native.js
CHANGED
|
@@ -2908,7 +2908,7 @@ var useUpdateStudentVocab = (page) => {
|
|
|
2908
2908
|
} catch (error) {
|
|
2909
2909
|
if (error instanceof Error && error.message === "not-found") {
|
|
2910
2910
|
data.firstSeen = serverTimestamp2();
|
|
2911
|
-
await api.setDoc(docPath, data);
|
|
2911
|
+
await api.setDoc(docPath, data, { merge: true });
|
|
2912
2912
|
} else {
|
|
2913
2913
|
console.log(error);
|
|
2914
2914
|
}
|
|
@@ -2930,7 +2930,7 @@ var useUpdateStudentVocab = (page) => {
|
|
|
2930
2930
|
} catch (error) {
|
|
2931
2931
|
if (error instanceof Error && error.message === "not-found") {
|
|
2932
2932
|
data.firstSeen = serverTimestamp2();
|
|
2933
|
-
await api.setDoc(docPath, data);
|
|
2933
|
+
await api.setDoc(docPath, data, { merge: true });
|
|
2934
2934
|
} else {
|
|
2935
2935
|
console.log(error);
|
|
2936
2936
|
}
|
|
@@ -3152,22 +3152,26 @@ var useBaseOpenAI = ({
|
|
|
3152
3152
|
return result;
|
|
3153
3153
|
}
|
|
3154
3154
|
let transcript;
|
|
3155
|
+
let audioUrl = void 0;
|
|
3155
3156
|
if (writtenResponse) {
|
|
3156
3157
|
transcript = writtenResponse;
|
|
3157
3158
|
onTranscriptSuccess(writtenResponse);
|
|
3158
3159
|
} else if (typeof audio === "string" && file) {
|
|
3159
3160
|
if (feedbackAccess == null ? void 0 : feedbackAccess.canAccessFeedback) {
|
|
3160
3161
|
transcript = await getTranscript(audio, language);
|
|
3162
|
+
audioUrl = audio;
|
|
3161
3163
|
onTranscriptSuccess(transcript);
|
|
3162
|
-
onGetAudioUrlAndTranscript == null ? void 0 : onGetAudioUrlAndTranscript({ transcript, audioUrl: audio });
|
|
3163
3164
|
} else {
|
|
3164
3165
|
console.info(
|
|
3165
3166
|
`Transcript not available: ${(feedbackAccess == null ? void 0 : feedbackAccess.reason) || "No feedback access"}`
|
|
3166
3167
|
);
|
|
3167
3168
|
}
|
|
3168
3169
|
} else {
|
|
3169
|
-
|
|
3170
|
+
const response = await uploadAudioAndGetTranscript(audio || "", language);
|
|
3171
|
+
transcript = response.transcript;
|
|
3172
|
+
audioUrl = response.audioUrl;
|
|
3170
3173
|
}
|
|
3174
|
+
onGetAudioUrlAndTranscript == null ? void 0 : onGetAudioUrlAndTranscript({ transcript, audioUrl });
|
|
3171
3175
|
if (feedbackAccess == null ? void 0 : feedbackAccess.canAccessFeedback) {
|
|
3172
3176
|
const results = await getAIResponse({
|
|
3173
3177
|
cardId,
|