@speakableio/core 0.1.88 → 0.1.90

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2368,7 +2368,10 @@ declare const useBaseOpenAI: ({ onTranscriptSuccess, onTranscriptError, onComple
2368
2368
  url: string;
2369
2369
  fileName: string;
2370
2370
  }>;
2371
- uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<string>;
2371
+ uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<{
2372
+ transcript: string;
2373
+ audioUrl: string;
2374
+ }>;
2372
2375
  onGetAudioUrlAndTranscript?: (args: {
2373
2376
  transcript?: string;
2374
2377
  audioUrl?: string;
@@ -2378,7 +2381,10 @@ declare const useBaseOpenAI: ({ onTranscriptSuccess, onTranscriptError, onComple
2378
2381
  url: string;
2379
2382
  fileName: string;
2380
2383
  }>;
2381
- uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<string>;
2384
+ uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<{
2385
+ transcript: string;
2386
+ audioUrl: string;
2387
+ }>;
2382
2388
  getTranscript: (audioUrl: string, language: string) => Promise<string>;
2383
2389
  getFreeResponseCompletion: (messages: string[], isFreeResponse: boolean, feedbackLanguage: string, gradingStandard?: string) => Promise<any>;
2384
2390
  getFeedback: ({ cardId, language, writtenResponse, audio, autoGrade, file, }: {
@@ -2368,7 +2368,10 @@ declare const useBaseOpenAI: ({ onTranscriptSuccess, onTranscriptError, onComple
2368
2368
  url: string;
2369
2369
  fileName: string;
2370
2370
  }>;
2371
- uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<string>;
2371
+ uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<{
2372
+ transcript: string;
2373
+ audioUrl: string;
2374
+ }>;
2372
2375
  onGetAudioUrlAndTranscript?: (args: {
2373
2376
  transcript?: string;
2374
2377
  audioUrl?: string;
@@ -2378,7 +2381,10 @@ declare const useBaseOpenAI: ({ onTranscriptSuccess, onTranscriptError, onComple
2378
2381
  url: string;
2379
2382
  fileName: string;
2380
2383
  }>;
2381
- uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<string>;
2384
+ uploadAudioAndGetTranscript: (audio: string, language: string) => Promise<{
2385
+ transcript: string;
2386
+ audioUrl: string;
2387
+ }>;
2382
2388
  getTranscript: (audioUrl: string, language: string) => Promise<string>;
2383
2389
  getFreeResponseCompletion: (messages: string[], isFreeResponse: boolean, feedbackLanguage: string, gradingStandard?: string) => Promise<any>;
2384
2390
  getFeedback: ({ cardId, language, writtenResponse, audio, autoGrade, file, }: {
@@ -2908,7 +2908,7 @@ var useUpdateStudentVocab = (page) => {
2908
2908
  } catch (error) {
2909
2909
  if (error instanceof Error && error.message === "not-found") {
2910
2910
  data.firstSeen = serverTimestamp2();
2911
- await api.setDoc(docPath, data);
2911
+ await api.setDoc(docPath, data, { merge: true });
2912
2912
  } else {
2913
2913
  console.log(error);
2914
2914
  }
@@ -2930,7 +2930,7 @@ var useUpdateStudentVocab = (page) => {
2930
2930
  } catch (error) {
2931
2931
  if (error instanceof Error && error.message === "not-found") {
2932
2932
  data.firstSeen = serverTimestamp2();
2933
- await api.setDoc(docPath, data);
2933
+ await api.setDoc(docPath, data, { merge: true });
2934
2934
  } else {
2935
2935
  console.log(error);
2936
2936
  }
@@ -3152,22 +3152,26 @@ var useBaseOpenAI = ({
3152
3152
  return result;
3153
3153
  }
3154
3154
  let transcript;
3155
+ let audioUrl = void 0;
3155
3156
  if (writtenResponse) {
3156
3157
  transcript = writtenResponse;
3157
3158
  onTranscriptSuccess(writtenResponse);
3158
3159
  } else if (typeof audio === "string" && file) {
3159
3160
  if (feedbackAccess == null ? void 0 : feedbackAccess.canAccessFeedback) {
3160
3161
  transcript = await getTranscript(audio, language);
3162
+ audioUrl = audio;
3161
3163
  onTranscriptSuccess(transcript);
3162
- onGetAudioUrlAndTranscript == null ? void 0 : onGetAudioUrlAndTranscript({ transcript, audioUrl: audio });
3163
3164
  } else {
3164
3165
  console.info(
3165
3166
  `Transcript not available: ${(feedbackAccess == null ? void 0 : feedbackAccess.reason) || "No feedback access"}`
3166
3167
  );
3167
3168
  }
3168
3169
  } else {
3169
- transcript = await uploadAudioAndGetTranscript(audio || "", language);
3170
+ const response = await uploadAudioAndGetTranscript(audio || "", language);
3171
+ transcript = response.transcript;
3172
+ audioUrl = response.audioUrl;
3170
3173
  }
3174
+ onGetAudioUrlAndTranscript == null ? void 0 : onGetAudioUrlAndTranscript({ transcript, audioUrl });
3171
3175
  if (feedbackAccess == null ? void 0 : feedbackAccess.canAccessFeedback) {
3172
3176
  const results = await getAIResponse({
3173
3177
  cardId,