@ai-sdk/revai 3.0.0-beta.2 → 3.0.0-beta.21

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/CHANGELOG.md CHANGED
@@ -1,5 +1,155 @@
1
1
  # @ai-sdk/revai
2
2
 
3
+ ## 3.0.0-beta.21
4
+
5
+ ### Major Changes
6
+
7
+ - ef992f8: Remove CommonJS exports from all packages. All packages are now ESM-only (`"type": "module"`). Consumers using `require()` must switch to ESM `import` syntax.
8
+
9
+ ### Patch Changes
10
+
11
+ - Updated dependencies [ef992f8]
12
+ - @ai-sdk/provider@4.0.0-beta.11
13
+ - @ai-sdk/provider-utils@5.0.0-beta.19
14
+
15
+ ## 3.0.0-beta.20
16
+
17
+ ### Patch Changes
18
+
19
+ - Updated dependencies [90e2d8a]
20
+ - @ai-sdk/provider-utils@5.0.0-beta.18
21
+
22
+ ## 3.0.0-beta.19
23
+
24
+ ### Patch Changes
25
+
26
+ - Updated dependencies [3ae1786]
27
+ - @ai-sdk/provider-utils@5.0.0-beta.17
28
+
29
+ ## 3.0.0-beta.18
30
+
31
+ ### Patch Changes
32
+
33
+ - Updated dependencies [176466a]
34
+ - @ai-sdk/provider@4.0.0-beta.10
35
+ - @ai-sdk/provider-utils@5.0.0-beta.16
36
+
37
+ ## 3.0.0-beta.17
38
+
39
+ ### Patch Changes
40
+
41
+ - Updated dependencies [e311194]
42
+ - @ai-sdk/provider@4.0.0-beta.9
43
+ - @ai-sdk/provider-utils@5.0.0-beta.15
44
+
45
+ ## 3.0.0-beta.16
46
+
47
+ ### Patch Changes
48
+
49
+ - Updated dependencies [34bd95d]
50
+ - Updated dependencies [008271d]
51
+ - @ai-sdk/provider@4.0.0-beta.8
52
+ - @ai-sdk/provider-utils@5.0.0-beta.14
53
+
54
+ ## 3.0.0-beta.15
55
+
56
+ ### Patch Changes
57
+
58
+ - Updated dependencies [b0c2869]
59
+ - Updated dependencies [7e26e81]
60
+ - @ai-sdk/provider-utils@5.0.0-beta.13
61
+
62
+ ## 3.0.0-beta.14
63
+
64
+ ### Patch Changes
65
+
66
+ - Updated dependencies [46d1149]
67
+ - @ai-sdk/provider-utils@5.0.0-beta.12
68
+
69
+ ## 3.0.0-beta.13
70
+
71
+ ### Patch Changes
72
+
73
+ - Updated dependencies [6fd51c0]
74
+ - @ai-sdk/provider-utils@5.0.0-beta.11
75
+ - @ai-sdk/provider@4.0.0-beta.7
76
+
77
+ ## 3.0.0-beta.12
78
+
79
+ ### Patch Changes
80
+
81
+ - Updated dependencies [c29a26f]
82
+ - @ai-sdk/provider-utils@5.0.0-beta.10
83
+ - @ai-sdk/provider@4.0.0-beta.6
84
+
85
+ ## 3.0.0-beta.11
86
+
87
+ ### Patch Changes
88
+
89
+ - 38fc777: Add AI Gateway hint to provider READMEs
90
+
91
+ ## 3.0.0-beta.10
92
+
93
+ ### Patch Changes
94
+
95
+ - Updated dependencies [2e17091]
96
+ - @ai-sdk/provider-utils@5.0.0-beta.9
97
+
98
+ ## 3.0.0-beta.9
99
+
100
+ ### Patch Changes
101
+
102
+ - Updated dependencies [986c6fd]
103
+ - Updated dependencies [493295c]
104
+ - @ai-sdk/provider-utils@5.0.0-beta.8
105
+
106
+ ## 3.0.0-beta.8
107
+
108
+ ### Patch Changes
109
+
110
+ - Updated dependencies [1f509d4]
111
+ - @ai-sdk/provider-utils@5.0.0-beta.7
112
+ - @ai-sdk/provider@4.0.0-beta.5
113
+
114
+ ## 3.0.0-beta.7
115
+
116
+ ### Patch Changes
117
+
118
+ - Updated dependencies [3887c70]
119
+ - @ai-sdk/provider-utils@5.0.0-beta.6
120
+ - @ai-sdk/provider@4.0.0-beta.4
121
+
122
+ ## 3.0.0-beta.6
123
+
124
+ ### Patch Changes
125
+
126
+ - Updated dependencies [776b617]
127
+ - @ai-sdk/provider-utils@5.0.0-beta.5
128
+ - @ai-sdk/provider@4.0.0-beta.3
129
+
130
+ ## 3.0.0-beta.5
131
+
132
+ ### Patch Changes
133
+
134
+ - Updated dependencies [61753c3]
135
+ - @ai-sdk/provider-utils@5.0.0-beta.4
136
+
137
+ ## 3.0.0-beta.4
138
+
139
+ ### Patch Changes
140
+
141
+ - Updated dependencies [f7d4f01]
142
+ - @ai-sdk/provider-utils@5.0.0-beta.3
143
+ - @ai-sdk/provider@4.0.0-beta.2
144
+
145
+ ## 3.0.0-beta.3
146
+
147
+ ### Patch Changes
148
+
149
+ - Updated dependencies [5c2a5a2]
150
+ - @ai-sdk/provider@4.0.0-beta.1
151
+ - @ai-sdk/provider-utils@5.0.0-beta.2
152
+
3
153
  ## 3.0.0-beta.2
4
154
 
5
155
  ### Patch Changes
@@ -216,13 +366,13 @@
216
366
  Before
217
367
 
218
368
  ```ts
219
- model.textEmbeddingModel('my-model-id');
369
+ model.textEmbeddingModel("my-model-id");
220
370
  ```
221
371
 
222
372
  After
223
373
 
224
374
  ```ts
225
- model.embeddingModel('my-model-id');
375
+ model.embeddingModel("my-model-id");
226
376
  ```
227
377
 
228
378
  - 95f65c2: chore: use import \* from zod/v4
@@ -422,13 +572,13 @@
422
572
  Before
423
573
 
424
574
  ```ts
425
- model.textEmbeddingModel('my-model-id');
575
+ model.textEmbeddingModel("my-model-id");
426
576
  ```
427
577
 
428
578
  After
429
579
 
430
580
  ```ts
431
- model.embeddingModel('my-model-id');
581
+ model.embeddingModel("my-model-id");
432
582
  ```
433
583
 
434
584
  - Updated dependencies [8d9e8ad]
package/README.md CHANGED
@@ -3,6 +3,8 @@
3
3
  The **[Rev.ai provider](https://ai-sdk.dev/providers/ai-sdk-providers/revai)** for the [AI SDK](https://ai-sdk.dev/docs)
4
4
  contains language model support for the Rev.ai transcription API.
5
5
 
6
+ > **Deploying to Vercel?** With Vercel's AI Gateway you can access Rev.ai (and hundreds of models from other providers) — no additional packages, API keys, or extra cost. [Get started with AI Gateway](https://vercel.com/ai-gateway).
7
+
6
8
  ## Setup
7
9
 
8
10
  The Rev.ai provider is available in the `@ai-sdk/revai` module. You can install it with
package/dist/index.js CHANGED
@@ -1,195 +1,183 @@
1
- "use strict";
2
- var __defProp = Object.defineProperty;
3
- var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
4
- var __getOwnPropNames = Object.getOwnPropertyNames;
5
- var __hasOwnProp = Object.prototype.hasOwnProperty;
6
- var __export = (target, all) => {
7
- for (var name in all)
8
- __defProp(target, name, { get: all[name], enumerable: true });
9
- };
10
- var __copyProps = (to, from, except, desc) => {
11
- if (from && typeof from === "object" || typeof from === "function") {
12
- for (let key of __getOwnPropNames(from))
13
- if (!__hasOwnProp.call(to, key) && key !== except)
14
- __defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
15
- }
16
- return to;
17
- };
18
- var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
19
-
20
- // src/index.ts
21
- var index_exports = {};
22
- __export(index_exports, {
23
- VERSION: () => VERSION,
24
- createRevai: () => createRevai,
25
- revai: () => revai
26
- });
27
- module.exports = __toCommonJS(index_exports);
28
-
29
1
  // src/revai-provider.ts
30
- var import_provider2 = require("@ai-sdk/provider");
31
- var import_provider_utils3 = require("@ai-sdk/provider-utils");
2
+ import {
3
+ NoSuchModelError
4
+ } from "@ai-sdk/provider";
5
+ import {
6
+ loadApiKey,
7
+ withUserAgentSuffix
8
+ } from "@ai-sdk/provider-utils";
32
9
 
33
10
  // src/revai-transcription-model.ts
34
- var import_provider = require("@ai-sdk/provider");
35
- var import_provider_utils2 = require("@ai-sdk/provider-utils");
36
- var import_v42 = require("zod/v4");
11
+ import {
12
+ AISDKError
13
+ } from "@ai-sdk/provider";
14
+ import {
15
+ combineHeaders,
16
+ convertBase64ToUint8Array,
17
+ createJsonResponseHandler,
18
+ mediaTypeToExtension,
19
+ delay,
20
+ getFromApi,
21
+ parseProviderOptions,
22
+ postFormDataToApi
23
+ } from "@ai-sdk/provider-utils";
24
+ import { z as z2 } from "zod/v4";
37
25
 
38
26
  // src/revai-error.ts
39
- var import_v4 = require("zod/v4");
40
- var import_provider_utils = require("@ai-sdk/provider-utils");
41
- var revaiErrorDataSchema = import_v4.z.object({
42
- error: import_v4.z.object({
43
- message: import_v4.z.string(),
44
- code: import_v4.z.number()
27
+ import { z } from "zod/v4";
28
+ import { createJsonErrorResponseHandler } from "@ai-sdk/provider-utils";
29
+ var revaiErrorDataSchema = z.object({
30
+ error: z.object({
31
+ message: z.string(),
32
+ code: z.number()
45
33
  })
46
34
  });
47
- var revaiFailedResponseHandler = (0, import_provider_utils.createJsonErrorResponseHandler)({
35
+ var revaiFailedResponseHandler = createJsonErrorResponseHandler({
48
36
  errorSchema: revaiErrorDataSchema,
49
37
  errorToMessage: (data) => data.error.message
50
38
  });
51
39
 
52
40
  // src/revai-transcription-model.ts
53
- var revaiTranscriptionModelOptionsSchema = import_v42.z.object({
41
+ var revaiTranscriptionModelOptionsSchema = z2.object({
54
42
  /**
55
43
  * Optional metadata string to associate with the transcription job.
56
44
  */
57
- metadata: import_v42.z.string().nullish(),
45
+ metadata: z2.string().nullish(),
58
46
  /**
59
47
  * Configuration for webhook notifications when job is complete.
60
48
  */
61
- notification_config: import_v42.z.object({
49
+ notification_config: z2.object({
62
50
  /**
63
51
  * URL to send the notification to.
64
52
  */
65
- url: import_v42.z.string(),
53
+ url: z2.string(),
66
54
  /**
67
55
  * Optional authorization headers for the notification request.
68
56
  */
69
- auth_headers: import_v42.z.object({
70
- Authorization: import_v42.z.string()
57
+ auth_headers: z2.object({
58
+ Authorization: z2.string()
71
59
  }).nullish()
72
60
  }).nullish(),
73
61
  /**
74
62
  * Number of seconds after which the job will be automatically deleted.
75
63
  */
76
- delete_after_seconds: import_v42.z.number().nullish(),
64
+ delete_after_seconds: z2.number().nullish(),
77
65
  /**
78
66
  * Whether to include filler words and false starts in the transcription.
79
67
  */
80
- verbatim: import_v42.z.boolean().optional(),
68
+ verbatim: z2.boolean().optional(),
81
69
  /**
82
70
  * Whether to prioritize the job for faster processing.
83
71
  */
84
- rush: import_v42.z.boolean().nullish().default(false),
72
+ rush: z2.boolean().nullish().default(false),
85
73
  /**
86
74
  * Whether to run the job in test mode.
87
75
  */
88
- test_mode: import_v42.z.boolean().nullish().default(false),
76
+ test_mode: z2.boolean().nullish().default(false),
89
77
  /**
90
78
  * Specific segments of the audio to transcribe.
91
79
  */
92
- segments_to_transcribe: import_v42.z.array(
93
- import_v42.z.object({
80
+ segments_to_transcribe: z2.array(
81
+ z2.object({
94
82
  /**
95
83
  * Start time of the segment in seconds.
96
84
  */
97
- start: import_v42.z.number(),
85
+ start: z2.number(),
98
86
  /**
99
87
  * End time of the segment in seconds.
100
88
  */
101
- end: import_v42.z.number()
89
+ end: z2.number()
102
90
  })
103
91
  ).nullish(),
104
92
  /**
105
93
  * Names to assign to speakers in the transcription.
106
94
  */
107
- speaker_names: import_v42.z.array(
108
- import_v42.z.object({
95
+ speaker_names: z2.array(
96
+ z2.object({
109
97
  /**
110
98
  * Display name for the speaker.
111
99
  */
112
- display_name: import_v42.z.string()
100
+ display_name: z2.string()
113
101
  })
114
102
  ).nullish(),
115
103
  /**
116
104
  * Whether to skip speaker diarization.
117
105
  */
118
- skip_diarization: import_v42.z.boolean().nullish().default(false),
106
+ skip_diarization: z2.boolean().nullish().default(false),
119
107
  /**
120
108
  * Whether to skip post-processing steps.
121
109
  */
122
- skip_postprocessing: import_v42.z.boolean().nullish().default(false),
110
+ skip_postprocessing: z2.boolean().nullish().default(false),
123
111
  /**
124
112
  * Whether to skip adding punctuation to the transcription.
125
113
  */
126
- skip_punctuation: import_v42.z.boolean().nullish().default(false),
114
+ skip_punctuation: z2.boolean().nullish().default(false),
127
115
  /**
128
116
  * Whether to remove disfluencies (um, uh, etc.) from the transcription.
129
117
  */
130
- remove_disfluencies: import_v42.z.boolean().nullish().default(false),
118
+ remove_disfluencies: z2.boolean().nullish().default(false),
131
119
  /**
132
120
  * Whether to remove atmospheric sounds from the transcription.
133
121
  */
134
- remove_atmospherics: import_v42.z.boolean().nullish().default(false),
122
+ remove_atmospherics: z2.boolean().nullish().default(false),
135
123
  /**
136
124
  * Whether to filter profanity from the transcription.
137
125
  */
138
- filter_profanity: import_v42.z.boolean().nullish().default(false),
126
+ filter_profanity: z2.boolean().nullish().default(false),
139
127
  /**
140
128
  * Number of speaker channels in the audio.
141
129
  */
142
- speaker_channels_count: import_v42.z.number().nullish(),
130
+ speaker_channels_count: z2.number().nullish(),
143
131
  /**
144
132
  * Expected number of speakers in the audio.
145
133
  */
146
- speakers_count: import_v42.z.number().nullish(),
134
+ speakers_count: z2.number().nullish(),
147
135
  /**
148
136
  * Type of diarization to use.
149
137
  */
150
- diarization_type: import_v42.z.enum(["standard", "premium"]).nullish().default("standard"),
138
+ diarization_type: z2.enum(["standard", "premium"]).nullish().default("standard"),
151
139
  /**
152
140
  * ID of a custom vocabulary to use for the transcription.
153
141
  */
154
- custom_vocabulary_id: import_v42.z.string().nullish(),
142
+ custom_vocabulary_id: z2.string().nullish(),
155
143
  /**
156
144
  * Custom vocabularies to use for the transcription.
157
145
  */
158
- custom_vocabularies: import_v42.z.array(import_v42.z.object({})).optional(),
146
+ custom_vocabularies: z2.array(z2.object({})).optional(),
159
147
  /**
160
148
  * Whether to strictly enforce custom vocabulary.
161
149
  */
162
- strict_custom_vocabulary: import_v42.z.boolean().optional(),
150
+ strict_custom_vocabulary: z2.boolean().optional(),
163
151
  /**
164
152
  * Configuration for generating a summary of the transcription.
165
153
  */
166
- summarization_config: import_v42.z.object({
154
+ summarization_config: z2.object({
167
155
  /**
168
156
  * Model to use for summarization.
169
157
  */
170
- model: import_v42.z.enum(["standard", "premium"]).nullish().default("standard"),
158
+ model: z2.enum(["standard", "premium"]).nullish().default("standard"),
171
159
  /**
172
160
  * Format of the summary.
173
161
  */
174
- type: import_v42.z.enum(["paragraph", "bullets"]).nullish().default("paragraph"),
162
+ type: z2.enum(["paragraph", "bullets"]).nullish().default("paragraph"),
175
163
  /**
176
164
  * Custom prompt for the summarization.
177
165
  */
178
- prompt: import_v42.z.string().nullish()
166
+ prompt: z2.string().nullish()
179
167
  }).nullish(),
180
168
  /**
181
169
  * Configuration for translating the transcription.
182
170
  */
183
- translation_config: import_v42.z.object({
171
+ translation_config: z2.object({
184
172
  /**
185
173
  * Target languages for translation.
186
174
  */
187
- target_languages: import_v42.z.array(
188
- import_v42.z.object({
175
+ target_languages: z2.array(
176
+ z2.object({
189
177
  /**
190
178
  * Language code for translation target.
191
179
  */
192
- language: import_v42.z.enum([
180
+ language: z2.enum([
193
181
  "en",
194
182
  "en-us",
195
183
  "en-gb",
@@ -213,16 +201,16 @@ var revaiTranscriptionModelOptionsSchema = import_v42.z.object({
213
201
  /**
214
202
  * Model to use for translation.
215
203
  */
216
- model: import_v42.z.enum(["standard", "premium"]).nullish().default("standard")
204
+ model: z2.enum(["standard", "premium"]).nullish().default("standard")
217
205
  }).nullish(),
218
206
  /**
219
207
  * Language of the audio content.
220
208
  */
221
- language: import_v42.z.string().nullish().default("en"),
209
+ language: z2.string().nullish().default("en"),
222
210
  /**
223
211
  * Whether to perform forced alignment.
224
212
  */
225
- forced_alignment: import_v42.z.boolean().nullish().default(false)
213
+ forced_alignment: z2.boolean().nullish().default(false)
226
214
  });
227
215
  var RevaiTranscriptionModel = class {
228
216
  constructor(modelId, config) {
@@ -240,14 +228,14 @@ var RevaiTranscriptionModel = class {
240
228
  }) {
241
229
  var _a, _b, _c, _d, _e, _f, _g, _h, _i, _j, _k, _l, _m, _n, _o, _p, _q, _r, _s, _t, _u, _v, _w, _x;
242
230
  const warnings = [];
243
- const revaiOptions = await (0, import_provider_utils2.parseProviderOptions)({
231
+ const revaiOptions = await parseProviderOptions({
244
232
  provider: "revai",
245
233
  providerOptions,
246
234
  schema: revaiTranscriptionModelOptionsSchema
247
235
  });
248
236
  const formData = new FormData();
249
- const blob = audio instanceof Uint8Array ? new Blob([audio]) : new Blob([(0, import_provider_utils2.convertBase64ToUint8Array)(audio)]);
250
- const fileExtension = (0, import_provider_utils2.mediaTypeToExtension)(mediaType);
237
+ const blob = audio instanceof Uint8Array ? new Blob([audio]) : new Blob([convertBase64ToUint8Array(audio)]);
238
+ const fileExtension = mediaTypeToExtension(mediaType);
251
239
  formData.append(
252
240
  "media",
253
241
  new File([blob], "audio", { type: mediaType }),
@@ -300,22 +288,22 @@ var RevaiTranscriptionModel = class {
300
288
  var _a, _b, _c, _d, _e, _f, _g, _h;
301
289
  const currentDate = (_c = (_b = (_a = this.config._internal) == null ? void 0 : _a.currentDate) == null ? void 0 : _b.call(_a)) != null ? _c : /* @__PURE__ */ new Date();
302
290
  const { formData, warnings } = await this.getArgs(options);
303
- const { value: submissionResponse } = await (0, import_provider_utils2.postFormDataToApi)({
291
+ const { value: submissionResponse } = await postFormDataToApi({
304
292
  url: this.config.url({
305
293
  path: "/speechtotext/v1/jobs",
306
294
  modelId: this.modelId
307
295
  }),
308
- headers: (0, import_provider_utils2.combineHeaders)(this.config.headers(), options.headers),
296
+ headers: combineHeaders(this.config.headers(), options.headers),
309
297
  formData,
310
298
  failedResponseHandler: revaiFailedResponseHandler,
311
- successfulResponseHandler: (0, import_provider_utils2.createJsonResponseHandler)(
299
+ successfulResponseHandler: createJsonResponseHandler(
312
300
  revaiTranscriptionJobResponseSchema
313
301
  ),
314
302
  abortSignal: options.abortSignal,
315
303
  fetch: this.config.fetch
316
304
  });
317
305
  if (submissionResponse.status === "failed") {
318
- throw new import_provider.AISDKError({
306
+ throw new AISDKError({
319
307
  message: "Failed to submit transcription job to Rev.ai",
320
308
  name: "TranscriptionJobSubmissionFailed",
321
309
  cause: submissionResponse
@@ -328,20 +316,20 @@ var RevaiTranscriptionModel = class {
328
316
  let jobResponse = submissionResponse;
329
317
  while (jobResponse.status !== "transcribed") {
330
318
  if (Date.now() - startTime > timeoutMs) {
331
- throw new import_provider.AISDKError({
319
+ throw new AISDKError({
332
320
  message: "Transcription job polling timed out",
333
321
  name: "TranscriptionJobPollingTimedOut",
334
322
  cause: submissionResponse
335
323
  });
336
324
  }
337
- const pollingResult = await (0, import_provider_utils2.getFromApi)({
325
+ const pollingResult = await getFromApi({
338
326
  url: this.config.url({
339
327
  path: `/speechtotext/v1/jobs/${jobId}`,
340
328
  modelId: this.modelId
341
329
  }),
342
- headers: (0, import_provider_utils2.combineHeaders)(this.config.headers(), options.headers),
330
+ headers: combineHeaders(this.config.headers(), options.headers),
343
331
  failedResponseHandler: revaiFailedResponseHandler,
344
- successfulResponseHandler: (0, import_provider_utils2.createJsonResponseHandler)(
332
+ successfulResponseHandler: createJsonResponseHandler(
345
333
  revaiTranscriptionJobResponseSchema
346
334
  ),
347
335
  abortSignal: options.abortSignal,
@@ -349,28 +337,28 @@ var RevaiTranscriptionModel = class {
349
337
  });
350
338
  jobResponse = pollingResult.value;
351
339
  if (jobResponse.status === "failed") {
352
- throw new import_provider.AISDKError({
340
+ throw new AISDKError({
353
341
  message: "Transcription job failed",
354
342
  name: "TranscriptionJobFailed",
355
343
  cause: jobResponse
356
344
  });
357
345
  }
358
346
  if (jobResponse.status !== "transcribed") {
359
- await (0, import_provider_utils2.delay)(pollingInterval);
347
+ await delay(pollingInterval);
360
348
  }
361
349
  }
362
350
  const {
363
351
  value: transcriptionResult,
364
352
  responseHeaders,
365
353
  rawValue: rawResponse
366
- } = await (0, import_provider_utils2.getFromApi)({
354
+ } = await getFromApi({
367
355
  url: this.config.url({
368
356
  path: `/speechtotext/v1/jobs/${jobId}/transcript`,
369
357
  modelId: this.modelId
370
358
  }),
371
- headers: (0, import_provider_utils2.combineHeaders)(this.config.headers(), options.headers),
359
+ headers: combineHeaders(this.config.headers(), options.headers),
372
360
  failedResponseHandler: revaiFailedResponseHandler,
373
- successfulResponseHandler: (0, import_provider_utils2.createJsonResponseHandler)(
361
+ successfulResponseHandler: createJsonResponseHandler(
374
362
  revaiTranscriptionResponseSchema
375
363
  ),
376
364
  abortSignal: options.abortSignal,
@@ -434,20 +422,20 @@ var RevaiTranscriptionModel = class {
434
422
  };
435
423
  }
436
424
  };
437
- var revaiTranscriptionJobResponseSchema = import_v42.z.object({
438
- id: import_v42.z.string().nullish(),
439
- status: import_v42.z.string().nullish(),
440
- language: import_v42.z.string().nullish()
425
+ var revaiTranscriptionJobResponseSchema = z2.object({
426
+ id: z2.string().nullish(),
427
+ status: z2.string().nullish(),
428
+ language: z2.string().nullish()
441
429
  });
442
- var revaiTranscriptionResponseSchema = import_v42.z.object({
443
- monologues: import_v42.z.array(
444
- import_v42.z.object({
445
- elements: import_v42.z.array(
446
- import_v42.z.object({
447
- type: import_v42.z.string().nullish(),
448
- value: import_v42.z.string().nullish(),
449
- ts: import_v42.z.number().nullish(),
450
- end_ts: import_v42.z.number().nullish()
430
+ var revaiTranscriptionResponseSchema = z2.object({
431
+ monologues: z2.array(
432
+ z2.object({
433
+ elements: z2.array(
434
+ z2.object({
435
+ type: z2.string().nullish(),
436
+ value: z2.string().nullish(),
437
+ ts: z2.number().nullish(),
438
+ end_ts: z2.number().nullish()
451
439
  })
452
440
  ).nullish()
453
441
  })
@@ -455,13 +443,13 @@ var revaiTranscriptionResponseSchema = import_v42.z.object({
455
443
  });
456
444
 
457
445
  // src/version.ts
458
- var VERSION = true ? "3.0.0-beta.2" : "0.0.0-test";
446
+ var VERSION = true ? "3.0.0-beta.21" : "0.0.0-test";
459
447
 
460
448
  // src/revai-provider.ts
461
449
  function createRevai(options = {}) {
462
- const getHeaders = () => (0, import_provider_utils3.withUserAgentSuffix)(
450
+ const getHeaders = () => withUserAgentSuffix(
463
451
  {
464
- authorization: `Bearer ${(0, import_provider_utils3.loadApiKey)({
452
+ authorization: `Bearer ${loadApiKey({
465
453
  apiKey: options.apiKey,
466
454
  environmentVariableName: "REVAI_API_KEY",
467
455
  description: "Rev.ai"
@@ -485,14 +473,14 @@ function createRevai(options = {}) {
485
473
  provider.transcription = createTranscriptionModel;
486
474
  provider.transcriptionModel = createTranscriptionModel;
487
475
  provider.languageModel = () => {
488
- throw new import_provider2.NoSuchModelError({
476
+ throw new NoSuchModelError({
489
477
  modelId: "unknown",
490
478
  modelType: "languageModel",
491
479
  message: "Rev.ai does not provide language models"
492
480
  });
493
481
  };
494
482
  provider.embeddingModel = () => {
495
- throw new import_provider2.NoSuchModelError({
483
+ throw new NoSuchModelError({
496
484
  modelId: "unknown",
497
485
  modelType: "embeddingModel",
498
486
  message: "Rev.ai does not provide text embedding models"
@@ -500,7 +488,7 @@ function createRevai(options = {}) {
500
488
  };
501
489
  provider.textEmbeddingModel = provider.embeddingModel;
502
490
  provider.imageModel = () => {
503
- throw new import_provider2.NoSuchModelError({
491
+ throw new NoSuchModelError({
504
492
  modelId: "unknown",
505
493
  modelType: "imageModel",
506
494
  message: "Rev.ai does not provide image models"
@@ -509,10 +497,9 @@ function createRevai(options = {}) {
509
497
  return provider;
510
498
  }
511
499
  var revai = createRevai();
512
- // Annotate the CommonJS export names for ESM import in node:
513
- 0 && (module.exports = {
500
+ export {
514
501
  VERSION,
515
502
  createRevai,
516
503
  revai
517
- });
504
+ };
518
505
  //# sourceMappingURL=index.js.map