voice-router-dev 0.7.5 → 0.7.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +59 -0
- package/dist/constants.d.mts +158 -110
- package/dist/constants.d.ts +158 -110
- package/dist/constants.js +252 -59
- package/dist/constants.mjs +247 -59
- package/dist/index.d.mts +3 -3
- package/dist/index.d.ts +3 -3
- package/dist/index.js +82 -60
- package/dist/index.mjs +82 -60
- package/dist/{transcriptWebhookNotification-CcoRrjT2.d.ts → transcriptWebhookNotification-BTxv69ck.d.ts} +3 -3
- package/dist/{transcriptWebhookNotification-CqJGQePy.d.mts → transcriptWebhookNotification-DCcbnAKP.d.mts} +3 -3
- package/dist/webhooks.d.mts +1 -1
- package/dist/webhooks.d.ts +1 -1
- package/package.json +32 -12
package/CHANGELOG.md
CHANGED
|
@@ -5,6 +5,65 @@ All notable changes to this project will be documented in this file.
|
|
|
5
5
|
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.1.0/),
|
|
6
6
|
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
|
|
7
7
|
|
|
8
|
+
## [0.7.7] - 2026-01-25
|
|
9
|
+
|
|
10
|
+
### Changed
|
|
11
|
+
|
|
12
|
+
#### Improved Package Exports for Bundler Compatibility
|
|
13
|
+
|
|
14
|
+
Fixed `ERR_PACKAGE_PATH_NOT_EXPORTED` errors in Next.js and other bundlers:
|
|
15
|
+
|
|
16
|
+
- Added `default` condition to all subpath exports as bundler fallback
|
|
17
|
+
- Added explicit file path exports (`./constants.mjs`, `./constants.js`, etc.) for direct aliasing
|
|
18
|
+
- Exported `./package.json` for tooling that needs version metadata
|
|
19
|
+
|
|
20
|
+
```json
|
|
21
|
+
// Now works without aliasing hacks
|
|
22
|
+
import { DeepgramModel } from 'voice-router-dev/constants'
|
|
23
|
+
import { DeepgramModel } from 'voice-router-dev/constants.mjs'
|
|
24
|
+
import pkg from 'voice-router-dev/package.json'
|
|
25
|
+
```
|
|
26
|
+
|
|
27
|
+
---
|
|
28
|
+
|
|
29
|
+
## [0.7.6] - 2026-01-25
|
|
30
|
+
|
|
31
|
+
### Added
|
|
32
|
+
|
|
33
|
+
#### Auto-Generated Model Constants
|
|
34
|
+
|
|
35
|
+
Replaced hardcoded model lists with auto-generated constants from provider APIs/specs:
|
|
36
|
+
|
|
37
|
+
```typescript
|
|
38
|
+
import {
|
|
39
|
+
DeepgramModel, // 48 models from Deepgram API
|
|
40
|
+
OpenAIModel, // 5 transcription models from OpenAI spec
|
|
41
|
+
OpenAIRealtimeModel // 14 realtime models from OpenAI spec
|
|
42
|
+
} from 'voice-router-dev/constants'
|
|
43
|
+
|
|
44
|
+
// All models now auto-sync when specs update
|
|
45
|
+
{ model: DeepgramModel["nova-3"] }
|
|
46
|
+
{ model: OpenAIModel["gpt-4o-transcribe"] }
|
|
47
|
+
{ model: OpenAIRealtimeModel["gpt-4o-realtime-preview"] }
|
|
48
|
+
```
|
|
49
|
+
|
|
50
|
+
**New generation scripts:**
|
|
51
|
+
- `generate-deepgram-models.js` - Fetches from `https://api.deepgram.com/v1/models`
|
|
52
|
+
- `generate-openai-models.js` - Extracts from orval-generated TypeScript types
|
|
53
|
+
|
|
54
|
+
**New exports:**
|
|
55
|
+
- `DeepgramModel`, `DeepgramModelCodes`, `DeepgramModelLabels`, `DeepgramModelCode`
|
|
56
|
+
- `OpenAIModelCodes`, `OpenAIModelLabels`, `OpenAIModelCode`
|
|
57
|
+
- `OpenAITranscriptionModel`, `OpenAITranscriptionModelCode`
|
|
58
|
+
- `OpenAIRealtimeModel`, `OpenAIRealtimeModelCode`
|
|
59
|
+
|
|
60
|
+
### Changed
|
|
61
|
+
|
|
62
|
+
- Models are now automatically synced during `prebuild` and `openapi:generate`
|
|
63
|
+
- Pipeline diagram updated to show MODEL EXTRACTION section
|
|
64
|
+
|
|
65
|
+
---
|
|
66
|
+
|
|
8
67
|
## [0.7.5] - 2026-01-25
|
|
9
68
|
|
|
10
69
|
### Added
|
package/dist/constants.d.mts
CHANGED
|
@@ -1,3 +1,88 @@
|
|
|
1
|
+
/**
|
|
2
|
+
* Generated from Deepgram API - DO NOT EDIT MANUALLY
|
|
3
|
+
*
|
|
4
|
+
* This file is auto-generated by scripts/generate-deepgram-models.js
|
|
5
|
+
* Run 'pnpm openapi:sync-deepgram-models' to regenerate.
|
|
6
|
+
*
|
|
7
|
+
* @see https://api.deepgram.com/v1/models
|
|
8
|
+
* @see https://developers.deepgram.com/docs/models-languages-overview
|
|
9
|
+
*
|
|
10
|
+
* @generated
|
|
11
|
+
*/
|
|
12
|
+
/**
|
|
13
|
+
* All Deepgram STT model codes
|
|
14
|
+
*/
|
|
15
|
+
declare const DeepgramModelCodes: readonly ["base", "conversationalai", "enhanced-automotive", "enhanced-drivethru", "enhanced-finance", "enhanced-general", "enhanced-meeting", "enhanced-phonecall", "finance", "general", "general-dQw4w9WgXcQ", "general-polaris", "meeting", "nova", "nova-2", "nova-2-atc", "nova-2-automotive", "nova-2-conversationalai", "nova-2-drivethru", "nova-2-ea", "nova-2-finance", "nova-2-general", "nova-2-medical", "nova-2-meeting", "nova-2-phonecall", "nova-2-video", "nova-2-voicemail", "nova-3", "nova-3-general", "nova-3-medical", "nova-drivethru", "nova-general", "nova-medical", "nova-phonecall", "nova-voicemail", "phonecall", "phonecall-dQw4w9WgXcQ", "phoneme", "polaris", "unknown", "video", "voicemail", "whisper", "whisper-base", "whisper-large", "whisper-medium", "whisper-small", "whisper-tiny"];
|
|
16
|
+
/**
|
|
17
|
+
* Type for Deepgram model codes
|
|
18
|
+
*/
|
|
19
|
+
type DeepgramModelCode = (typeof DeepgramModelCodes)[number];
|
|
20
|
+
/**
|
|
21
|
+
* Deepgram model constant object for autocomplete
|
|
22
|
+
*
|
|
23
|
+
* @example
|
|
24
|
+
* ```typescript
|
|
25
|
+
* import { DeepgramModel } from 'voice-router-dev/constants'
|
|
26
|
+
*
|
|
27
|
+
* { model: DeepgramModel["nova-3"] }
|
|
28
|
+
* { model: DeepgramModel["nova-2-medical"] }
|
|
29
|
+
* ```
|
|
30
|
+
*/
|
|
31
|
+
declare const DeepgramModel: {
|
|
32
|
+
readonly base: "base";
|
|
33
|
+
readonly conversationalai: "conversationalai";
|
|
34
|
+
readonly "enhanced-automotive": "enhanced-automotive";
|
|
35
|
+
readonly "enhanced-drivethru": "enhanced-drivethru";
|
|
36
|
+
readonly "enhanced-finance": "enhanced-finance";
|
|
37
|
+
readonly "enhanced-general": "enhanced-general";
|
|
38
|
+
readonly "enhanced-meeting": "enhanced-meeting";
|
|
39
|
+
readonly "enhanced-phonecall": "enhanced-phonecall";
|
|
40
|
+
readonly finance: "finance";
|
|
41
|
+
readonly general: "general";
|
|
42
|
+
readonly "general-dQw4w9WgXcQ": "general-dQw4w9WgXcQ";
|
|
43
|
+
readonly "general-polaris": "general-polaris";
|
|
44
|
+
readonly meeting: "meeting";
|
|
45
|
+
readonly nova: "nova";
|
|
46
|
+
readonly "nova-2": "nova-2";
|
|
47
|
+
readonly "nova-2-atc": "nova-2-atc";
|
|
48
|
+
readonly "nova-2-automotive": "nova-2-automotive";
|
|
49
|
+
readonly "nova-2-conversationalai": "nova-2-conversationalai";
|
|
50
|
+
readonly "nova-2-drivethru": "nova-2-drivethru";
|
|
51
|
+
readonly "nova-2-ea": "nova-2-ea";
|
|
52
|
+
readonly "nova-2-finance": "nova-2-finance";
|
|
53
|
+
readonly "nova-2-general": "nova-2-general";
|
|
54
|
+
readonly "nova-2-medical": "nova-2-medical";
|
|
55
|
+
readonly "nova-2-meeting": "nova-2-meeting";
|
|
56
|
+
readonly "nova-2-phonecall": "nova-2-phonecall";
|
|
57
|
+
readonly "nova-2-video": "nova-2-video";
|
|
58
|
+
readonly "nova-2-voicemail": "nova-2-voicemail";
|
|
59
|
+
readonly "nova-3": "nova-3";
|
|
60
|
+
readonly "nova-3-general": "nova-3-general";
|
|
61
|
+
readonly "nova-3-medical": "nova-3-medical";
|
|
62
|
+
readonly "nova-drivethru": "nova-drivethru";
|
|
63
|
+
readonly "nova-general": "nova-general";
|
|
64
|
+
readonly "nova-medical": "nova-medical";
|
|
65
|
+
readonly "nova-phonecall": "nova-phonecall";
|
|
66
|
+
readonly "nova-voicemail": "nova-voicemail";
|
|
67
|
+
readonly phonecall: "phonecall";
|
|
68
|
+
readonly "phonecall-dQw4w9WgXcQ": "phonecall-dQw4w9WgXcQ";
|
|
69
|
+
readonly phoneme: "phoneme";
|
|
70
|
+
readonly polaris: "polaris";
|
|
71
|
+
readonly unknown: "unknown";
|
|
72
|
+
readonly video: "video";
|
|
73
|
+
readonly voicemail: "voicemail";
|
|
74
|
+
readonly whisper: "whisper";
|
|
75
|
+
readonly "whisper-base": "whisper-base";
|
|
76
|
+
readonly "whisper-large": "whisper-large";
|
|
77
|
+
readonly "whisper-medium": "whisper-medium";
|
|
78
|
+
readonly "whisper-small": "whisper-small";
|
|
79
|
+
readonly "whisper-tiny": "whisper-tiny";
|
|
80
|
+
};
|
|
81
|
+
/**
|
|
82
|
+
* Model name to display label mapping
|
|
83
|
+
*/
|
|
84
|
+
declare const DeepgramModelLabels: Record<DeepgramModelCode, string>;
|
|
85
|
+
|
|
1
86
|
/**
|
|
2
87
|
* Generated from Deepgram API - DO NOT EDIT MANUALLY
|
|
3
88
|
*
|
|
@@ -1594,6 +1679,44 @@ declare const AzureLocale: {
|
|
|
1594
1679
|
readonly "zu-ZA": "zu-ZA";
|
|
1595
1680
|
};
|
|
1596
1681
|
|
|
1682
|
+
/**
|
|
1683
|
+
* Generated from OpenAI TypeScript types - DO NOT EDIT MANUALLY
|
|
1684
|
+
*
|
|
1685
|
+
* This file is auto-generated by scripts/generate-openai-models.js
|
|
1686
|
+
* Run 'pnpm openapi:sync-openai-models' to regenerate.
|
|
1687
|
+
*
|
|
1688
|
+
* Source files:
|
|
1689
|
+
* - src/generated/openai/schema/createTranscriptionRequestModel.ts
|
|
1690
|
+
* - src/generated/openai/schema/realtimeSessionCreateRequestGAModel.ts
|
|
1691
|
+
* - src/generated/openai/schema/audioTranscriptionModel.ts
|
|
1692
|
+
*
|
|
1693
|
+
* @generated
|
|
1694
|
+
*/
|
|
1695
|
+
/**
|
|
1696
|
+
* All OpenAI audio model codes (transcription + realtime)
|
|
1697
|
+
*/
|
|
1698
|
+
declare const OpenAIModelCodes$1: readonly ["gpt-4o-mini-realtime-preview", "gpt-4o-mini-realtime-preview-2024-12-17", "gpt-4o-mini-transcribe", "gpt-4o-mini-transcribe-2025-12-15", "gpt-4o-realtime-preview", "gpt-4o-realtime-preview-2024-10-01", "gpt-4o-realtime-preview-2024-12-17", "gpt-4o-realtime-preview-2025-06-03", "gpt-4o-transcribe", "gpt-4o-transcribe-diarize", "gpt-audio-mini", "gpt-audio-mini-2025-10-06", "gpt-audio-mini-2025-12-15", "gpt-realtime", "gpt-realtime-2025-08-28", "gpt-realtime-mini", "gpt-realtime-mini-2025-10-06", "gpt-realtime-mini-2025-12-15", "whisper-1"];
|
|
1699
|
+
/**
|
|
1700
|
+
* Type for all OpenAI model codes
|
|
1701
|
+
*/
|
|
1702
|
+
type OpenAIModelCode = (typeof OpenAIModelCodes$1)[number];
|
|
1703
|
+
/**
|
|
1704
|
+
* OpenAI transcription model codes (for batch transcription)
|
|
1705
|
+
*/
|
|
1706
|
+
declare const OpenAITranscriptionModelCodes: readonly ["gpt-4o-mini-transcribe", "gpt-4o-mini-transcribe-2025-12-15", "gpt-4o-transcribe", "gpt-4o-transcribe-diarize", "whisper-1"];
|
|
1707
|
+
/**
|
|
1708
|
+
* Type for OpenAI transcription model codes
|
|
1709
|
+
*/
|
|
1710
|
+
type OpenAITranscriptionModelCode = (typeof OpenAITranscriptionModelCodes)[number];
|
|
1711
|
+
/**
|
|
1712
|
+
* OpenAI realtime model codes (for streaming)
|
|
1713
|
+
*/
|
|
1714
|
+
declare const OpenAIRealtimeModelCodes$1: readonly ["gpt-4o-mini-realtime-preview", "gpt-4o-mini-realtime-preview-2024-12-17", "gpt-4o-realtime-preview", "gpt-4o-realtime-preview-2024-10-01", "gpt-4o-realtime-preview-2024-12-17", "gpt-4o-realtime-preview-2025-06-03", "gpt-audio-mini", "gpt-audio-mini-2025-10-06", "gpt-audio-mini-2025-12-15", "gpt-realtime", "gpt-realtime-2025-08-28", "gpt-realtime-mini", "gpt-realtime-mini-2025-10-06", "gpt-realtime-mini-2025-12-15"];
|
|
1715
|
+
/**
|
|
1716
|
+
* Type for OpenAI realtime model codes
|
|
1717
|
+
*/
|
|
1718
|
+
type OpenAIRealtimeModelCode = (typeof OpenAIRealtimeModelCodes$1)[number];
|
|
1719
|
+
|
|
1597
1720
|
/**
|
|
1598
1721
|
* Browser-safe constants for speech-to-text providers
|
|
1599
1722
|
*
|
|
@@ -1720,48 +1843,6 @@ declare const DeepgramSampleRate: {
|
|
|
1720
1843
|
readonly NUMBER_44100: 44100;
|
|
1721
1844
|
readonly NUMBER_48000: 48000;
|
|
1722
1845
|
};
|
|
1723
|
-
/**
|
|
1724
|
-
* Deepgram transcription models
|
|
1725
|
-
*
|
|
1726
|
-
* @example
|
|
1727
|
-
* ```typescript
|
|
1728
|
-
* import { DeepgramModel } from 'voice-router-dev/constants'
|
|
1729
|
-
*
|
|
1730
|
-
* { model: DeepgramModel["nova-3"] }
|
|
1731
|
-
* { model: DeepgramModel["nova-2-medical"] }
|
|
1732
|
-
* ```
|
|
1733
|
-
*/
|
|
1734
|
-
declare const DeepgramModel: {
|
|
1735
|
-
readonly "nova-3": "nova-3";
|
|
1736
|
-
readonly "nova-3-general": "nova-3-general";
|
|
1737
|
-
readonly "nova-3-medical": "nova-3-medical";
|
|
1738
|
-
readonly "nova-2": "nova-2";
|
|
1739
|
-
readonly "nova-2-general": "nova-2-general";
|
|
1740
|
-
readonly "nova-2-meeting": "nova-2-meeting";
|
|
1741
|
-
readonly "nova-2-finance": "nova-2-finance";
|
|
1742
|
-
readonly "nova-2-conversationalai": "nova-2-conversationalai";
|
|
1743
|
-
readonly "nova-2-voicemail": "nova-2-voicemail";
|
|
1744
|
-
readonly "nova-2-video": "nova-2-video";
|
|
1745
|
-
readonly "nova-2-medical": "nova-2-medical";
|
|
1746
|
-
readonly "nova-2-drivethru": "nova-2-drivethru";
|
|
1747
|
-
readonly "nova-2-automotive": "nova-2-automotive";
|
|
1748
|
-
readonly nova: "nova";
|
|
1749
|
-
readonly "nova-general": "nova-general";
|
|
1750
|
-
readonly "nova-phonecall": "nova-phonecall";
|
|
1751
|
-
readonly "nova-medical": "nova-medical";
|
|
1752
|
-
readonly enhanced: "enhanced";
|
|
1753
|
-
readonly "enhanced-general": "enhanced-general";
|
|
1754
|
-
readonly "enhanced-meeting": "enhanced-meeting";
|
|
1755
|
-
readonly "enhanced-phonecall": "enhanced-phonecall";
|
|
1756
|
-
readonly "enhanced-finance": "enhanced-finance";
|
|
1757
|
-
readonly base: "base";
|
|
1758
|
-
readonly meeting: "meeting";
|
|
1759
|
-
readonly phonecall: "phonecall";
|
|
1760
|
-
readonly finance: "finance";
|
|
1761
|
-
readonly conversationalai: "conversationalai";
|
|
1762
|
-
readonly voicemail: "voicemail";
|
|
1763
|
-
readonly video: "video";
|
|
1764
|
-
};
|
|
1765
1846
|
|
|
1766
1847
|
/**
|
|
1767
1848
|
* Speechmatics operating point (model quality tier)
|
|
@@ -2402,8 +2483,6 @@ type DeepgramIntentModeType = (typeof DeepgramIntentMode)[keyof typeof DeepgramI
|
|
|
2402
2483
|
type DeepgramCallbackMethodType = (typeof DeepgramCallbackMethod)[keyof typeof DeepgramCallbackMethod];
|
|
2403
2484
|
/** Deepgram sample rate type derived from const object */
|
|
2404
2485
|
type DeepgramSampleRateType = (typeof DeepgramSampleRate)[keyof typeof DeepgramSampleRate];
|
|
2405
|
-
/** Deepgram model type derived from const object */
|
|
2406
|
-
type DeepgramModelType = (typeof DeepgramModel)[keyof typeof DeepgramModel];
|
|
2407
2486
|
|
|
2408
2487
|
/** Gladia encoding type derived from const object */
|
|
2409
2488
|
type GladiaEncodingType = (typeof GladiaEncoding)[keyof typeof GladiaEncoding];
|
|
@@ -2703,31 +2782,40 @@ type DeepgramTTSContainerType = (typeof DeepgramTTSContainer)[keyof typeof Deepg
|
|
|
2703
2782
|
/** Deepgram TTS sample rate type derived from const object */
|
|
2704
2783
|
type DeepgramTTSSampleRateType = (typeof DeepgramTTSSampleRate)[keyof typeof DeepgramTTSSampleRate];
|
|
2705
2784
|
/**
|
|
2706
|
-
* OpenAI
|
|
2707
|
-
*
|
|
2708
|
-
* Values from official spec (auto-synced from Stainless):
|
|
2709
|
-
* - `whisper-1`: Open source Whisper V2 model
|
|
2710
|
-
* - `gpt-4o-transcribe`: GPT-4o based transcription (more accurate)
|
|
2711
|
-
* - `gpt-4o-mini-transcribe`: Faster, cost-effective GPT-4o mini
|
|
2712
|
-
* - `gpt-4o-mini-transcribe-2025-12-15`: Dated version of GPT-4o mini
|
|
2713
|
-
* - `gpt-4o-transcribe-diarize`: GPT-4o with speaker diarization
|
|
2714
|
-
*
|
|
2715
|
-
* @example
|
|
2716
|
-
* ```typescript
|
|
2717
|
-
* import { OpenAIModel } from 'voice-router-dev/constants'
|
|
2718
|
-
*
|
|
2719
|
-
* { model: OpenAIModel["whisper-1"] }
|
|
2720
|
-
* { model: OpenAIModel["gpt-4o-transcribe"] }
|
|
2721
|
-
* { model: OpenAIModel["gpt-4o-transcribe-diarize"] }
|
|
2722
|
-
* ```
|
|
2785
|
+
* OpenAI transcription models (batch)
|
|
2786
|
+
* @see scripts/generate-openai-models.js
|
|
2723
2787
|
*/
|
|
2724
2788
|
declare const OpenAIModel: {
|
|
2725
|
-
readonly "whisper-1": "whisper-1";
|
|
2726
|
-
readonly "gpt-4o-transcribe": "gpt-4o-transcribe";
|
|
2727
2789
|
readonly "gpt-4o-mini-transcribe": "gpt-4o-mini-transcribe";
|
|
2728
2790
|
readonly "gpt-4o-mini-transcribe-2025-12-15": "gpt-4o-mini-transcribe-2025-12-15";
|
|
2791
|
+
readonly "gpt-4o-transcribe": "gpt-4o-transcribe";
|
|
2729
2792
|
readonly "gpt-4o-transcribe-diarize": "gpt-4o-transcribe-diarize";
|
|
2793
|
+
readonly "whisper-1": "whisper-1";
|
|
2794
|
+
};
|
|
2795
|
+
declare const OpenAIModelCodes: readonly ["gpt-4o-mini-realtime-preview", "gpt-4o-mini-realtime-preview-2024-12-17", "gpt-4o-mini-transcribe", "gpt-4o-mini-transcribe-2025-12-15", "gpt-4o-realtime-preview", "gpt-4o-realtime-preview-2024-10-01", "gpt-4o-realtime-preview-2024-12-17", "gpt-4o-realtime-preview-2025-06-03", "gpt-4o-transcribe", "gpt-4o-transcribe-diarize", "gpt-audio-mini", "gpt-audio-mini-2025-10-06", "gpt-audio-mini-2025-12-15", "gpt-realtime", "gpt-realtime-2025-08-28", "gpt-realtime-mini", "gpt-realtime-mini-2025-10-06", "gpt-realtime-mini-2025-12-15", "whisper-1"];
|
|
2796
|
+
declare const OpenAIModelLabels: Record<"gpt-4o-mini-realtime-preview" | "gpt-4o-mini-realtime-preview-2024-12-17" | "gpt-4o-mini-transcribe" | "gpt-4o-mini-transcribe-2025-12-15" | "gpt-4o-realtime-preview" | "gpt-4o-realtime-preview-2024-10-01" | "gpt-4o-realtime-preview-2024-12-17" | "gpt-4o-realtime-preview-2025-06-03" | "gpt-4o-transcribe" | "gpt-4o-transcribe-diarize" | "gpt-audio-mini" | "gpt-audio-mini-2025-10-06" | "gpt-audio-mini-2025-12-15" | "gpt-realtime" | "gpt-realtime-2025-08-28" | "gpt-realtime-mini" | "gpt-realtime-mini-2025-10-06" | "gpt-realtime-mini-2025-12-15" | "whisper-1", string>;
|
|
2797
|
+
/**
|
|
2798
|
+
* OpenAI Realtime API models (streaming)
|
|
2799
|
+
* @see scripts/generate-openai-models.js
|
|
2800
|
+
*/
|
|
2801
|
+
declare const OpenAIRealtimeModel: {
|
|
2802
|
+
readonly "gpt-4o-mini-realtime-preview": "gpt-4o-mini-realtime-preview";
|
|
2803
|
+
readonly "gpt-4o-mini-realtime-preview-2024-12-17": "gpt-4o-mini-realtime-preview-2024-12-17";
|
|
2804
|
+
readonly "gpt-4o-realtime-preview": "gpt-4o-realtime-preview";
|
|
2805
|
+
readonly "gpt-4o-realtime-preview-2024-10-01": "gpt-4o-realtime-preview-2024-10-01";
|
|
2806
|
+
readonly "gpt-4o-realtime-preview-2024-12-17": "gpt-4o-realtime-preview-2024-12-17";
|
|
2807
|
+
readonly "gpt-4o-realtime-preview-2025-06-03": "gpt-4o-realtime-preview-2025-06-03";
|
|
2808
|
+
readonly "gpt-audio-mini": "gpt-audio-mini";
|
|
2809
|
+
readonly "gpt-audio-mini-2025-10-06": "gpt-audio-mini-2025-10-06";
|
|
2810
|
+
readonly "gpt-audio-mini-2025-12-15": "gpt-audio-mini-2025-12-15";
|
|
2811
|
+
readonly "gpt-realtime": "gpt-realtime";
|
|
2812
|
+
readonly "gpt-realtime-2025-08-28": "gpt-realtime-2025-08-28";
|
|
2813
|
+
readonly "gpt-realtime-mini": "gpt-realtime-mini";
|
|
2814
|
+
readonly "gpt-realtime-mini-2025-10-06": "gpt-realtime-mini-2025-10-06";
|
|
2815
|
+
readonly "gpt-realtime-mini-2025-12-15": "gpt-realtime-mini-2025-12-15";
|
|
2730
2816
|
};
|
|
2817
|
+
declare const OpenAIRealtimeModelCodes: readonly ["gpt-4o-mini-realtime-preview", "gpt-4o-mini-realtime-preview-2024-12-17", "gpt-4o-realtime-preview", "gpt-4o-realtime-preview-2024-10-01", "gpt-4o-realtime-preview-2024-12-17", "gpt-4o-realtime-preview-2025-06-03", "gpt-audio-mini", "gpt-audio-mini-2025-10-06", "gpt-audio-mini-2025-12-15", "gpt-realtime", "gpt-realtime-2025-08-28", "gpt-realtime-mini", "gpt-realtime-mini-2025-10-06", "gpt-realtime-mini-2025-12-15"];
|
|
2818
|
+
|
|
2731
2819
|
/**
|
|
2732
2820
|
* OpenAI transcription response formats
|
|
2733
2821
|
*
|
|
@@ -2739,9 +2827,6 @@ declare const OpenAIModel: {
|
|
|
2739
2827
|
* - `vtt`: VTT subtitle format
|
|
2740
2828
|
* - `diarized_json`: JSON with speaker annotations (gpt-4o-transcribe-diarize only)
|
|
2741
2829
|
*
|
|
2742
|
-
* Note: GPT-4o transcribe models only support `json` format.
|
|
2743
|
-
* For diarization, use `diarized_json` with `gpt-4o-transcribe-diarize` model.
|
|
2744
|
-
*
|
|
2745
2830
|
* @example
|
|
2746
2831
|
* ```typescript
|
|
2747
2832
|
* import { OpenAIResponseFormat } from 'voice-router-dev/constants'
|
|
@@ -2762,36 +2847,8 @@ declare const OpenAIResponseFormat: {
|
|
|
2762
2847
|
type OpenAIModelType = (typeof OpenAIModel)[keyof typeof OpenAIModel];
|
|
2763
2848
|
/** OpenAI response format type derived from const object */
|
|
2764
2849
|
type OpenAIResponseFormatType = (typeof OpenAIResponseFormat)[keyof typeof OpenAIResponseFormat];
|
|
2765
|
-
/**
|
|
2766
|
-
|
|
2767
|
-
*
|
|
2768
|
-
* Values from official spec (auto-synced from Stainless):
|
|
2769
|
-
* - `gpt-4o-realtime-preview`: Latest GPT-4o realtime preview
|
|
2770
|
-
* - `gpt-4o-realtime-preview-2024-10-01`: October 2024 version
|
|
2771
|
-
* - `gpt-4o-realtime-preview-2024-12-17`: December 2024 version
|
|
2772
|
-
* - `gpt-4o-realtime-preview-2025-06-03`: June 2025 version
|
|
2773
|
-
* - `gpt-4o-mini-realtime-preview`: GPT-4o mini realtime
|
|
2774
|
-
* - `gpt-4o-mini-realtime-preview-2024-12-17`: December 2024 mini version
|
|
2775
|
-
*
|
|
2776
|
-
* @example
|
|
2777
|
-
* ```typescript
|
|
2778
|
-
* import { OpenAIRealtimeModel } from 'voice-router-dev/constants'
|
|
2779
|
-
*
|
|
2780
|
-
* await adapter.transcribeStream({
|
|
2781
|
-
* openaiStreaming: {
|
|
2782
|
-
* model: OpenAIRealtimeModel["gpt-4o-realtime-preview"]
|
|
2783
|
-
* }
|
|
2784
|
-
* })
|
|
2785
|
-
* ```
|
|
2786
|
-
*/
|
|
2787
|
-
declare const OpenAIRealtimeModel: {
|
|
2788
|
-
readonly "gpt-4o-realtime-preview": "gpt-4o-realtime-preview";
|
|
2789
|
-
readonly "gpt-4o-realtime-preview-2024-10-01": "gpt-4o-realtime-preview-2024-10-01";
|
|
2790
|
-
readonly "gpt-4o-realtime-preview-2024-12-17": "gpt-4o-realtime-preview-2024-12-17";
|
|
2791
|
-
readonly "gpt-4o-realtime-preview-2025-06-03": "gpt-4o-realtime-preview-2025-06-03";
|
|
2792
|
-
readonly "gpt-4o-mini-realtime-preview": "gpt-4o-mini-realtime-preview";
|
|
2793
|
-
readonly "gpt-4o-mini-realtime-preview-2024-12-17": "gpt-4o-mini-realtime-preview-2024-12-17";
|
|
2794
|
-
};
|
|
2850
|
+
/** OpenAI Realtime model type derived from const object */
|
|
2851
|
+
type OpenAIRealtimeModelType = (typeof OpenAIRealtimeModel)[keyof typeof OpenAIRealtimeModel];
|
|
2795
2852
|
/**
|
|
2796
2853
|
* OpenAI Realtime audio input formats
|
|
2797
2854
|
*
|
|
@@ -2839,25 +2896,16 @@ declare const OpenAIRealtimeTurnDetection: {
|
|
|
2839
2896
|
};
|
|
2840
2897
|
/**
|
|
2841
2898
|
* OpenAI Realtime input transcription models
|
|
2842
|
-
*
|
|
2843
|
-
*
|
|
2844
|
-
*
|
|
2845
|
-
* @example
|
|
2846
|
-
* ```typescript
|
|
2847
|
-
* import { OpenAIRealtimeTranscriptionModel } from 'voice-router-dev/constants'
|
|
2848
|
-
*
|
|
2849
|
-
* // whisper-1 is the standard transcription model
|
|
2850
|
-
* ```
|
|
2899
|
+
* (Alias for OpenAITranscriptionModel - same models work for realtime input transcription)
|
|
2900
|
+
* @see scripts/generate-openai-models.js
|
|
2851
2901
|
*/
|
|
2852
2902
|
declare const OpenAIRealtimeTranscriptionModel: {
|
|
2853
|
-
readonly "whisper-1": "whisper-1";
|
|
2854
|
-
readonly "gpt-4o-transcribe": "gpt-4o-transcribe";
|
|
2855
2903
|
readonly "gpt-4o-mini-transcribe": "gpt-4o-mini-transcribe";
|
|
2856
2904
|
readonly "gpt-4o-mini-transcribe-2025-12-15": "gpt-4o-mini-transcribe-2025-12-15";
|
|
2905
|
+
readonly "gpt-4o-transcribe": "gpt-4o-transcribe";
|
|
2857
2906
|
readonly "gpt-4o-transcribe-diarize": "gpt-4o-transcribe-diarize";
|
|
2907
|
+
readonly "whisper-1": "whisper-1";
|
|
2858
2908
|
};
|
|
2859
|
-
/** OpenAI Realtime model type */
|
|
2860
|
-
type OpenAIRealtimeModelType = (typeof OpenAIRealtimeModel)[keyof typeof OpenAIRealtimeModel];
|
|
2861
2909
|
/** OpenAI Realtime audio format type */
|
|
2862
2910
|
type OpenAIRealtimeAudioFormatType = (typeof OpenAIRealtimeAudioFormat)[keyof typeof OpenAIRealtimeAudioFormat];
|
|
2863
2911
|
/** OpenAI Realtime turn detection type */
|
|
@@ -2925,4 +2973,4 @@ declare const OpenAILanguage: {
|
|
|
2925
2973
|
/** OpenAI language type */
|
|
2926
2974
|
type OpenAILanguageType = (typeof OpenAILanguageCodes)[number];
|
|
2927
2975
|
|
|
2928
|
-
export { AssemblyAIEncoding, type AssemblyAIEncodingType, AssemblyAILanguage, type AssemblyAILanguageType, AssemblyAISampleRate, type AssemblyAISampleRateType, AssemblyAISpeechModel, type AssemblyAISpeechModelType, AssemblyAIStatus, type AssemblyAIStatusType, AssemblyAITranscriptionModel, type AssemblyAITranscriptionModelType, AzureLocale, type AzureLocaleCode, AzureLocaleCodes, AzureLocaleLabels, type AzureLocaleType, AzureLocales, AzureStatus, type AzureStatusType, type DeepgramArchitecture, DeepgramArchitectureLanguages, DeepgramArchitectures, DeepgramCallbackMethod, type DeepgramCallbackMethodType, DeepgramEncoding, type DeepgramEncodingType, DeepgramIntentMode, type DeepgramIntentModeType, DeepgramLanguage, type DeepgramLanguageCode, DeepgramLanguageCodes, type DeepgramLanguageCode as DeepgramLanguageType, DeepgramModel, type DeepgramModelType, type DeepgramMultilingualArchitecture, DeepgramMultilingualArchitectures, DeepgramRedact, type DeepgramRedactType, DeepgramRegion, type DeepgramRegionType, DeepgramSampleRate, type DeepgramSampleRateType, DeepgramStatus, type DeepgramStatusType, DeepgramTTSContainer, type DeepgramTTSContainerType, DeepgramTTSEncoding, type DeepgramTTSEncodingType, DeepgramTTSModel, type DeepgramTTSModelType, DeepgramTTSSampleRate, type DeepgramTTSSampleRateType, DeepgramTopicMode, type DeepgramTopicModeType, GladiaBitDepth, type GladiaBitDepthType, GladiaEncoding, type GladiaEncodingType, GladiaLanguage, type GladiaLanguageType, GladiaModel, type GladiaModelType, GladiaRegion, type GladiaRegionType, GladiaSampleRate, type GladiaSampleRateType, GladiaStatus, type GladiaStatusType, GladiaTranslationLanguage, type GladiaTranslationLanguageType, OpenAILanguage, OpenAILanguageCodes, type OpenAILanguageType, OpenAIModel, type OpenAIModelType, OpenAIRealtimeAudioFormat, type OpenAIRealtimeAudioFormatType, OpenAIRealtimeModel, type OpenAIRealtimeModelType, OpenAIRealtimeTranscriptionModel, type OpenAIRealtimeTranscriptionModelType, OpenAIRealtimeTurnDetection, type OpenAIRealtimeTurnDetectionType, OpenAIResponseFormat, type OpenAIResponseFormatType, SonioxAsyncModel, type SonioxAsyncModelCode, SonioxAsyncModelCodes, SonioxLanguage, type SonioxLanguageCode, SonioxLanguageCodes, SonioxLanguageLabels, type SonioxLanguageType, SonioxLanguages, SonioxModel, type SonioxModelCode, SonioxModelCodes, SonioxModelLabels, SonioxModels, SonioxRealtimeModel, type SonioxRealtimeModelCode, SonioxRealtimeModelCodes, SonioxRegion, type SonioxRegionType, SpeechmaticsLanguage, type SpeechmaticsLanguageCode, SpeechmaticsLanguageCodes, SpeechmaticsLanguageLabels, type SpeechmaticsLanguageType, SpeechmaticsLanguages, SpeechmaticsOperatingPoint, type SpeechmaticsOperatingPointType, SpeechmaticsRegion, type SpeechmaticsRegionType };
|
|
2976
|
+
export { AssemblyAIEncoding, type AssemblyAIEncodingType, AssemblyAILanguage, type AssemblyAILanguageType, AssemblyAISampleRate, type AssemblyAISampleRateType, AssemblyAISpeechModel, type AssemblyAISpeechModelType, AssemblyAIStatus, type AssemblyAIStatusType, AssemblyAITranscriptionModel, type AssemblyAITranscriptionModelType, AzureLocale, type AzureLocaleCode, AzureLocaleCodes, AzureLocaleLabels, type AzureLocaleType, AzureLocales, AzureStatus, type AzureStatusType, type DeepgramArchitecture, DeepgramArchitectureLanguages, DeepgramArchitectures, DeepgramCallbackMethod, type DeepgramCallbackMethodType, DeepgramEncoding, type DeepgramEncodingType, DeepgramIntentMode, type DeepgramIntentModeType, DeepgramLanguage, type DeepgramLanguageCode, DeepgramLanguageCodes, type DeepgramLanguageCode as DeepgramLanguageType, DeepgramModel, type DeepgramModelCode, DeepgramModelCodes, DeepgramModelLabels, type DeepgramModelCode as DeepgramModelType, type DeepgramMultilingualArchitecture, DeepgramMultilingualArchitectures, DeepgramRedact, type DeepgramRedactType, DeepgramRegion, type DeepgramRegionType, DeepgramSampleRate, type DeepgramSampleRateType, DeepgramStatus, type DeepgramStatusType, DeepgramTTSContainer, type DeepgramTTSContainerType, DeepgramTTSEncoding, type DeepgramTTSEncodingType, DeepgramTTSModel, type DeepgramTTSModelType, DeepgramTTSSampleRate, type DeepgramTTSSampleRateType, DeepgramTopicMode, type DeepgramTopicModeType, GladiaBitDepth, type GladiaBitDepthType, GladiaEncoding, type GladiaEncodingType, GladiaLanguage, type GladiaLanguageType, GladiaModel, type GladiaModelType, GladiaRegion, type GladiaRegionType, GladiaSampleRate, type GladiaSampleRateType, GladiaStatus, type GladiaStatusType, GladiaTranslationLanguage, type GladiaTranslationLanguageType, OpenAILanguage, OpenAILanguageCodes, type OpenAILanguageType, OpenAIModel, type OpenAIModelCode, OpenAIModelCodes, OpenAIModelLabels, type OpenAIModelType, OpenAIRealtimeAudioFormat, type OpenAIRealtimeAudioFormatType, OpenAIRealtimeModel, type OpenAIRealtimeModelCode, OpenAIRealtimeModelCodes, type OpenAIRealtimeModelType, OpenAIRealtimeTranscriptionModel, type OpenAIRealtimeTranscriptionModelType, OpenAIRealtimeTurnDetection, type OpenAIRealtimeTurnDetectionType, OpenAIResponseFormat, type OpenAIResponseFormatType, type OpenAITranscriptionModelCode, SonioxAsyncModel, type SonioxAsyncModelCode, SonioxAsyncModelCodes, SonioxLanguage, type SonioxLanguageCode, SonioxLanguageCodes, SonioxLanguageLabels, type SonioxLanguageType, SonioxLanguages, SonioxModel, type SonioxModelCode, SonioxModelCodes, SonioxModelLabels, SonioxModels, SonioxRealtimeModel, type SonioxRealtimeModelCode, SonioxRealtimeModelCodes, SonioxRegion, type SonioxRegionType, SpeechmaticsLanguage, type SpeechmaticsLanguageCode, SpeechmaticsLanguageCodes, SpeechmaticsLanguageLabels, type SpeechmaticsLanguageType, SpeechmaticsLanguages, SpeechmaticsOperatingPoint, type SpeechmaticsOperatingPointType, SpeechmaticsRegion, type SpeechmaticsRegionType };
|