voice-router-dev 0.2.6 → 0.2.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +154 -74
- package/dist/index.d.mts +709 -540
- package/dist/index.d.ts +709 -540
- package/dist/index.js +81 -20
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +81 -20
- package/dist/index.mjs.map +1 -1
- package/package.json +1 -1
package/dist/index.mjs
CHANGED
|
@@ -1935,12 +1935,27 @@ var GladiaAdapter = class extends BaseAdapter {
|
|
|
1935
1935
|
utterances: this.extractUtterances(transcription),
|
|
1936
1936
|
summary: result?.summarization?.results || void 0,
|
|
1937
1937
|
metadata: {
|
|
1938
|
-
requestParams: response.request_params
|
|
1939
|
-
customMetadata: response.custom_metadata
|
|
1938
|
+
requestParams: response.request_params
|
|
1940
1939
|
},
|
|
1941
1940
|
createdAt: response.created_at,
|
|
1942
1941
|
completedAt: response.completed_at || void 0
|
|
1943
1942
|
},
|
|
1943
|
+
// Extended data - fully typed from OpenAPI specs
|
|
1944
|
+
extended: {
|
|
1945
|
+
translation: result?.translation || void 0,
|
|
1946
|
+
moderation: result?.moderation || void 0,
|
|
1947
|
+
entities: result?.named_entity_recognition || void 0,
|
|
1948
|
+
sentiment: result?.sentiment_analysis || void 0,
|
|
1949
|
+
audioToLlm: result?.audio_to_llm || void 0,
|
|
1950
|
+
chapters: result?.chapterization || void 0,
|
|
1951
|
+
speakerReidentification: result?.speaker_reidentification || void 0,
|
|
1952
|
+
structuredData: result?.structured_data_extraction || void 0,
|
|
1953
|
+
customMetadata: response.custom_metadata || void 0
|
|
1954
|
+
},
|
|
1955
|
+
// Request tracking
|
|
1956
|
+
tracking: {
|
|
1957
|
+
requestId: response.request_id
|
|
1958
|
+
},
|
|
1944
1959
|
raw: response
|
|
1945
1960
|
};
|
|
1946
1961
|
}
|
|
@@ -2796,18 +2811,29 @@ var AssemblyAIAdapter = class extends BaseAdapter {
|
|
|
2796
2811
|
status,
|
|
2797
2812
|
language: response.language_code,
|
|
2798
2813
|
duration: response.audio_duration ? response.audio_duration / 1e3 : void 0,
|
|
2799
|
-
// Convert ms to seconds
|
|
2800
2814
|
speakers: this.extractSpeakers(response),
|
|
2801
2815
|
words: this.extractWords(response),
|
|
2802
2816
|
utterances: this.extractUtterances(response),
|
|
2803
2817
|
summary: response.summary || void 0,
|
|
2804
2818
|
metadata: {
|
|
2805
|
-
audioUrl: response.audio_url
|
|
2806
|
-
entities: response.entities,
|
|
2807
|
-
sentimentAnalysis: response.sentiment_analysis_results,
|
|
2808
|
-
contentModeration: response.content_safety_labels
|
|
2819
|
+
audioUrl: response.audio_url
|
|
2809
2820
|
}
|
|
2810
2821
|
},
|
|
2822
|
+
// Extended data - fully typed from OpenAPI specs
|
|
2823
|
+
extended: {
|
|
2824
|
+
chapters: response.chapters || void 0,
|
|
2825
|
+
entities: response.entities || void 0,
|
|
2826
|
+
sentimentResults: response.sentiment_analysis_results || void 0,
|
|
2827
|
+
highlights: response.auto_highlights_result || void 0,
|
|
2828
|
+
contentSafety: response.content_safety_labels || void 0,
|
|
2829
|
+
topics: response.iab_categories_result || void 0,
|
|
2830
|
+
languageConfidence: response.language_confidence ?? void 0,
|
|
2831
|
+
throttled: response.throttled ?? void 0
|
|
2832
|
+
},
|
|
2833
|
+
// Request tracking
|
|
2834
|
+
tracking: {
|
|
2835
|
+
requestId: response.id
|
|
2836
|
+
},
|
|
2811
2837
|
raw: response
|
|
2812
2838
|
};
|
|
2813
2839
|
}
|
|
@@ -3278,14 +3304,20 @@ var DeepgramAdapter = class extends BaseAdapter {
|
|
|
3278
3304
|
speakers: this.extractSpeakers(response),
|
|
3279
3305
|
words: this.extractWords(alternative),
|
|
3280
3306
|
utterances: this.extractUtterances(response),
|
|
3281
|
-
summary: this.extractSummary(alternative)
|
|
3282
|
-
|
|
3283
|
-
|
|
3284
|
-
|
|
3285
|
-
|
|
3286
|
-
|
|
3287
|
-
|
|
3288
|
-
|
|
3307
|
+
summary: this.extractSummary(alternative)
|
|
3308
|
+
},
|
|
3309
|
+
// Extended data - fully typed from OpenAPI specs
|
|
3310
|
+
extended: {
|
|
3311
|
+
metadata: response.metadata,
|
|
3312
|
+
requestId: response.metadata?.request_id,
|
|
3313
|
+
sha256: response.metadata?.sha256,
|
|
3314
|
+
modelInfo: response.metadata?.model_info,
|
|
3315
|
+
tags: response.metadata?.tags
|
|
3316
|
+
},
|
|
3317
|
+
// Request tracking
|
|
3318
|
+
tracking: {
|
|
3319
|
+
requestId: response.metadata?.request_id,
|
|
3320
|
+
audioHash: response.metadata?.sha256
|
|
3289
3321
|
},
|
|
3290
3322
|
raw: response
|
|
3291
3323
|
};
|
|
@@ -3801,11 +3833,12 @@ var AzureSTTAdapter = class extends BaseAdapter {
|
|
|
3801
3833
|
id: String(speakerId),
|
|
3802
3834
|
label: `Speaker ${speakerId}`
|
|
3803
3835
|
})) : void 0;
|
|
3836
|
+
const transcriptionId = transcription.self?.split("/").pop() || "";
|
|
3804
3837
|
return {
|
|
3805
3838
|
success: true,
|
|
3806
3839
|
provider: this.name,
|
|
3807
3840
|
data: {
|
|
3808
|
-
id:
|
|
3841
|
+
id: transcriptionId,
|
|
3809
3842
|
text: fullText,
|
|
3810
3843
|
confidence: recognizedPhrases[0]?.nBest?.[0]?.confidence,
|
|
3811
3844
|
status: "completed",
|
|
@@ -3816,6 +3849,10 @@ var AzureSTTAdapter = class extends BaseAdapter {
|
|
|
3816
3849
|
createdAt: transcription.createdDateTime,
|
|
3817
3850
|
completedAt: transcription.lastActionDateTime
|
|
3818
3851
|
},
|
|
3852
|
+
extended: {},
|
|
3853
|
+
tracking: {
|
|
3854
|
+
requestId: transcriptionId
|
|
3855
|
+
},
|
|
3819
3856
|
raw: {
|
|
3820
3857
|
transcription,
|
|
3821
3858
|
transcriptionData
|
|
@@ -4010,16 +4047,21 @@ var OpenAIWhisperAdapter = class extends BaseAdapter {
|
|
|
4010
4047
|
*/
|
|
4011
4048
|
normalizeResponse(response, model, isDiarization) {
|
|
4012
4049
|
if ("text" in response && Object.keys(response).length === 1) {
|
|
4050
|
+
const requestId2 = `openai-${Date.now()}`;
|
|
4013
4051
|
return {
|
|
4014
4052
|
success: true,
|
|
4015
4053
|
provider: this.name,
|
|
4016
4054
|
data: {
|
|
4017
|
-
id:
|
|
4055
|
+
id: requestId2,
|
|
4018
4056
|
text: response.text,
|
|
4019
4057
|
status: "completed",
|
|
4020
4058
|
language: void 0,
|
|
4021
4059
|
confidence: void 0
|
|
4022
4060
|
},
|
|
4061
|
+
extended: {},
|
|
4062
|
+
tracking: {
|
|
4063
|
+
requestId: requestId2
|
|
4064
|
+
},
|
|
4023
4065
|
raw: response
|
|
4024
4066
|
};
|
|
4025
4067
|
}
|
|
@@ -4038,11 +4080,12 @@ var OpenAIWhisperAdapter = class extends BaseAdapter {
|
|
|
4038
4080
|
end: segment.end,
|
|
4039
4081
|
confidence: void 0
|
|
4040
4082
|
}));
|
|
4083
|
+
const requestId2 = `openai-${Date.now()}`;
|
|
4041
4084
|
return {
|
|
4042
4085
|
success: true,
|
|
4043
4086
|
provider: this.name,
|
|
4044
4087
|
data: {
|
|
4045
|
-
id:
|
|
4088
|
+
id: requestId2,
|
|
4046
4089
|
text: diarizedResponse.text,
|
|
4047
4090
|
status: "completed",
|
|
4048
4091
|
language: void 0,
|
|
@@ -4050,6 +4093,10 @@ var OpenAIWhisperAdapter = class extends BaseAdapter {
|
|
|
4050
4093
|
speakers,
|
|
4051
4094
|
utterances
|
|
4052
4095
|
},
|
|
4096
|
+
extended: {},
|
|
4097
|
+
tracking: {
|
|
4098
|
+
requestId: requestId2
|
|
4099
|
+
},
|
|
4053
4100
|
raw: response
|
|
4054
4101
|
};
|
|
4055
4102
|
}
|
|
@@ -4061,28 +4108,38 @@ var OpenAIWhisperAdapter = class extends BaseAdapter {
|
|
|
4061
4108
|
end: word.end,
|
|
4062
4109
|
confidence: void 0
|
|
4063
4110
|
}));
|
|
4111
|
+
const requestId2 = `openai-${Date.now()}`;
|
|
4064
4112
|
return {
|
|
4065
4113
|
success: true,
|
|
4066
4114
|
provider: this.name,
|
|
4067
4115
|
data: {
|
|
4068
|
-
id:
|
|
4116
|
+
id: requestId2,
|
|
4069
4117
|
text: verboseResponse.text,
|
|
4070
4118
|
status: "completed",
|
|
4071
4119
|
language: verboseResponse.language,
|
|
4072
4120
|
duration: verboseResponse.duration,
|
|
4073
4121
|
words
|
|
4074
4122
|
},
|
|
4123
|
+
extended: {},
|
|
4124
|
+
tracking: {
|
|
4125
|
+
requestId: requestId2
|
|
4126
|
+
},
|
|
4075
4127
|
raw: response
|
|
4076
4128
|
};
|
|
4077
4129
|
}
|
|
4130
|
+
const requestId = `openai-${Date.now()}`;
|
|
4078
4131
|
return {
|
|
4079
4132
|
success: true,
|
|
4080
4133
|
provider: this.name,
|
|
4081
4134
|
data: {
|
|
4082
|
-
id:
|
|
4135
|
+
id: requestId,
|
|
4083
4136
|
text: "text" in response ? response.text : "",
|
|
4084
4137
|
status: "completed"
|
|
4085
4138
|
},
|
|
4139
|
+
extended: {},
|
|
4140
|
+
tracking: {
|
|
4141
|
+
requestId
|
|
4142
|
+
},
|
|
4086
4143
|
raw: response
|
|
4087
4144
|
};
|
|
4088
4145
|
}
|
|
@@ -4381,6 +4438,10 @@ var SpeechmaticsAdapter = class extends BaseAdapter {
|
|
|
4381
4438
|
summary: response.summary?.content,
|
|
4382
4439
|
createdAt: response.job.created_at
|
|
4383
4440
|
},
|
|
4441
|
+
extended: {},
|
|
4442
|
+
tracking: {
|
|
4443
|
+
requestId: response.job.id
|
|
4444
|
+
},
|
|
4384
4445
|
raw: response
|
|
4385
4446
|
};
|
|
4386
4447
|
}
|