voice-router-dev 0.2.6 → 0.2.7

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.mjs CHANGED
@@ -1935,12 +1935,27 @@ var GladiaAdapter = class extends BaseAdapter {
1935
1935
  utterances: this.extractUtterances(transcription),
1936
1936
  summary: result?.summarization?.results || void 0,
1937
1937
  metadata: {
1938
- requestParams: response.request_params,
1939
- customMetadata: response.custom_metadata
1938
+ requestParams: response.request_params
1940
1939
  },
1941
1940
  createdAt: response.created_at,
1942
1941
  completedAt: response.completed_at || void 0
1943
1942
  },
1943
+ // Extended data - fully typed from OpenAPI specs
1944
+ extended: {
1945
+ translation: result?.translation || void 0,
1946
+ moderation: result?.moderation || void 0,
1947
+ entities: result?.named_entity_recognition || void 0,
1948
+ sentiment: result?.sentiment_analysis || void 0,
1949
+ audioToLlm: result?.audio_to_llm || void 0,
1950
+ chapters: result?.chapterization || void 0,
1951
+ speakerReidentification: result?.speaker_reidentification || void 0,
1952
+ structuredData: result?.structured_data_extraction || void 0,
1953
+ customMetadata: response.custom_metadata || void 0
1954
+ },
1955
+ // Request tracking
1956
+ tracking: {
1957
+ requestId: response.request_id
1958
+ },
1944
1959
  raw: response
1945
1960
  };
1946
1961
  }
@@ -2796,18 +2811,29 @@ var AssemblyAIAdapter = class extends BaseAdapter {
2796
2811
  status,
2797
2812
  language: response.language_code,
2798
2813
  duration: response.audio_duration ? response.audio_duration / 1e3 : void 0,
2799
- // Convert ms to seconds
2800
2814
  speakers: this.extractSpeakers(response),
2801
2815
  words: this.extractWords(response),
2802
2816
  utterances: this.extractUtterances(response),
2803
2817
  summary: response.summary || void 0,
2804
2818
  metadata: {
2805
- audioUrl: response.audio_url,
2806
- entities: response.entities,
2807
- sentimentAnalysis: response.sentiment_analysis_results,
2808
- contentModeration: response.content_safety_labels
2819
+ audioUrl: response.audio_url
2809
2820
  }
2810
2821
  },
2822
+ // Extended data - fully typed from OpenAPI specs
2823
+ extended: {
2824
+ chapters: response.chapters || void 0,
2825
+ entities: response.entities || void 0,
2826
+ sentimentResults: response.sentiment_analysis_results || void 0,
2827
+ highlights: response.auto_highlights_result || void 0,
2828
+ contentSafety: response.content_safety_labels || void 0,
2829
+ topics: response.iab_categories_result || void 0,
2830
+ languageConfidence: response.language_confidence ?? void 0,
2831
+ throttled: response.throttled ?? void 0
2832
+ },
2833
+ // Request tracking
2834
+ tracking: {
2835
+ requestId: response.id
2836
+ },
2811
2837
  raw: response
2812
2838
  };
2813
2839
  }
@@ -3278,14 +3304,20 @@ var DeepgramAdapter = class extends BaseAdapter {
3278
3304
  speakers: this.extractSpeakers(response),
3279
3305
  words: this.extractWords(alternative),
3280
3306
  utterances: this.extractUtterances(response),
3281
- summary: this.extractSummary(alternative),
3282
- metadata: {
3283
- modelInfo: response.metadata?.model_info,
3284
- channels: response.metadata?.channels,
3285
- sentiment: response.results.sentiments,
3286
- intents: response.results.intents,
3287
- topics: response.results.topics
3288
- }
3307
+ summary: this.extractSummary(alternative)
3308
+ },
3309
+ // Extended data - fully typed from OpenAPI specs
3310
+ extended: {
3311
+ metadata: response.metadata,
3312
+ requestId: response.metadata?.request_id,
3313
+ sha256: response.metadata?.sha256,
3314
+ modelInfo: response.metadata?.model_info,
3315
+ tags: response.metadata?.tags
3316
+ },
3317
+ // Request tracking
3318
+ tracking: {
3319
+ requestId: response.metadata?.request_id,
3320
+ audioHash: response.metadata?.sha256
3289
3321
  },
3290
3322
  raw: response
3291
3323
  };
@@ -3801,11 +3833,12 @@ var AzureSTTAdapter = class extends BaseAdapter {
3801
3833
  id: String(speakerId),
3802
3834
  label: `Speaker ${speakerId}`
3803
3835
  })) : void 0;
3836
+ const transcriptionId = transcription.self?.split("/").pop() || "";
3804
3837
  return {
3805
3838
  success: true,
3806
3839
  provider: this.name,
3807
3840
  data: {
3808
- id: transcription.self?.split("/").pop() || "",
3841
+ id: transcriptionId,
3809
3842
  text: fullText,
3810
3843
  confidence: recognizedPhrases[0]?.nBest?.[0]?.confidence,
3811
3844
  status: "completed",
@@ -3816,6 +3849,10 @@ var AzureSTTAdapter = class extends BaseAdapter {
3816
3849
  createdAt: transcription.createdDateTime,
3817
3850
  completedAt: transcription.lastActionDateTime
3818
3851
  },
3852
+ extended: {},
3853
+ tracking: {
3854
+ requestId: transcriptionId
3855
+ },
3819
3856
  raw: {
3820
3857
  transcription,
3821
3858
  transcriptionData
@@ -4010,16 +4047,21 @@ var OpenAIWhisperAdapter = class extends BaseAdapter {
4010
4047
  */
4011
4048
  normalizeResponse(response, model, isDiarization) {
4012
4049
  if ("text" in response && Object.keys(response).length === 1) {
4050
+ const requestId2 = `openai-${Date.now()}`;
4013
4051
  return {
4014
4052
  success: true,
4015
4053
  provider: this.name,
4016
4054
  data: {
4017
- id: `openai-${Date.now()}`,
4055
+ id: requestId2,
4018
4056
  text: response.text,
4019
4057
  status: "completed",
4020
4058
  language: void 0,
4021
4059
  confidence: void 0
4022
4060
  },
4061
+ extended: {},
4062
+ tracking: {
4063
+ requestId: requestId2
4064
+ },
4023
4065
  raw: response
4024
4066
  };
4025
4067
  }
@@ -4038,11 +4080,12 @@ var OpenAIWhisperAdapter = class extends BaseAdapter {
4038
4080
  end: segment.end,
4039
4081
  confidence: void 0
4040
4082
  }));
4083
+ const requestId2 = `openai-${Date.now()}`;
4041
4084
  return {
4042
4085
  success: true,
4043
4086
  provider: this.name,
4044
4087
  data: {
4045
- id: `openai-${Date.now()}`,
4088
+ id: requestId2,
4046
4089
  text: diarizedResponse.text,
4047
4090
  status: "completed",
4048
4091
  language: void 0,
@@ -4050,6 +4093,10 @@ var OpenAIWhisperAdapter = class extends BaseAdapter {
4050
4093
  speakers,
4051
4094
  utterances
4052
4095
  },
4096
+ extended: {},
4097
+ tracking: {
4098
+ requestId: requestId2
4099
+ },
4053
4100
  raw: response
4054
4101
  };
4055
4102
  }
@@ -4061,28 +4108,38 @@ var OpenAIWhisperAdapter = class extends BaseAdapter {
4061
4108
  end: word.end,
4062
4109
  confidence: void 0
4063
4110
  }));
4111
+ const requestId2 = `openai-${Date.now()}`;
4064
4112
  return {
4065
4113
  success: true,
4066
4114
  provider: this.name,
4067
4115
  data: {
4068
- id: `openai-${Date.now()}`,
4116
+ id: requestId2,
4069
4117
  text: verboseResponse.text,
4070
4118
  status: "completed",
4071
4119
  language: verboseResponse.language,
4072
4120
  duration: verboseResponse.duration,
4073
4121
  words
4074
4122
  },
4123
+ extended: {},
4124
+ tracking: {
4125
+ requestId: requestId2
4126
+ },
4075
4127
  raw: response
4076
4128
  };
4077
4129
  }
4130
+ const requestId = `openai-${Date.now()}`;
4078
4131
  return {
4079
4132
  success: true,
4080
4133
  provider: this.name,
4081
4134
  data: {
4082
- id: `openai-${Date.now()}`,
4135
+ id: requestId,
4083
4136
  text: "text" in response ? response.text : "",
4084
4137
  status: "completed"
4085
4138
  },
4139
+ extended: {},
4140
+ tracking: {
4141
+ requestId
4142
+ },
4086
4143
  raw: response
4087
4144
  };
4088
4145
  }
@@ -4381,6 +4438,10 @@ var SpeechmaticsAdapter = class extends BaseAdapter {
4381
4438
  summary: response.summary?.content,
4382
4439
  createdAt: response.job.created_at
4383
4440
  },
4441
+ extended: {},
4442
+ tracking: {
4443
+ requestId: response.job.id
4444
+ },
4384
4445
  raw: response
4385
4446
  };
4386
4447
  }