@wix/auto_sdk_ai-gateway_prompts 1.0.11 → 1.0.12
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/build/cjs/index.d.ts +1 -1
- package/build/cjs/index.js +130 -0
- package/build/cjs/index.js.map +1 -1
- package/build/cjs/index.typings.d.ts +271 -1
- package/build/cjs/index.typings.js +122 -0
- package/build/cjs/index.typings.js.map +1 -1
- package/build/cjs/meta.d.ts +258 -2
- package/build/cjs/meta.js +109 -0
- package/build/cjs/meta.js.map +1 -1
- package/build/es/index.d.mts +1 -1
- package/build/es/index.mjs +126 -0
- package/build/es/index.mjs.map +1 -1
- package/build/es/index.typings.d.mts +271 -1
- package/build/es/index.typings.mjs +118 -0
- package/build/es/index.typings.mjs.map +1 -1
- package/build/es/meta.d.mts +258 -2
- package/build/es/meta.mjs +105 -0
- package/build/es/meta.mjs.map +1 -1
- package/build/internal/cjs/index.d.ts +13 -5
- package/build/internal/cjs/index.js +130 -0
- package/build/internal/cjs/index.js.map +1 -1
- package/build/internal/cjs/index.typings.d.ts +279 -3
- package/build/internal/cjs/index.typings.js +122 -0
- package/build/internal/cjs/index.typings.js.map +1 -1
- package/build/internal/cjs/meta.d.ts +258 -2
- package/build/internal/cjs/meta.js +109 -0
- package/build/internal/cjs/meta.js.map +1 -1
- package/build/internal/es/index.d.mts +13 -5
- package/build/internal/es/index.mjs +126 -0
- package/build/internal/es/index.mjs.map +1 -1
- package/build/internal/es/index.typings.d.mts +279 -3
- package/build/internal/es/index.typings.mjs +118 -0
- package/build/internal/es/index.typings.mjs.map +1 -1
- package/build/internal/es/meta.d.mts +258 -2
- package/build/internal/es/meta.mjs +105 -0
- package/build/internal/es/meta.mjs.map +1 -1
- package/package.json +2 -2
|
@@ -750,6 +750,70 @@ function generateContentByPromptObjectAsync(payload) {
|
|
|
750
750
|
}
|
|
751
751
|
return __generateContentByPromptObjectAsync;
|
|
752
752
|
}
|
|
753
|
+
function generateAudioStreamed(payload) {
|
|
754
|
+
function __generateAudioStreamed({ host }) {
|
|
755
|
+
const serializedData = transformPaths(payload, [
|
|
756
|
+
{
|
|
757
|
+
transformFn: transformSDKFloatToRESTFloat,
|
|
758
|
+
paths: [
|
|
759
|
+
{ path: "openAiCreateSpeechRequest.speed" },
|
|
760
|
+
{ path: "elevenlabsTextToSpeechRequest.voiceSettings.style" },
|
|
761
|
+
{ path: "elevenlabsTextToSpeechRequest.voiceSettings.stability" },
|
|
762
|
+
{
|
|
763
|
+
path: "elevenlabsTextToSpeechRequest.voiceSettings.similarityBoost"
|
|
764
|
+
}
|
|
765
|
+
]
|
|
766
|
+
}
|
|
767
|
+
]);
|
|
768
|
+
const metadata = {
|
|
769
|
+
entityFqdn: "wix.api_infra.v1.prompt_proxy",
|
|
770
|
+
method: "POST",
|
|
771
|
+
methodFqn: "wix.api_infra.v1.WixAiExternalGateway.GenerateAudioStreamed",
|
|
772
|
+
packageName: PACKAGE_NAME,
|
|
773
|
+
migrationOptions: {
|
|
774
|
+
optInTransformResponse: true
|
|
775
|
+
},
|
|
776
|
+
url: resolveWixApiInfraV1WixAiExternalGatewayUrl({
|
|
777
|
+
protoPath: "/v1/generate-audio-streamed",
|
|
778
|
+
data: serializedData,
|
|
779
|
+
host
|
|
780
|
+
}),
|
|
781
|
+
data: serializedData,
|
|
782
|
+
transformResponse: (payload2) => transformPaths(payload2, [
|
|
783
|
+
{
|
|
784
|
+
transformFn: transformRESTBytesToSDKBytes,
|
|
785
|
+
paths: [
|
|
786
|
+
{ path: "openAiSpeechChunk.content" },
|
|
787
|
+
{ path: "elevenlabsSpeechChunk.audioBase64" }
|
|
788
|
+
]
|
|
789
|
+
},
|
|
790
|
+
{
|
|
791
|
+
transformFn: transformRESTFloatToSDKFloat,
|
|
792
|
+
paths: [
|
|
793
|
+
{
|
|
794
|
+
path: "elevenlabsSpeechChunk.alignment.characterStartTimesSeconds",
|
|
795
|
+
isRepeated: true
|
|
796
|
+
},
|
|
797
|
+
{
|
|
798
|
+
path: "elevenlabsSpeechChunk.alignment.characterEndTimesSeconds",
|
|
799
|
+
isRepeated: true
|
|
800
|
+
},
|
|
801
|
+
{
|
|
802
|
+
path: "elevenlabsSpeechChunk.normalizedAlignment.characterStartTimesSeconds",
|
|
803
|
+
isRepeated: true
|
|
804
|
+
},
|
|
805
|
+
{
|
|
806
|
+
path: "elevenlabsSpeechChunk.normalizedAlignment.characterEndTimesSeconds",
|
|
807
|
+
isRepeated: true
|
|
808
|
+
}
|
|
809
|
+
]
|
|
810
|
+
}
|
|
811
|
+
])
|
|
812
|
+
};
|
|
813
|
+
return metadata;
|
|
814
|
+
}
|
|
815
|
+
return __generateAudioStreamed;
|
|
816
|
+
}
|
|
753
817
|
|
|
754
818
|
// src/api-infra-v1-prompt-proxy-prompts.universal.ts
|
|
755
819
|
var OpenaiproxyV1ChatCompletionMessageMessageRole = /* @__PURE__ */ ((OpenaiproxyV1ChatCompletionMessageMessageRole2) => {
|
|
@@ -1394,6 +1458,12 @@ var ResponsesMessageRole = /* @__PURE__ */ ((ResponsesMessageRole2) => {
|
|
|
1394
1458
|
ResponsesMessageRole2["DEVELOPER"] = "DEVELOPER";
|
|
1395
1459
|
return ResponsesMessageRole2;
|
|
1396
1460
|
})(ResponsesMessageRole || {});
|
|
1461
|
+
var V1VideoModel = /* @__PURE__ */ ((V1VideoModel2) => {
|
|
1462
|
+
V1VideoModel2["UNKNOWN_VIDEO_MODEL"] = "UNKNOWN_VIDEO_MODEL";
|
|
1463
|
+
V1VideoModel2["SORA_2"] = "SORA_2";
|
|
1464
|
+
V1VideoModel2["SORA_2_PRO"] = "SORA_2_PRO";
|
|
1465
|
+
return V1VideoModel2;
|
|
1466
|
+
})(V1VideoModel || {});
|
|
1397
1467
|
var GatewayMessageDefinitionRole = /* @__PURE__ */ ((GatewayMessageDefinitionRole2) => {
|
|
1398
1468
|
GatewayMessageDefinitionRole2["UNKNOWN"] = "UNKNOWN";
|
|
1399
1469
|
GatewayMessageDefinitionRole2["USER"] = "USER";
|
|
@@ -1403,6 +1473,19 @@ var GatewayMessageDefinitionRole = /* @__PURE__ */ ((GatewayMessageDefinitionRol
|
|
|
1403
1473
|
GatewayMessageDefinitionRole2["DEVELOPER"] = "DEVELOPER";
|
|
1404
1474
|
return GatewayMessageDefinitionRole2;
|
|
1405
1475
|
})(GatewayMessageDefinitionRole || {});
|
|
1476
|
+
var SpeechModel = /* @__PURE__ */ ((SpeechModel2) => {
|
|
1477
|
+
SpeechModel2["UNKNOWN_SPEECH_MODEL"] = "UNKNOWN_SPEECH_MODEL";
|
|
1478
|
+
SpeechModel2["TTS_1"] = "TTS_1";
|
|
1479
|
+
SpeechModel2["TTS_1_HD"] = "TTS_1_HD";
|
|
1480
|
+
return SpeechModel2;
|
|
1481
|
+
})(SpeechModel || {});
|
|
1482
|
+
var ElevenLabsTextToSpeechModel = /* @__PURE__ */ ((ElevenLabsTextToSpeechModel2) => {
|
|
1483
|
+
ElevenLabsTextToSpeechModel2["UNKNOWN_ELEVEN_LABS_TEXT_TO_SPEECH_MODEL"] = "UNKNOWN_ELEVEN_LABS_TEXT_TO_SPEECH_MODEL";
|
|
1484
|
+
ElevenLabsTextToSpeechModel2["ELEVEN_MULTILINGUAL_V2"] = "ELEVEN_MULTILINGUAL_V2";
|
|
1485
|
+
ElevenLabsTextToSpeechModel2["ELEVEN_FLASH_V2_5"] = "ELEVEN_FLASH_V2_5";
|
|
1486
|
+
ElevenLabsTextToSpeechModel2["ELEVEN_FLASH_V2"] = "ELEVEN_FLASH_V2";
|
|
1487
|
+
return ElevenLabsTextToSpeechModel2;
|
|
1488
|
+
})(ElevenLabsTextToSpeechModel || {});
|
|
1406
1489
|
var WebhookIdentityType = /* @__PURE__ */ ((WebhookIdentityType2) => {
|
|
1407
1490
|
WebhookIdentityType2["UNKNOWN"] = "UNKNOWN";
|
|
1408
1491
|
WebhookIdentityType2["ANONYMOUS_VISITOR"] = "ANONYMOUS_VISITOR";
|
|
@@ -1532,6 +1615,37 @@ async function generateContentByPromptObjectAsync2(options) {
|
|
|
1532
1615
|
throw transformedError;
|
|
1533
1616
|
}
|
|
1534
1617
|
}
|
|
1618
|
+
async function generateAudioStreamed2(options) {
|
|
1619
|
+
const { httpClient, sideEffects } = arguments[1];
|
|
1620
|
+
const payload = renameKeysFromSDKRequestToRESTRequest({
|
|
1621
|
+
openAiCreateSpeechRequest: options?.openAiCreateSpeechRequest,
|
|
1622
|
+
elevenlabsTextToSpeechRequest: options?.elevenlabsTextToSpeechRequest,
|
|
1623
|
+
userRequestInfo: options?.userRequestInfo
|
|
1624
|
+
});
|
|
1625
|
+
const reqOpts = generateAudioStreamed(payload);
|
|
1626
|
+
sideEffects?.onSiteCall?.();
|
|
1627
|
+
try {
|
|
1628
|
+
const result = await httpClient.request(reqOpts);
|
|
1629
|
+
sideEffects?.onSuccess?.(result);
|
|
1630
|
+
return renameKeysFromRESTResponseToSDKResponse(result.data);
|
|
1631
|
+
} catch (err) {
|
|
1632
|
+
const transformedError = sdkTransformError(
|
|
1633
|
+
err,
|
|
1634
|
+
{
|
|
1635
|
+
spreadPathsToArguments: {},
|
|
1636
|
+
explicitPathsToArguments: {
|
|
1637
|
+
openAiCreateSpeechRequest: "$[0].openAiCreateSpeechRequest",
|
|
1638
|
+
elevenlabsTextToSpeechRequest: "$[0].elevenlabsTextToSpeechRequest",
|
|
1639
|
+
userRequestInfo: "$[0].userRequestInfo"
|
|
1640
|
+
},
|
|
1641
|
+
singleArgumentUnchanged: false
|
|
1642
|
+
},
|
|
1643
|
+
["options"]
|
|
1644
|
+
);
|
|
1645
|
+
sideEffects?.onError?.(err);
|
|
1646
|
+
throw transformedError;
|
|
1647
|
+
}
|
|
1648
|
+
}
|
|
1535
1649
|
export {
|
|
1536
1650
|
AnthropicModel,
|
|
1537
1651
|
CacheControlType,
|
|
@@ -1544,6 +1658,7 @@ export {
|
|
|
1544
1658
|
CreatePredictionModel,
|
|
1545
1659
|
DynamicRetrievalConfigMode,
|
|
1546
1660
|
EditImageWithPromptRequestModel,
|
|
1661
|
+
ElevenLabsTextToSpeechModel,
|
|
1547
1662
|
FinishReason,
|
|
1548
1663
|
GatewayMessageDefinitionRole,
|
|
1549
1664
|
GenerateAnImageModel,
|
|
@@ -1587,6 +1702,7 @@ export {
|
|
|
1587
1702
|
ResponsesModel,
|
|
1588
1703
|
Role,
|
|
1589
1704
|
Sampler,
|
|
1705
|
+
SpeechModel,
|
|
1590
1706
|
StylePreset,
|
|
1591
1707
|
TextBisonModel,
|
|
1592
1708
|
TextToImageRequestModel,
|
|
@@ -1604,9 +1720,11 @@ export {
|
|
|
1604
1720
|
V1ResponseTypeType,
|
|
1605
1721
|
V1ResponsesModel,
|
|
1606
1722
|
V1ToolChoiceType,
|
|
1723
|
+
V1VideoModel,
|
|
1607
1724
|
VideoGenModel,
|
|
1608
1725
|
VideoModel,
|
|
1609
1726
|
WebhookIdentityType,
|
|
1727
|
+
generateAudioStreamed2 as generateAudioStreamed,
|
|
1610
1728
|
generateContentByPromptObject2 as generateContentByPromptObject,
|
|
1611
1729
|
generateContentByPromptObjectAsync2 as generateContentByPromptObjectAsync,
|
|
1612
1730
|
generateTextByPromptObjectStreamed2 as generateTextByPromptObjectStreamed
|