@llmops/gateway 0.2.8 → 0.2.10
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +28 -15
- package/dist/index.mjs +28 -15
- package/package.json +1 -1
package/dist/index.cjs
CHANGED
|
@@ -10375,7 +10375,7 @@ const retryRequest = async (url, options, retryCount, statusCodesToRetry, timeou
|
|
|
10375
10375
|
|
|
10376
10376
|
//#endregion
|
|
10377
10377
|
//#region package.json
|
|
10378
|
-
var version = "0.2.
|
|
10378
|
+
var version = "0.2.10";
|
|
10379
10379
|
|
|
10380
10380
|
//#endregion
|
|
10381
10381
|
//#region src/providers/bytez/api.ts
|
|
@@ -12219,7 +12219,10 @@ const AzureOpenAIChatCompleteConfig = {
|
|
|
12219
12219
|
tool_choice: { param: "tool_choice" },
|
|
12220
12220
|
response_format: { param: "response_format" },
|
|
12221
12221
|
store: { param: "store" },
|
|
12222
|
-
metadata: {
|
|
12222
|
+
metadata: {
|
|
12223
|
+
param: "metadata",
|
|
12224
|
+
transform: (params) => params.store ? params.metadata : void 0
|
|
12225
|
+
},
|
|
12223
12226
|
modalities: { param: "modalities" },
|
|
12224
12227
|
audio: { param: "audio" },
|
|
12225
12228
|
seed: { param: "seed" },
|
|
@@ -12529,7 +12532,10 @@ const OpenAIChatCompleteConfig = {
|
|
|
12529
12532
|
parallel_tool_calls: { param: "parallel_tool_calls" },
|
|
12530
12533
|
max_completion_tokens: { param: "max_completion_tokens" },
|
|
12531
12534
|
store: { param: "store" },
|
|
12532
|
-
metadata: {
|
|
12535
|
+
metadata: {
|
|
12536
|
+
param: "metadata",
|
|
12537
|
+
transform: (params) => params.store ? params.metadata : void 0
|
|
12538
|
+
},
|
|
12533
12539
|
modalities: { param: "modalities" },
|
|
12534
12540
|
audio: { param: "audio" },
|
|
12535
12541
|
prediction: { param: "prediction" },
|
|
@@ -26772,7 +26778,10 @@ const AzureAIInferenceChatCompleteConfig = {
|
|
|
26772
26778
|
top_logprobs: { param: "top_logprobs" },
|
|
26773
26779
|
logit_bias: { param: "logit_bias" },
|
|
26774
26780
|
store: { param: "store" },
|
|
26775
|
-
metadata: {
|
|
26781
|
+
metadata: {
|
|
26782
|
+
param: "metadata",
|
|
26783
|
+
transform: (params) => params.store ? params.metadata : void 0
|
|
26784
|
+
},
|
|
26776
26785
|
modalities: { param: "modalities" },
|
|
26777
26786
|
audio: { param: "audio" },
|
|
26778
26787
|
seed: { param: "seed" },
|
|
@@ -31103,9 +31112,8 @@ var ResponseService = class {
|
|
|
31103
31112
|
finalMappedResponse = response;
|
|
31104
31113
|
originalResponseJSON = originalResponseJson;
|
|
31105
31114
|
} else ({response: finalMappedResponse, originalResponseJson: originalResponseJSON, responseJson: responseJson} = await this.getResponse(response, responseTransformer, cache.isCacheHit));
|
|
31106
|
-
this.updateHeaders(finalMappedResponse, cache.cacheStatus, retryAttempt);
|
|
31107
31115
|
return {
|
|
31108
|
-
response: finalMappedResponse,
|
|
31116
|
+
response: this.updateHeaders(finalMappedResponse, cache.cacheStatus, retryAttempt),
|
|
31109
31117
|
responseJson,
|
|
31110
31118
|
originalResponseJson: originalResponseJSON
|
|
31111
31119
|
};
|
|
@@ -31115,17 +31123,22 @@ var ResponseService = class {
|
|
|
31115
31123
|
return await responseHandler(this.context.honoContext, response, this.context.isStreaming, this.context.providerOption, responseTransformer, url, isCacheHit, this.context.params, this.context.strictOpenAiCompliance, this.context.honoContext.req.url, this.hooksService.areSyncHooksAvailable, this.hooksService.hookSpan?.id);
|
|
31116
31124
|
}
|
|
31117
31125
|
updateHeaders(response, cacheStatus, retryAttempt) {
|
|
31118
|
-
response.headers
|
|
31119
|
-
|
|
31120
|
-
|
|
31121
|
-
|
|
31122
|
-
if (
|
|
31126
|
+
const headers = new Headers(response.headers);
|
|
31127
|
+
headers.append(RESPONSE_HEADER_KEYS.LAST_USED_OPTION_INDEX, this.context.index.toString());
|
|
31128
|
+
headers.append(RESPONSE_HEADER_KEYS.TRACE_ID, this.context.traceId);
|
|
31129
|
+
headers.append(RESPONSE_HEADER_KEYS.RETRY_ATTEMPT_COUNT, retryAttempt.toString());
|
|
31130
|
+
if (cacheStatus) headers.append(RESPONSE_HEADER_KEYS.CACHE_STATUS, cacheStatus);
|
|
31131
|
+
if (this.context.provider && this.context.provider !== POWERED_BY) headers.append(HEADER_KEYS.PROVIDER, this.context.provider);
|
|
31123
31132
|
if ((0, hono_adapter.getRuntimeKey)() == "node") {
|
|
31124
|
-
|
|
31125
|
-
|
|
31133
|
+
headers.delete("content-encoding");
|
|
31134
|
+
headers.delete("transfer-encoding");
|
|
31126
31135
|
}
|
|
31127
|
-
|
|
31128
|
-
return response
|
|
31136
|
+
headers.delete("content-length");
|
|
31137
|
+
return new Response(response.body, {
|
|
31138
|
+
status: response.status,
|
|
31139
|
+
statusText: response.statusText,
|
|
31140
|
+
headers
|
|
31141
|
+
});
|
|
31129
31142
|
}
|
|
31130
31143
|
};
|
|
31131
31144
|
|
package/dist/index.mjs
CHANGED
|
@@ -10348,7 +10348,7 @@ const retryRequest = async (url, options, retryCount, statusCodesToRetry, timeou
|
|
|
10348
10348
|
|
|
10349
10349
|
//#endregion
|
|
10350
10350
|
//#region package.json
|
|
10351
|
-
var version = "0.2.
|
|
10351
|
+
var version = "0.2.10";
|
|
10352
10352
|
|
|
10353
10353
|
//#endregion
|
|
10354
10354
|
//#region src/providers/bytez/api.ts
|
|
@@ -12192,7 +12192,10 @@ const AzureOpenAIChatCompleteConfig = {
|
|
|
12192
12192
|
tool_choice: { param: "tool_choice" },
|
|
12193
12193
|
response_format: { param: "response_format" },
|
|
12194
12194
|
store: { param: "store" },
|
|
12195
|
-
metadata: {
|
|
12195
|
+
metadata: {
|
|
12196
|
+
param: "metadata",
|
|
12197
|
+
transform: (params) => params.store ? params.metadata : void 0
|
|
12198
|
+
},
|
|
12196
12199
|
modalities: { param: "modalities" },
|
|
12197
12200
|
audio: { param: "audio" },
|
|
12198
12201
|
seed: { param: "seed" },
|
|
@@ -12502,7 +12505,10 @@ const OpenAIChatCompleteConfig = {
|
|
|
12502
12505
|
parallel_tool_calls: { param: "parallel_tool_calls" },
|
|
12503
12506
|
max_completion_tokens: { param: "max_completion_tokens" },
|
|
12504
12507
|
store: { param: "store" },
|
|
12505
|
-
metadata: {
|
|
12508
|
+
metadata: {
|
|
12509
|
+
param: "metadata",
|
|
12510
|
+
transform: (params) => params.store ? params.metadata : void 0
|
|
12511
|
+
},
|
|
12506
12512
|
modalities: { param: "modalities" },
|
|
12507
12513
|
audio: { param: "audio" },
|
|
12508
12514
|
prediction: { param: "prediction" },
|
|
@@ -26745,7 +26751,10 @@ const AzureAIInferenceChatCompleteConfig = {
|
|
|
26745
26751
|
top_logprobs: { param: "top_logprobs" },
|
|
26746
26752
|
logit_bias: { param: "logit_bias" },
|
|
26747
26753
|
store: { param: "store" },
|
|
26748
|
-
metadata: {
|
|
26754
|
+
metadata: {
|
|
26755
|
+
param: "metadata",
|
|
26756
|
+
transform: (params) => params.store ? params.metadata : void 0
|
|
26757
|
+
},
|
|
26749
26758
|
modalities: { param: "modalities" },
|
|
26750
26759
|
audio: { param: "audio" },
|
|
26751
26760
|
seed: { param: "seed" },
|
|
@@ -31076,9 +31085,8 @@ var ResponseService = class {
|
|
|
31076
31085
|
finalMappedResponse = response;
|
|
31077
31086
|
originalResponseJSON = originalResponseJson;
|
|
31078
31087
|
} else ({response: finalMappedResponse, originalResponseJson: originalResponseJSON, responseJson: responseJson} = await this.getResponse(response, responseTransformer, cache.isCacheHit));
|
|
31079
|
-
this.updateHeaders(finalMappedResponse, cache.cacheStatus, retryAttempt);
|
|
31080
31088
|
return {
|
|
31081
|
-
response: finalMappedResponse,
|
|
31089
|
+
response: this.updateHeaders(finalMappedResponse, cache.cacheStatus, retryAttempt),
|
|
31082
31090
|
responseJson,
|
|
31083
31091
|
originalResponseJson: originalResponseJSON
|
|
31084
31092
|
};
|
|
@@ -31088,17 +31096,22 @@ var ResponseService = class {
|
|
|
31088
31096
|
return await responseHandler(this.context.honoContext, response, this.context.isStreaming, this.context.providerOption, responseTransformer, url, isCacheHit, this.context.params, this.context.strictOpenAiCompliance, this.context.honoContext.req.url, this.hooksService.areSyncHooksAvailable, this.hooksService.hookSpan?.id);
|
|
31089
31097
|
}
|
|
31090
31098
|
updateHeaders(response, cacheStatus, retryAttempt) {
|
|
31091
|
-
response.headers
|
|
31092
|
-
|
|
31093
|
-
|
|
31094
|
-
|
|
31095
|
-
if (
|
|
31099
|
+
const headers = new Headers(response.headers);
|
|
31100
|
+
headers.append(RESPONSE_HEADER_KEYS.LAST_USED_OPTION_INDEX, this.context.index.toString());
|
|
31101
|
+
headers.append(RESPONSE_HEADER_KEYS.TRACE_ID, this.context.traceId);
|
|
31102
|
+
headers.append(RESPONSE_HEADER_KEYS.RETRY_ATTEMPT_COUNT, retryAttempt.toString());
|
|
31103
|
+
if (cacheStatus) headers.append(RESPONSE_HEADER_KEYS.CACHE_STATUS, cacheStatus);
|
|
31104
|
+
if (this.context.provider && this.context.provider !== POWERED_BY) headers.append(HEADER_KEYS.PROVIDER, this.context.provider);
|
|
31096
31105
|
if (getRuntimeKey() == "node") {
|
|
31097
|
-
|
|
31098
|
-
|
|
31106
|
+
headers.delete("content-encoding");
|
|
31107
|
+
headers.delete("transfer-encoding");
|
|
31099
31108
|
}
|
|
31100
|
-
|
|
31101
|
-
return response
|
|
31109
|
+
headers.delete("content-length");
|
|
31110
|
+
return new Response(response.body, {
|
|
31111
|
+
status: response.status,
|
|
31112
|
+
statusText: response.statusText,
|
|
31113
|
+
headers
|
|
31114
|
+
});
|
|
31102
31115
|
}
|
|
31103
31116
|
};
|
|
31104
31117
|
|