@llmops/gateway 0.2.8 → 0.2.10

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -10375,7 +10375,7 @@ const retryRequest = async (url, options, retryCount, statusCodesToRetry, timeou
10375
10375
 
10376
10376
  //#endregion
10377
10377
  //#region package.json
10378
- var version = "0.2.8";
10378
+ var version = "0.2.10";
10379
10379
 
10380
10380
  //#endregion
10381
10381
  //#region src/providers/bytez/api.ts
@@ -12219,7 +12219,10 @@ const AzureOpenAIChatCompleteConfig = {
12219
12219
  tool_choice: { param: "tool_choice" },
12220
12220
  response_format: { param: "response_format" },
12221
12221
  store: { param: "store" },
12222
- metadata: { param: "metadata" },
12222
+ metadata: {
12223
+ param: "metadata",
12224
+ transform: (params) => params.store ? params.metadata : void 0
12225
+ },
12223
12226
  modalities: { param: "modalities" },
12224
12227
  audio: { param: "audio" },
12225
12228
  seed: { param: "seed" },
@@ -12529,7 +12532,10 @@ const OpenAIChatCompleteConfig = {
12529
12532
  parallel_tool_calls: { param: "parallel_tool_calls" },
12530
12533
  max_completion_tokens: { param: "max_completion_tokens" },
12531
12534
  store: { param: "store" },
12532
- metadata: { param: "metadata" },
12535
+ metadata: {
12536
+ param: "metadata",
12537
+ transform: (params) => params.store ? params.metadata : void 0
12538
+ },
12533
12539
  modalities: { param: "modalities" },
12534
12540
  audio: { param: "audio" },
12535
12541
  prediction: { param: "prediction" },
@@ -26772,7 +26778,10 @@ const AzureAIInferenceChatCompleteConfig = {
26772
26778
  top_logprobs: { param: "top_logprobs" },
26773
26779
  logit_bias: { param: "logit_bias" },
26774
26780
  store: { param: "store" },
26775
- metadata: { param: "metadata" },
26781
+ metadata: {
26782
+ param: "metadata",
26783
+ transform: (params) => params.store ? params.metadata : void 0
26784
+ },
26776
26785
  modalities: { param: "modalities" },
26777
26786
  audio: { param: "audio" },
26778
26787
  seed: { param: "seed" },
@@ -31103,9 +31112,8 @@ var ResponseService = class {
31103
31112
  finalMappedResponse = response;
31104
31113
  originalResponseJSON = originalResponseJson;
31105
31114
  } else ({response: finalMappedResponse, originalResponseJson: originalResponseJSON, responseJson: responseJson} = await this.getResponse(response, responseTransformer, cache.isCacheHit));
31106
- this.updateHeaders(finalMappedResponse, cache.cacheStatus, retryAttempt);
31107
31115
  return {
31108
- response: finalMappedResponse,
31116
+ response: this.updateHeaders(finalMappedResponse, cache.cacheStatus, retryAttempt),
31109
31117
  responseJson,
31110
31118
  originalResponseJson: originalResponseJSON
31111
31119
  };
@@ -31115,17 +31123,22 @@ var ResponseService = class {
31115
31123
  return await responseHandler(this.context.honoContext, response, this.context.isStreaming, this.context.providerOption, responseTransformer, url, isCacheHit, this.context.params, this.context.strictOpenAiCompliance, this.context.honoContext.req.url, this.hooksService.areSyncHooksAvailable, this.hooksService.hookSpan?.id);
31116
31124
  }
31117
31125
  updateHeaders(response, cacheStatus, retryAttempt) {
31118
- response.headers.append(RESPONSE_HEADER_KEYS.LAST_USED_OPTION_INDEX, this.context.index.toString());
31119
- response.headers.append(RESPONSE_HEADER_KEYS.TRACE_ID, this.context.traceId);
31120
- response.headers.append(RESPONSE_HEADER_KEYS.RETRY_ATTEMPT_COUNT, retryAttempt.toString());
31121
- if (cacheStatus) response.headers.append(RESPONSE_HEADER_KEYS.CACHE_STATUS, cacheStatus);
31122
- if (this.context.provider && this.context.provider !== POWERED_BY) response.headers.append(HEADER_KEYS.PROVIDER, this.context.provider);
31126
+ const headers = new Headers(response.headers);
31127
+ headers.append(RESPONSE_HEADER_KEYS.LAST_USED_OPTION_INDEX, this.context.index.toString());
31128
+ headers.append(RESPONSE_HEADER_KEYS.TRACE_ID, this.context.traceId);
31129
+ headers.append(RESPONSE_HEADER_KEYS.RETRY_ATTEMPT_COUNT, retryAttempt.toString());
31130
+ if (cacheStatus) headers.append(RESPONSE_HEADER_KEYS.CACHE_STATUS, cacheStatus);
31131
+ if (this.context.provider && this.context.provider !== POWERED_BY) headers.append(HEADER_KEYS.PROVIDER, this.context.provider);
31123
31132
  if ((0, hono_adapter.getRuntimeKey)() == "node") {
31124
- response.headers.delete("content-encoding");
31125
- response.headers.delete("transfer-encoding");
31133
+ headers.delete("content-encoding");
31134
+ headers.delete("transfer-encoding");
31126
31135
  }
31127
- response.headers.delete("content-length");
31128
- return response;
31136
+ headers.delete("content-length");
31137
+ return new Response(response.body, {
31138
+ status: response.status,
31139
+ statusText: response.statusText,
31140
+ headers
31141
+ });
31129
31142
  }
31130
31143
  };
31131
31144
 
package/dist/index.mjs CHANGED
@@ -10348,7 +10348,7 @@ const retryRequest = async (url, options, retryCount, statusCodesToRetry, timeou
10348
10348
 
10349
10349
  //#endregion
10350
10350
  //#region package.json
10351
- var version = "0.2.8";
10351
+ var version = "0.2.10";
10352
10352
 
10353
10353
  //#endregion
10354
10354
  //#region src/providers/bytez/api.ts
@@ -12192,7 +12192,10 @@ const AzureOpenAIChatCompleteConfig = {
12192
12192
  tool_choice: { param: "tool_choice" },
12193
12193
  response_format: { param: "response_format" },
12194
12194
  store: { param: "store" },
12195
- metadata: { param: "metadata" },
12195
+ metadata: {
12196
+ param: "metadata",
12197
+ transform: (params) => params.store ? params.metadata : void 0
12198
+ },
12196
12199
  modalities: { param: "modalities" },
12197
12200
  audio: { param: "audio" },
12198
12201
  seed: { param: "seed" },
@@ -12502,7 +12505,10 @@ const OpenAIChatCompleteConfig = {
12502
12505
  parallel_tool_calls: { param: "parallel_tool_calls" },
12503
12506
  max_completion_tokens: { param: "max_completion_tokens" },
12504
12507
  store: { param: "store" },
12505
- metadata: { param: "metadata" },
12508
+ metadata: {
12509
+ param: "metadata",
12510
+ transform: (params) => params.store ? params.metadata : void 0
12511
+ },
12506
12512
  modalities: { param: "modalities" },
12507
12513
  audio: { param: "audio" },
12508
12514
  prediction: { param: "prediction" },
@@ -26745,7 +26751,10 @@ const AzureAIInferenceChatCompleteConfig = {
26745
26751
  top_logprobs: { param: "top_logprobs" },
26746
26752
  logit_bias: { param: "logit_bias" },
26747
26753
  store: { param: "store" },
26748
- metadata: { param: "metadata" },
26754
+ metadata: {
26755
+ param: "metadata",
26756
+ transform: (params) => params.store ? params.metadata : void 0
26757
+ },
26749
26758
  modalities: { param: "modalities" },
26750
26759
  audio: { param: "audio" },
26751
26760
  seed: { param: "seed" },
@@ -31076,9 +31085,8 @@ var ResponseService = class {
31076
31085
  finalMappedResponse = response;
31077
31086
  originalResponseJSON = originalResponseJson;
31078
31087
  } else ({response: finalMappedResponse, originalResponseJson: originalResponseJSON, responseJson: responseJson} = await this.getResponse(response, responseTransformer, cache.isCacheHit));
31079
- this.updateHeaders(finalMappedResponse, cache.cacheStatus, retryAttempt);
31080
31088
  return {
31081
- response: finalMappedResponse,
31089
+ response: this.updateHeaders(finalMappedResponse, cache.cacheStatus, retryAttempt),
31082
31090
  responseJson,
31083
31091
  originalResponseJson: originalResponseJSON
31084
31092
  };
@@ -31088,17 +31096,22 @@ var ResponseService = class {
31088
31096
  return await responseHandler(this.context.honoContext, response, this.context.isStreaming, this.context.providerOption, responseTransformer, url, isCacheHit, this.context.params, this.context.strictOpenAiCompliance, this.context.honoContext.req.url, this.hooksService.areSyncHooksAvailable, this.hooksService.hookSpan?.id);
31089
31097
  }
31090
31098
  updateHeaders(response, cacheStatus, retryAttempt) {
31091
- response.headers.append(RESPONSE_HEADER_KEYS.LAST_USED_OPTION_INDEX, this.context.index.toString());
31092
- response.headers.append(RESPONSE_HEADER_KEYS.TRACE_ID, this.context.traceId);
31093
- response.headers.append(RESPONSE_HEADER_KEYS.RETRY_ATTEMPT_COUNT, retryAttempt.toString());
31094
- if (cacheStatus) response.headers.append(RESPONSE_HEADER_KEYS.CACHE_STATUS, cacheStatus);
31095
- if (this.context.provider && this.context.provider !== POWERED_BY) response.headers.append(HEADER_KEYS.PROVIDER, this.context.provider);
31099
+ const headers = new Headers(response.headers);
31100
+ headers.append(RESPONSE_HEADER_KEYS.LAST_USED_OPTION_INDEX, this.context.index.toString());
31101
+ headers.append(RESPONSE_HEADER_KEYS.TRACE_ID, this.context.traceId);
31102
+ headers.append(RESPONSE_HEADER_KEYS.RETRY_ATTEMPT_COUNT, retryAttempt.toString());
31103
+ if (cacheStatus) headers.append(RESPONSE_HEADER_KEYS.CACHE_STATUS, cacheStatus);
31104
+ if (this.context.provider && this.context.provider !== POWERED_BY) headers.append(HEADER_KEYS.PROVIDER, this.context.provider);
31096
31105
  if (getRuntimeKey() == "node") {
31097
- response.headers.delete("content-encoding");
31098
- response.headers.delete("transfer-encoding");
31106
+ headers.delete("content-encoding");
31107
+ headers.delete("transfer-encoding");
31099
31108
  }
31100
- response.headers.delete("content-length");
31101
- return response;
31109
+ headers.delete("content-length");
31110
+ return new Response(response.body, {
31111
+ status: response.status,
31112
+ statusText: response.statusText,
31113
+ headers
31114
+ });
31102
31115
  }
31103
31116
  };
31104
31117
 
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@llmops/gateway",
3
- "version": "0.2.8",
3
+ "version": "0.2.10",
4
4
  "description": "AI gateway for LLMOps (forked from Portkey)",
5
5
  "type": "module",
6
6
  "license": "Apache-2.0",