llm-proxy 1.3.3 → 1.3.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.d.ts +1 -1
- package/dist/index.js +4 -4
- package/dist/index.js.map +1 -1
- package/dist/services/AwsBedrockAnthropicService.d.ts +2 -2
- package/dist/services/AwsBedrockAnthropicService.js +4 -4
- package/dist/services/AwsBedrockAnthropicService.js.map +1 -1
- package/dist/services/ClientService.d.ts +2 -2
- package/dist/services/OpenAIService.d.ts +2 -2
- package/dist/services/OpenAIService.js +4 -4
- package/dist/services/OpenAIService.js.map +1 -1
- package/package.json +2 -2
package/dist/index.d.ts
CHANGED
package/dist/index.js
CHANGED
|
@@ -54,7 +54,7 @@ const types_1 = require("./types");
|
|
|
54
54
|
// Main function for non-streaming requests
|
|
55
55
|
function generateLLMResponse(params) {
|
|
56
56
|
return __awaiter(this, void 0, void 0, function* () {
|
|
57
|
-
const { messages, model,
|
|
57
|
+
const { messages, model, max_tokens, temperature, credentials } = params;
|
|
58
58
|
// Step 1: Identify the provider based on the model
|
|
59
59
|
const provider = ProviderFinder_1.ProviderFinder.getProvider(model);
|
|
60
60
|
// Initialize the correct service based on the provider
|
|
@@ -79,7 +79,7 @@ function generateLLMResponse(params) {
|
|
|
79
79
|
const { adaptedMessages, systemPrompt } = InputFormatAdapter_1.InputFormatAdapter.adaptMessages(messages, provider);
|
|
80
80
|
// Step 3: Generate the completion
|
|
81
81
|
const response = yield service.generateCompletion(adaptedMessages, // TODO: fix this any
|
|
82
|
-
model,
|
|
82
|
+
model, max_tokens, temperature, systemPrompt);
|
|
83
83
|
// Step 4: Adapt the response if needed
|
|
84
84
|
return provider === types_1.Providers.OPENAI
|
|
85
85
|
? response
|
|
@@ -89,7 +89,7 @@ function generateLLMResponse(params) {
|
|
|
89
89
|
// Main function for streaming requests
|
|
90
90
|
function generateLLMStreamResponse(params) {
|
|
91
91
|
return __awaiter(this, void 0, void 0, function* () {
|
|
92
|
-
const { messages, model,
|
|
92
|
+
const { messages, model, max_tokens, temperature, credentials } = params;
|
|
93
93
|
// Step 1: Identify the provider based on the model
|
|
94
94
|
const provider = ProviderFinder_1.ProviderFinder.getProvider(model);
|
|
95
95
|
// Initialize the correct service based on the provider
|
|
@@ -114,7 +114,7 @@ function generateLLMStreamResponse(params) {
|
|
|
114
114
|
const { adaptedMessages, systemPrompt } = InputFormatAdapter_1.InputFormatAdapter.adaptMessages(messages, provider);
|
|
115
115
|
// Step 3: Generate the streaming completion
|
|
116
116
|
const stream = service.generateStreamCompletion(adaptedMessages, // TODO: Fix this any
|
|
117
|
-
model,
|
|
117
|
+
model, max_tokens, temperature, systemPrompt);
|
|
118
118
|
// Step 4: Create and return the async generator
|
|
119
119
|
function streamGenerator() {
|
|
120
120
|
return __asyncGenerator(this, arguments, function* streamGenerator_1() {
|
package/dist/index.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.js","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAuBA,kDAgDC;AAGD,8DAyDC;AAnID,gEAA6D;AAC7D,wEAAqE;AACrE,0EAAuE;AACvE,sFAAmF;AACnF,4DAAyD;AACzD,mCAA8D;AAiB9D,2CAA2C;AAC3C,SAAsB,mBAAmB,CACvC,MAAiC;;QAEjC,MAAM,EAAE,QAAQ,EAAE,KAAK,EAAE,
|
|
1
|
+
{"version":3,"file":"index.js","sourceRoot":"","sources":["../src/index.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAuBA,kDAgDC;AAGD,8DAyDC;AAnID,gEAA6D;AAC7D,wEAAqE;AACrE,0EAAuE;AACvE,sFAAmF;AACnF,4DAAyD;AACzD,mCAA8D;AAiB9D,2CAA2C;AAC3C,SAAsB,mBAAmB,CACvC,MAAiC;;QAEjC,MAAM,EAAE,QAAQ,EAAE,KAAK,EAAE,UAAU,EAAE,WAAW,EAAE,WAAW,EAAE,GAAG,MAAM,CAAC;QAEzE,mDAAmD;QACnD,MAAM,QAAQ,GAAG,+BAAc,CAAC,WAAW,CAAC,KAAK,CAAC,CAAC;QAEnD,uDAAuD;QACvD,IAAI,OAAmD,CAAC;QACxD,IAAI,QAAQ,KAAK,iBAAS,CAAC,MAAM,EAAE,CAAC;YAClC,IAAI,CAAC,WAAW,CAAC,MAAM,EAAE,CAAC;gBACxB,MAAM,IAAI,KAAK,CAAC,+CAA+C,CAAC,CAAC;YACnE,CAAC;YACD,OAAO,GAAG,IAAI,6BAAa,CAAC,WAAW,CAAC,MAAM,CAAC,CAAC;QAClD,CAAC;aAAM,IAAI,QAAQ,KAAK,iBAAS,CAAC,iBAAiB,EAAE,CAAC;YACpD,MAAM,SAAS,GAAG,WAAW,CAAC,SAAS,CAAC;YACxC,IAAI,CAAC,SAAS,EAAE,CAAC;gBACf,MAAM,IAAI,KAAK,CAAC,kDAAkD,CAAC,CAAC;YACtE,CAAC;YACD,OAAO,GAAG,IAAI,uDAA0B,CACtC,SAAS,CAAC,WAAW,EACrB,SAAS,CAAC,eAAe,EACzB,SAAS,CAAC,MAAM,CACjB,CAAC;QACJ,CAAC;aAAM,CAAC;YACN,MAAM,IAAI,KAAK,CAAC,sBAAsB,CAAC,CAAC;QAC1C,CAAC;QAED,uDAAuD;QACvD,MAAM,EAAE,eAAe,EAAE,YAAY,EAAE,GAAG,uCAAkB,CAAC,aAAa,CACxE,QAAQ,EACR,QAAQ,CACT,CAAC;QAEF,kCAAkC;QAClC,MAAM,QAAQ,GAAG,MAAM,OAAO,CAAC,kBAAkB,CAC/C,eAAsB,EAAE,qBAAqB;QAC7C,KAAK,EACL,UAAU,EACV,WAAW,EACX,YAAY,CACb,CAAC;QAEF,uCAAuC;QACvC,OAAO,QAAQ,KAAK,iBAAS,CAAC,MAAM;YAClC,CAAC,CAAE,QAA2B;YAC9B,CAAC,CAAE,yCAAmB,CAAC,aAAa,CAAC,QAAQ,EAAE,QAAQ,CAAoB,CAAC;IAChF,CAAC;CAAA;AAED,uCAAuC;AACvC,SAAsB,yBAAyB,CAC7C,MAAiC;;QAEjC,MAAM,EAAE,QAAQ,EAAE,KAAK,EAAE,UAAU,EAAE,WAAW,EAAE,WAAW,EAAE,GAAG,MAAM,CAAC;QAEzE,mDAAmD;QACnD,MAAM,QAAQ,GAAG,+BAAc,CAAC,WAAW,CAAC,KAAK,CAAC,CAAC;QAEnD,uDAAuD;QACvD,IAAI,OAAmD,CAAC;QACxD,IAAI,QAAQ,KAAK,iBAAS,CAAC,MAAM,EAAE,CAAC;YAClC,IAAI,CAAC,WAAW,CAAC,MAAM,EAAE,CAAC;gBACxB,MAAM,IAAI,KAAK,CAAC,+CAA+C,CAAC,CAAC;YACnE,CAAC;YACD,OAAO,GAAG,IAAI,6BAAa,CAAC,WAAW,CAAC,MAAM,CAAC,CAAC;QAClD,CAAC;aAAM,IAAI,QAAQ,KAAK,iBAAS,CAAC,iBAAiB,EAAE,CAAC;YACpD,MAAM,SAAS,GAAG,WAAW,CAAC,SAAS,CAAC;YACxC,IAAI,CAAC,SAAS,EAAE,CAAC;gBACf,MAAM,IAAI,KAAK,CAAC,kDAAkD,CAAC,CAAC;YACtE,CAAC;YACD,OAAO,GAAG,IAAI,uDAA0B,CACtC,SAAS,CAAC,WAAW,EACrB,SAAS,CAAC,eAAe,EACzB,SAAS,CAAC,MAAM,CACjB,CAAC;QACJ,CAAC;aAAM,CAAC;YACN,MAAM,IAAI,KAAK,CAAC,sBAAsB,CAAC,CAAC;QAC1C,CAAC;QAED,uDAAuD;QACvD,MAAM,EAAE,eAAe,EAAE,YAAY,EAAE,GAAG,uCAAkB,CAAC,aAAa,CACxE,QAAQ,EACR,QAAQ,CACT,CAAC;QAEF,4CAA4C;QAC5C,MAAM,MAAM,GAAG,OAAO,CAAC,wBAAwB,CAC7C,eAAsB,EAAE,qBAAqB;QAC7C,KAAK,EACL,UAAU,EACV,WAAW,EACX,YAAY,CACb,CAAC;QAEF,gDAAgD;QAChD,SAAgB,eAAe;;;;oBAC7B,KAA0B,eAAA,WAAA,cAAA,MAAM,CAAA,YAAA,qFAAE,CAAC;wBAAT,sBAAM;wBAAN,WAAM;wBAArB,MAAM,KAAK,KAAA,CAAA;wBACpB,oBAAM,QAAQ,KAAK,iBAAS,CAAC,MAAM;4BACjC,CAAC,CAAE,KAAwB;4BAC3B,CAAC,CAAE,yCAAmB,CAAC,aAAa,CAChC,KAAK,EACL,QAAQ,CACU,CAAA,CAAC;oBAC3B,CAAC;;;;;;;;;YACH,CAAC;SAAA;QAED,OAAO,eAAe,EAAE,CAAC;IAC3B,CAAC;CAAA;AAED,0CAAwB"}
|
|
@@ -3,6 +3,6 @@ import { ClientService } from "./ClientService";
|
|
|
3
3
|
export declare class AwsBedrockAnthropicService implements ClientService {
|
|
4
4
|
private bedrock;
|
|
5
5
|
constructor(awsAccessKey: string, awsSecretKey: string, region: string);
|
|
6
|
-
generateCompletion(messages: Messages, model?: string,
|
|
7
|
-
generateStreamCompletion(messages: Messages, model?: string,
|
|
6
|
+
generateCompletion(messages: Messages, model?: string, max_tokens?: number, temperature?: number, systemPrompt?: string): Promise<BedrockAnthropicResponse>;
|
|
7
|
+
generateStreamCompletion(messages: Messages, model?: string, max_tokens?: number, temperature?: number, systemPrompt?: string): AsyncGenerator<BedrockAnthropicParsedChunk, void, unknown>;
|
|
8
8
|
}
|
|
@@ -41,14 +41,14 @@ class AwsBedrockAnthropicService {
|
|
|
41
41
|
},
|
|
42
42
|
});
|
|
43
43
|
}
|
|
44
|
-
generateCompletion(messages, model,
|
|
44
|
+
generateCompletion(messages, model, max_tokens, temperature, systemPrompt) {
|
|
45
45
|
return __awaiter(this, void 0, void 0, function* () {
|
|
46
46
|
if (!model) {
|
|
47
47
|
throw new Error("Model ID is required for AwsBedrockAnthropicService");
|
|
48
48
|
}
|
|
49
49
|
const body = JSON.stringify({
|
|
50
50
|
anthropic_version: "bedrock-2023-05-31",
|
|
51
|
-
max_tokens
|
|
51
|
+
max_tokens,
|
|
52
52
|
temperature,
|
|
53
53
|
messages,
|
|
54
54
|
system: systemPrompt,
|
|
@@ -63,7 +63,7 @@ class AwsBedrockAnthropicService {
|
|
|
63
63
|
return JSON.parse(new TextDecoder().decode(response.body));
|
|
64
64
|
});
|
|
65
65
|
}
|
|
66
|
-
generateStreamCompletion(messages, model,
|
|
66
|
+
generateStreamCompletion(messages, model, max_tokens, temperature, systemPrompt) {
|
|
67
67
|
return __asyncGenerator(this, arguments, function* generateStreamCompletion_1() {
|
|
68
68
|
var _a, e_1, _b, _c;
|
|
69
69
|
var _d;
|
|
@@ -72,7 +72,7 @@ class AwsBedrockAnthropicService {
|
|
|
72
72
|
}
|
|
73
73
|
const body = JSON.stringify({
|
|
74
74
|
anthropic_version: "bedrock-2023-05-31",
|
|
75
|
-
max_tokens
|
|
75
|
+
max_tokens,
|
|
76
76
|
temperature,
|
|
77
77
|
messages,
|
|
78
78
|
system: systemPrompt,
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"AwsBedrockAnthropicService.js","sourceRoot":"","sources":["../../src/services/AwsBedrockAnthropicService.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAKA,4EAIyC;AAGzC,MAAa,0BAA0B;IAGrC,YAAY,YAAoB,EAAE,YAAoB,EAAE,MAAc;QACpE,IAAI,CAAC,OAAO,GAAG,IAAI,6CAAoB,CAAC;YACtC,MAAM;YACN,WAAW,EAAE;gBACX,WAAW,EAAE,YAAY;gBACzB,eAAe,EAAE,YAAY;aAC9B;SACF,CAAC,CAAC;IACL,CAAC;IAEK,kBAAkB,CACtB,QAAkB,EAClB,KAAc,EACd,
|
|
1
|
+
{"version":3,"file":"AwsBedrockAnthropicService.js","sourceRoot":"","sources":["../../src/services/AwsBedrockAnthropicService.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAKA,4EAIyC;AAGzC,MAAa,0BAA0B;IAGrC,YAAY,YAAoB,EAAE,YAAoB,EAAE,MAAc;QACpE,IAAI,CAAC,OAAO,GAAG,IAAI,6CAAoB,CAAC;YACtC,MAAM;YACN,WAAW,EAAE;gBACX,WAAW,EAAE,YAAY;gBACzB,eAAe,EAAE,YAAY;aAC9B;SACF,CAAC,CAAC;IACL,CAAC;IAEK,kBAAkB,CACtB,QAAkB,EAClB,KAAc,EACd,UAAmB,EACnB,WAAoB,EACpB,YAAqB;;YAErB,IAAI,CAAC,KAAK,EAAE,CAAC;gBACX,MAAM,IAAI,KAAK,CAAC,qDAAqD,CAAC,CAAC;YACzE,CAAC;YAED,MAAM,IAAI,GAAG,IAAI,CAAC,SAAS,CAAC;gBAC1B,iBAAiB,EAAE,oBAAoB;gBACvC,UAAU;gBACV,WAAW;gBACX,QAAQ;gBACR,MAAM,EAAE,YAAY;aACrB,CAAC,CAAC;YAEH,MAAM,OAAO,GAAG,IAAI,2CAAkB,CAAC;gBACrC,OAAO,EAAE,KAAK;gBACd,IAAI;gBACJ,WAAW,EAAE,kBAAkB;gBAC/B,MAAM,EAAE,kBAAkB;aAC3B,CAAC,CAAC;YAEH,MAAM,QAAQ,GAAG,MAAM,IAAI,CAAC,OAAO,CAAC,IAAI,CAAC,OAAO,CAAC,CAAC;YAClD,OAAO,IAAI,CAAC,KAAK,CAAC,IAAI,WAAW,EAAE,CAAC,MAAM,CAAC,QAAQ,CAAC,IAAI,CAAC,CAAC,CAAC;QAC7D,CAAC;KAAA;IAEM,wBAAwB,CAC7B,QAAkB,EAClB,KAAc,EACd,UAAmB,EACnB,WAAoB,EACpB,YAAqB;;;;YAErB,IAAI,CAAC,KAAK,EAAE,CAAC;gBACX,MAAM,IAAI,KAAK,CAAC,qDAAqD,CAAC,CAAC;YACzE,CAAC;YAED,MAAM,IAAI,GAAG,IAAI,CAAC,SAAS,CAAC;gBAC1B,iBAAiB,EAAE,oBAAoB;gBACvC,UAAU;gBACV,WAAW;gBACX,QAAQ;gBACR,MAAM,EAAE,YAAY;aACrB,CAAC,CAAC;YAEH,MAAM,OAAO,GAAG,IAAI,6DAAoC,CAAC;gBACvD,OAAO,EAAE,KAAK;gBACd,IAAI;gBACJ,WAAW,EAAE,kBAAkB;gBAC/B,MAAM,EAAE,kBAAkB;aAC3B,CAAC,CAAC;YAEH,MAAM,QAAQ,GAAG,cAAM,IAAI,CAAC,OAAO,CAAC,IAAI,CAAC,OAAO,CAAC,CAAA,CAAC;YAElD,IAAI,QAAQ,CAAC,IAAI,EAAE,CAAC;gBAClB,MAAM,OAAO,GAAG,IAAI,WAAW,CAAC,OAAO,CAAC,CAAC;;oBAEzC,KAA4B,eAAA,KAAA,cAAA,QAAQ,CAAC,IAAI,CAAA,IAAA,+DAAE,CAAC;wBAAhB,cAAa;wBAAb,WAAa;wBAA9B,MAAM,OAAO,KAAA,CAAA;wBACtB,MAAM,aAAa,GAAG,OAAO,CAAC,MAAM,CAAC,MAAA,OAAO,CAAC,KAAK,0CAAE,KAAK,EAAE;4BACzD,MAAM,EAAE,IAAI;yBACb,CAAC,CAAC;wBAEH,IAAI,CAAC;4BACH,MAAM,UAAU,GAAG,IAAI,CAAC,KAAK,CAAC,aAAa,CAAC,CAAC;4BAC7C,oBAAM,UAAU,CAAA,CAAC;wBACnB,CAAC;wBAAC,OAAO,KAAK,EAAE,CAAC;4BACf,OAAO,CAAC,KAAK,CAAC,gCAAgC,EAAE,KAAK,CAAC,CAAC;wBACzD,CAAC;oBACH,CAAC;;;;;;;;;YACH,CAAC;QACH,CAAC;KAAA;CACF;AAxFD,gEAwFC"}
|
|
@@ -1,5 +1,5 @@
|
|
|
1
1
|
import { BedrockAnthropicParsedChunk, LLMResponse, Messages } from "../types";
|
|
2
2
|
export interface ClientService {
|
|
3
|
-
generateCompletion(messages: Messages, model?: string,
|
|
4
|
-
generateStreamCompletion(messages: Messages, model?: string,
|
|
3
|
+
generateCompletion(messages: Messages, model?: string, max_tokens?: number, temperature?: number, systemPrompt?: string): Promise<LLMResponse>;
|
|
4
|
+
generateStreamCompletion(messages: Messages, model?: string, max_tokens?: number, temperature?: number, systemPrompt?: string): AsyncGenerator<BedrockAnthropicParsedChunk, void, unknown>;
|
|
5
5
|
}
|
|
@@ -3,6 +3,6 @@ import { ClientService } from "./ClientService";
|
|
|
3
3
|
export declare class OpenAIService implements ClientService {
|
|
4
4
|
private openai;
|
|
5
5
|
constructor(apiKey: string);
|
|
6
|
-
generateCompletion(messages: OpenAIMessages, model: string,
|
|
7
|
-
generateStreamCompletion(messages: OpenAIMessages, model: string,
|
|
6
|
+
generateCompletion(messages: OpenAIMessages, model: string, max_tokens: number, temperature: number, systemPrompt?: string): Promise<OpenAIResponse>;
|
|
7
|
+
generateStreamCompletion(messages: OpenAIMessages, model: string, max_tokens: number, temperature: number, systemPrompt?: string): AsyncGenerator<any, void, unknown>;
|
|
8
8
|
}
|
|
@@ -38,7 +38,7 @@ class OpenAIService {
|
|
|
38
38
|
constructor(apiKey) {
|
|
39
39
|
this.openai = new openai_1.default({ apiKey });
|
|
40
40
|
}
|
|
41
|
-
generateCompletion(messages, model,
|
|
41
|
+
generateCompletion(messages, model, max_tokens, temperature, systemPrompt) {
|
|
42
42
|
return __awaiter(this, void 0, void 0, function* () {
|
|
43
43
|
if (!model) {
|
|
44
44
|
throw new Error("Model ID is required for OpenAIService.");
|
|
@@ -47,7 +47,7 @@ class OpenAIService {
|
|
|
47
47
|
const response = yield this.openai.chat.completions.create({
|
|
48
48
|
model, // Use the string directly
|
|
49
49
|
messages,
|
|
50
|
-
max_tokens
|
|
50
|
+
max_tokens,
|
|
51
51
|
temperature,
|
|
52
52
|
});
|
|
53
53
|
return response;
|
|
@@ -58,7 +58,7 @@ class OpenAIService {
|
|
|
58
58
|
}
|
|
59
59
|
});
|
|
60
60
|
}
|
|
61
|
-
generateStreamCompletion(messages, model,
|
|
61
|
+
generateStreamCompletion(messages, model, max_tokens, temperature, systemPrompt) {
|
|
62
62
|
return __asyncGenerator(this, arguments, function* generateStreamCompletion_1() {
|
|
63
63
|
var _a, e_1, _b, _c;
|
|
64
64
|
if (!model) {
|
|
@@ -68,7 +68,7 @@ class OpenAIService {
|
|
|
68
68
|
const stream = yield __await(this.openai.chat.completions.create({
|
|
69
69
|
model,
|
|
70
70
|
messages,
|
|
71
|
-
max_tokens
|
|
71
|
+
max_tokens,
|
|
72
72
|
temperature,
|
|
73
73
|
stream: true,
|
|
74
74
|
stream_options: {
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"OpenAIService.js","sourceRoot":"","sources":["../../src/services/OpenAIService.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA,oDAA4B;AAI5B,MAAa,aAAa;IAGxB,YAAY,MAAc;QACxB,IAAI,CAAC,MAAM,GAAG,IAAI,gBAAM,CAAC,EAAE,MAAM,EAAE,CAAC,CAAC;IACvC,CAAC;IAEK,kBAAkB,CACtB,QAAwB,EACxB,KAAa,EACb,
|
|
1
|
+
{"version":3,"file":"OpenAIService.js","sourceRoot":"","sources":["../../src/services/OpenAIService.ts"],"names":[],"mappings":";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAAA,oDAA4B;AAI5B,MAAa,aAAa;IAGxB,YAAY,MAAc;QACxB,IAAI,CAAC,MAAM,GAAG,IAAI,gBAAM,CAAC,EAAE,MAAM,EAAE,CAAC,CAAC;IACvC,CAAC;IAEK,kBAAkB,CACtB,QAAwB,EACxB,KAAa,EACb,UAAkB,EAClB,WAAmB,EACnB,YAAqB;;YAErB,IAAI,CAAC,KAAK,EAAE,CAAC;gBACX,MAAM,IAAI,KAAK,CAAC,yCAAyC,CAAC,CAAC;YAC7D,CAAC;YAED,IAAI,CAAC;gBACH,MAAM,QAAQ,GAAG,MAAM,IAAI,CAAC,MAAM,CAAC,IAAI,CAAC,WAAW,CAAC,MAAM,CAAC;oBACzD,KAAK,EAAE,0BAA0B;oBACjC,QAAQ;oBACR,UAAU;oBACV,WAAW;iBACZ,CAAC,CAAC;gBACH,OAAO,QAA0B,CAAC;YACpC,CAAC;YAAC,OAAO,KAAK,EAAE,CAAC;gBACf,OAAO,CAAC,KAAK,CAAC,wBAAwB,EAAE,KAAK,CAAC,CAAC;gBAC/C,MAAM,KAAK,CAAC;YACd,CAAC;QACH,CAAC;KAAA;IAEM,wBAAwB,CAC7B,QAAwB,EACxB,KAAa,EACb,UAAkB,EAClB,WAAmB,EACnB,YAAqB;;;YAErB,IAAI,CAAC,KAAK,EAAE,CAAC;gBACX,MAAM,IAAI,KAAK,CAAC,yCAAyC,CAAC,CAAC;YAC7D,CAAC;YAED,IAAI,CAAC;gBACH,MAAM,MAAM,GAAG,cAAM,IAAI,CAAC,MAAM,CAAC,IAAI,CAAC,WAAW,CAAC,MAAM,CAAC;oBACvD,KAAK;oBACL,QAAQ;oBACR,UAAU;oBACV,WAAW;oBACX,MAAM,EAAE,IAAI;oBACZ,cAAc,EAAE;wBACd,aAAa,EAAE,IAAI;qBACpB;iBACF,CAAC,CAAA,CAAC;;oBAEH,KAA0B,eAAA,WAAA,cAAA,MAAM,CAAA,YAAA,qFAAE,CAAC;wBAAT,sBAAM;wBAAN,WAAM;wBAArB,MAAM,KAAK,KAAA,CAAA;wBACpB,oBAAM,KAAK,CAAA,CAAC;oBACd,CAAC;;;;;;;;;YACH,CAAC;YAAC,OAAO,KAAK,EAAE,CAAC;gBACf,OAAO,CAAC,KAAK,CAAC,6BAA6B,EAAE,KAAK,CAAC,CAAC;gBACpD,MAAM,KAAK,CAAC;YACd,CAAC;QACH,CAAC;KAAA;CACF;AA/DD,sCA+DC"}
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "llm-proxy",
|
|
3
|
-
"version": "1.3.
|
|
3
|
+
"version": "1.3.4",
|
|
4
4
|
"description": "An LLM Proxy that allows the user to interact with different language models from different providers using unified request and response formats.",
|
|
5
5
|
"main": "dist/index.js",
|
|
6
6
|
"types": "dist/index.d.ts",
|
|
@@ -30,7 +30,7 @@
|
|
|
30
30
|
"aws-sdk": "^2.1691.0",
|
|
31
31
|
"axios": "^1.7.7",
|
|
32
32
|
"dotenv": "^16.4.5",
|
|
33
|
-
"llm-proxy": "^1.3.
|
|
33
|
+
"llm-proxy": "^1.3.4",
|
|
34
34
|
"openai": "^4.69.0"
|
|
35
35
|
}
|
|
36
36
|
}
|