@chatluna/v1-shared-adapter 1.0.15 → 1.0.17

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/lib/index.cjs CHANGED
@@ -75,7 +75,8 @@ function getModelMaxContextSize(info) {
75
75
  "gemini-1.0-pro": 30720,
76
76
  "gemini-2.0-flash": 1048576,
77
77
  "gemini-2.0-pro": 2097152,
78
- "gemini-2.5-pro": 2097152,
78
+ "gemini-2.5": 2097152,
79
+ "gemini-3.0-pro": 1097152,
79
80
  "gemini-2.0": 2097152,
80
81
  deepseek: 128e3,
81
82
  "llama3.1": 128e3,
@@ -443,6 +444,7 @@ __name(convertDeltaToMessageChunk, "convertDeltaToMessageChunk");
443
444
  var import_messages2 = require("@langchain/core/messages");
444
445
  var import_string2 = require("koishi-plugin-chatluna/utils/string");
445
446
  var import_logger = require("koishi-plugin-chatluna/utils/logger");
447
+ var import_object = require("koishi-plugin-chatluna/utils/object");
446
448
  async function buildChatCompletionParams(params, plugin, enableGoogleSearch, supportImageInput2) {
447
449
  const base = {
448
450
  model: params.model,
@@ -477,7 +479,7 @@ async function buildChatCompletionParams(params, plugin, enableGoogleSearch, sup
477
479
  delete base.n;
478
480
  delete base.top_p;
479
481
  }
480
- return base;
482
+ return (0, import_object.deepAssign)({}, base, params.overrideRequestParams ?? {});
481
483
  }
482
484
  __name(buildChatCompletionParams, "buildChatCompletionParams");
483
485
  function processReasoningContent(delta, reasoningState) {
@@ -514,15 +516,17 @@ async function* processStreamResponse(requestContext, iterator) {
514
516
  const choice = data.choices?.[0];
515
517
  if (data.usage) {
516
518
  yield new import_outputs.ChatGenerationChunk({
517
- message: new import_messages2.AIMessageChunk(""),
518
- text: "",
519
- generationInfo: {
520
- tokenUsage: {
521
- promptTokens: data.usage.prompt_tokens,
522
- completionTokens: data.usage.completion_tokens,
523
- totalTokens: data.usage.total_tokens
519
+ message: new import_messages2.AIMessageChunk({
520
+ content: "",
521
+ response_metadata: {
522
+ tokenUsage: {
523
+ promptTokens: data.usage.prompt_tokens,
524
+ completionTokens: data.usage.completion_tokens,
525
+ totalTokens: data.usage.total_tokens
526
+ }
524
527
  }
525
- }
528
+ }),
529
+ text: ""
526
530
  });
527
531
  }
528
532
  if (!choice) continue;
package/lib/index.mjs CHANGED
@@ -32,7 +32,8 @@ function getModelMaxContextSize(info) {
32
32
  "gemini-1.0-pro": 30720,
33
33
  "gemini-2.0-flash": 1048576,
34
34
  "gemini-2.0-pro": 2097152,
35
- "gemini-2.5-pro": 2097152,
35
+ "gemini-2.5": 2097152,
36
+ "gemini-3.0-pro": 1097152,
36
37
  "gemini-2.0": 2097152,
37
38
  deepseek: 128e3,
38
39
  "llama3.1": 128e3,
@@ -413,6 +414,7 @@ __name(convertDeltaToMessageChunk, "convertDeltaToMessageChunk");
413
414
  import { AIMessageChunk as AIMessageChunk2 } from "@langchain/core/messages";
414
415
  import { getMessageContent } from "koishi-plugin-chatluna/utils/string";
415
416
  import { trackLogToLocal } from "koishi-plugin-chatluna/utils/logger";
417
+ import { deepAssign } from "koishi-plugin-chatluna/utils/object";
416
418
  async function buildChatCompletionParams(params, plugin, enableGoogleSearch, supportImageInput2) {
417
419
  const base = {
418
420
  model: params.model,
@@ -447,7 +449,7 @@ async function buildChatCompletionParams(params, plugin, enableGoogleSearch, sup
447
449
  delete base.n;
448
450
  delete base.top_p;
449
451
  }
450
- return base;
452
+ return deepAssign({}, base, params.overrideRequestParams ?? {});
451
453
  }
452
454
  __name(buildChatCompletionParams, "buildChatCompletionParams");
453
455
  function processReasoningContent(delta, reasoningState) {
@@ -484,15 +486,17 @@ async function* processStreamResponse(requestContext, iterator) {
484
486
  const choice = data.choices?.[0];
485
487
  if (data.usage) {
486
488
  yield new ChatGenerationChunk({
487
- message: new AIMessageChunk2(""),
488
- text: "",
489
- generationInfo: {
490
- tokenUsage: {
491
- promptTokens: data.usage.prompt_tokens,
492
- completionTokens: data.usage.completion_tokens,
493
- totalTokens: data.usage.total_tokens
489
+ message: new AIMessageChunk2({
490
+ content: "",
491
+ response_metadata: {
492
+ tokenUsage: {
493
+ promptTokens: data.usage.prompt_tokens,
494
+ completionTokens: data.usage.completion_tokens,
495
+ totalTokens: data.usage.total_tokens
496
+ }
494
497
  }
495
- }
498
+ }),
499
+ text: ""
496
500
  });
497
501
  }
498
502
  if (!choice) continue;
@@ -30,7 +30,7 @@ export declare function buildChatCompletionParams(params: ModelRequestParams, pl
30
30
  stream_options: {
31
31
  include_usage: boolean;
32
32
  };
33
- }>;
33
+ } & Record<string, any>>;
34
34
  export declare function processReasoningContent(delta: {
35
35
  reasoning_content?: string;
36
36
  content?: string;
package/lib/utils.d.ts CHANGED
@@ -9,5 +9,5 @@ export declare function messageTypeToOpenAIRole(type: MessageType): ChatCompleti
9
9
  export declare function formatToolsToOpenAITools(tools: StructuredTool[], includeGoogleSearch: boolean): ChatCompletionTool[];
10
10
  export declare function formatToolToOpenAITool(tool: StructuredTool): ChatCompletionTool;
11
11
  export declare function removeAdditionalProperties(schema: JsonSchema7Type): JsonSchema7Type;
12
- export declare function convertMessageToMessageChunk(message: ChatCompletionResponseMessage): HumanMessageChunk | AIMessageChunk | SystemMessageChunk | FunctionMessageChunk | ToolMessageChunk | ChatMessageChunk;
13
- export declare function convertDeltaToMessageChunk(delta: Record<string, any>, defaultRole?: ChatCompletionResponseMessageRoleEnum): HumanMessageChunk | AIMessageChunk | SystemMessageChunk | FunctionMessageChunk | ToolMessageChunk | ChatMessageChunk;
12
+ export declare function convertMessageToMessageChunk(message: ChatCompletionResponseMessage): AIMessageChunk | HumanMessageChunk | SystemMessageChunk | FunctionMessageChunk | ToolMessageChunk | ChatMessageChunk;
13
+ export declare function convertDeltaToMessageChunk(delta: Record<string, any>, defaultRole?: ChatCompletionResponseMessageRoleEnum): AIMessageChunk | HumanMessageChunk | SystemMessageChunk | FunctionMessageChunk | ToolMessageChunk | ChatMessageChunk;
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "@chatluna/v1-shared-adapter",
3
3
  "description": "chatluna shared adapter",
4
- "version": "1.0.15",
4
+ "version": "1.0.17",
5
5
  "main": "lib/index.cjs",
6
6
  "module": "lib/index.mjs",
7
7
  "typings": "lib/index.d.ts",
@@ -70,6 +70,6 @@
70
70
  },
71
71
  "peerDependencies": {
72
72
  "koishi": "^4.18.9",
73
- "koishi-plugin-chatluna": "^1.3.0-alpha.76"
73
+ "koishi-plugin-chatluna": "^1.3.1"
74
74
  }
75
75
  }