@aigne/aigne-hub 0.4.8 → 0.5.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (41) hide show
  1. package/CHANGELOG.md +62 -0
  2. package/lib/cjs/blocklet-aigne-hub-model.js +4 -3
  3. package/lib/cjs/cli-aigne-hub-model.js +1 -1
  4. package/lib/cjs/index.d.ts +5 -1
  5. package/lib/cjs/index.js +8 -3
  6. package/lib/cjs/util/constants.d.ts +13 -0
  7. package/lib/cjs/util/constants.js +29 -0
  8. package/lib/cjs/util/credential.d.ts +24 -0
  9. package/lib/cjs/util/credential.js +223 -0
  10. package/lib/cjs/util/crypto.d.ts +4 -0
  11. package/lib/cjs/util/crypto.js +19 -0
  12. package/lib/cjs/util/model.d.ts +12 -0
  13. package/lib/cjs/util/model.js +201 -0
  14. package/lib/cjs/util/type.d.ts +58 -0
  15. package/lib/cjs/util/type.js +2 -0
  16. package/lib/dts/index.d.ts +5 -1
  17. package/lib/dts/util/constants.d.ts +13 -0
  18. package/lib/dts/util/credential.d.ts +24 -0
  19. package/lib/dts/util/crypto.d.ts +4 -0
  20. package/lib/dts/util/model.d.ts +12 -0
  21. package/lib/dts/util/type.d.ts +58 -0
  22. package/lib/esm/blocklet-aigne-hub-model.js +2 -1
  23. package/lib/esm/cli-aigne-hub-model.js +1 -1
  24. package/lib/esm/index.d.ts +5 -1
  25. package/lib/esm/index.js +7 -2
  26. package/lib/esm/util/constants.d.ts +13 -0
  27. package/lib/esm/util/constants.js +15 -0
  28. package/lib/esm/util/credential.d.ts +24 -0
  29. package/lib/esm/util/credential.js +211 -0
  30. package/lib/esm/util/crypto.d.ts +4 -0
  31. package/lib/esm/util/crypto.js +9 -0
  32. package/lib/esm/util/model.d.ts +12 -0
  33. package/lib/esm/util/model.js +189 -0
  34. package/lib/esm/util/type.d.ts +58 -0
  35. package/lib/esm/util/type.js +1 -0
  36. package/package.json +22 -15
  37. package/lib/cjs/constants.d.ts +0 -28
  38. package/lib/cjs/constants.js +0 -123
  39. package/lib/dts/constants.d.ts +0 -28
  40. package/lib/esm/constants.d.ts +0 -28
  41. package/lib/esm/constants.js +0 -116
@@ -1,123 +0,0 @@
1
- "use strict";
2
- Object.defineProperty(exports, "__esModule", { value: true });
3
- exports.AIGNE_HUB_URL = void 0;
4
- exports.availableModels = availableModels;
5
- exports.findModel = findModel;
6
- exports.loadModel = loadModel;
7
- exports.getAIGNEHubMountPoint = getAIGNEHubMountPoint;
8
- const anthropic_1 = require("@aigne/anthropic");
9
- const bedrock_1 = require("@aigne/bedrock");
10
- const deepseek_1 = require("@aigne/deepseek");
11
- const gemini_1 = require("@aigne/gemini");
12
- const ollama_1 = require("@aigne/ollama");
13
- const open_router_1 = require("@aigne/open-router");
14
- const openai_1 = require("@aigne/openai");
15
- const index_js_1 = require("@aigne/platform-helpers/nodejs/index.js");
16
- const xai_1 = require("@aigne/xai");
17
- const node_http_handler_1 = require("@smithy/node-http-handler");
18
- const https_proxy_agent_1 = require("https-proxy-agent");
19
- const ufo_1 = require("ufo");
20
- const cli_aigne_hub_model_js_1 = require("./cli-aigne-hub-model.js");
21
- const AIGNE_HUB_DID = "z8ia3xzq2tMq8CRHfaXj1BTYJyYnEcHbqP8cJ";
22
- exports.AIGNE_HUB_URL = "https://hub.aigne.io/";
23
- function availableModels() {
24
- const proxy = ["HTTPS_PROXY", "https_proxy", "HTTP_PROXY", "http_proxy", "ALL_PROXY", "all_proxy"]
25
- .map((i) => process.env[i])
26
- .filter(Boolean)[0];
27
- const httpAgent = proxy ? new https_proxy_agent_1.HttpsProxyAgent(proxy) : undefined;
28
- const clientOptions = {
29
- fetchOptions: {
30
- // @ts-ignore
31
- agent: httpAgent,
32
- },
33
- };
34
- return [
35
- {
36
- name: openai_1.OpenAIChatModel.name,
37
- apiKeyEnvName: "OPENAI_API_KEY",
38
- create: (params) => new openai_1.OpenAIChatModel({ ...params, clientOptions }),
39
- },
40
- {
41
- name: anthropic_1.AnthropicChatModel.name,
42
- apiKeyEnvName: "ANTHROPIC_API_KEY",
43
- create: (params) => new anthropic_1.AnthropicChatModel({ ...params, clientOptions }),
44
- },
45
- {
46
- name: bedrock_1.BedrockChatModel.name,
47
- apiKeyEnvName: "AWS_ACCESS_KEY_ID",
48
- create: (params) => new bedrock_1.BedrockChatModel({
49
- ...params,
50
- clientOptions: {
51
- requestHandler: node_http_handler_1.NodeHttpHandler.create({ httpAgent, httpsAgent: httpAgent }),
52
- streamCollector: node_http_handler_1.streamCollector,
53
- },
54
- }),
55
- },
56
- {
57
- name: deepseek_1.DeepSeekChatModel.name,
58
- apiKeyEnvName: "DEEPSEEK_API_KEY",
59
- create: (params) => new deepseek_1.DeepSeekChatModel({ ...params, clientOptions }),
60
- },
61
- {
62
- name: [gemini_1.GeminiChatModel.name, "google"],
63
- apiKeyEnvName: ["GEMINI_API_KEY", "GOOGLE_API_KEY"],
64
- create: (params) => new gemini_1.GeminiChatModel({ ...params, clientOptions }),
65
- },
66
- {
67
- name: ollama_1.OllamaChatModel.name,
68
- apiKeyEnvName: "OLLAMA_API_KEY",
69
- create: (params) => new ollama_1.OllamaChatModel({ ...params, clientOptions }),
70
- },
71
- {
72
- name: open_router_1.OpenRouterChatModel.name,
73
- apiKeyEnvName: "OPEN_ROUTER_API_KEY",
74
- create: (params) => new open_router_1.OpenRouterChatModel({ ...params, clientOptions }),
75
- },
76
- {
77
- name: xai_1.XAIChatModel.name,
78
- apiKeyEnvName: "XAI_API_KEY",
79
- create: (params) => new xai_1.XAIChatModel({ ...params, clientOptions }),
80
- },
81
- {
82
- name: cli_aigne_hub_model_js_1.CliAIGNEHubChatModel.name,
83
- apiKeyEnvName: "AIGNE_HUB_API_KEY",
84
- create: (params) => new cli_aigne_hub_model_js_1.CliAIGNEHubChatModel({ ...params, clientOptions }),
85
- },
86
- ];
87
- }
88
- function findModel(models, provider) {
89
- return models.find((m) => {
90
- if (typeof m.name === "string") {
91
- return m.name.toLowerCase().includes(provider.toLowerCase());
92
- }
93
- return m.name.some((n) => n.toLowerCase().includes(provider.toLowerCase()));
94
- });
95
- }
96
- const { MODEL_PROVIDER, MODEL_NAME } = index_js_1.nodejs.env;
97
- const DEFAULT_MODEL_PROVIDER = "openai";
98
- async function loadModel(model, modelOptions, credential) {
99
- const params = {
100
- model: MODEL_NAME ?? model?.name ?? undefined,
101
- temperature: model?.temperature ?? undefined,
102
- topP: model?.topP ?? undefined,
103
- frequencyPenalty: model?.frequencyPenalty ?? undefined,
104
- presencePenalty: model?.presencePenalty ?? undefined,
105
- };
106
- const provider = (MODEL_PROVIDER ?? model?.provider ?? DEFAULT_MODEL_PROVIDER).replace(/-/g, "");
107
- const m = findModel(availableModels(), provider);
108
- if (!m)
109
- throw new Error(`Unsupported model: ${model?.provider} ${model?.name}`);
110
- return m.create({
111
- ...(credential || {}),
112
- model: params.model,
113
- modelOptions: { ...params, ...modelOptions },
114
- });
115
- }
116
- async function getAIGNEHubMountPoint(url) {
117
- const { origin } = new URL(url);
118
- const BLOCKLET_JSON_PATH = "__blocklet__.js?type=json";
119
- const blockletInfo = await fetch((0, ufo_1.joinURL)(origin, BLOCKLET_JSON_PATH));
120
- const blocklet = await blockletInfo.json();
121
- const aigneHubMount = (blocklet?.componentMountPoints || []).find((m) => m.did === AIGNE_HUB_DID);
122
- return (0, ufo_1.joinURL)(origin, aigneHubMount?.mountPoint || "");
123
- }
@@ -1,28 +0,0 @@
1
- import type { ChatModel, ChatModelOptions } from "@aigne/core/agents/chat-model.js";
2
- export declare const AIGNE_HUB_URL = "https://hub.aigne.io/";
3
- export declare function availableModels(): LoadableModel[];
4
- export interface LoadableModel {
5
- name: string | string[];
6
- apiKeyEnvName?: string | string[];
7
- create: (options: {
8
- model?: string;
9
- modelOptions?: ChatModelOptions;
10
- apiKey?: string;
11
- url?: string;
12
- }) => ChatModel;
13
- }
14
- export declare function findModel(models: LoadableModel[], provider: string): LoadableModel | undefined;
15
- type Model = {
16
- provider?: string | null;
17
- name?: string | null;
18
- temperature?: number | null;
19
- topP?: number | null;
20
- frequencyPenalty?: number | null;
21
- presencePenalty?: number | null;
22
- } | undefined;
23
- export declare function loadModel(model?: Model, modelOptions?: ChatModelOptions, credential?: {
24
- apiKey?: string;
25
- url?: string;
26
- }): Promise<ChatModel | undefined>;
27
- export declare function getAIGNEHubMountPoint(url: string): Promise<string>;
28
- export {};
@@ -1,28 +0,0 @@
1
- import type { ChatModel, ChatModelOptions } from "@aigne/core/agents/chat-model.js";
2
- export declare const AIGNE_HUB_URL = "https://hub.aigne.io/";
3
- export declare function availableModels(): LoadableModel[];
4
- export interface LoadableModel {
5
- name: string | string[];
6
- apiKeyEnvName?: string | string[];
7
- create: (options: {
8
- model?: string;
9
- modelOptions?: ChatModelOptions;
10
- apiKey?: string;
11
- url?: string;
12
- }) => ChatModel;
13
- }
14
- export declare function findModel(models: LoadableModel[], provider: string): LoadableModel | undefined;
15
- type Model = {
16
- provider?: string | null;
17
- name?: string | null;
18
- temperature?: number | null;
19
- topP?: number | null;
20
- frequencyPenalty?: number | null;
21
- presencePenalty?: number | null;
22
- } | undefined;
23
- export declare function loadModel(model?: Model, modelOptions?: ChatModelOptions, credential?: {
24
- apiKey?: string;
25
- url?: string;
26
- }): Promise<ChatModel | undefined>;
27
- export declare function getAIGNEHubMountPoint(url: string): Promise<string>;
28
- export {};
@@ -1,116 +0,0 @@
1
- import { AnthropicChatModel } from "@aigne/anthropic";
2
- import { BedrockChatModel } from "@aigne/bedrock";
3
- import { DeepSeekChatModel } from "@aigne/deepseek";
4
- import { GeminiChatModel } from "@aigne/gemini";
5
- import { OllamaChatModel } from "@aigne/ollama";
6
- import { OpenRouterChatModel } from "@aigne/open-router";
7
- import { OpenAIChatModel } from "@aigne/openai";
8
- import { nodejs } from "@aigne/platform-helpers/nodejs/index.js";
9
- import { XAIChatModel } from "@aigne/xai";
10
- import { NodeHttpHandler, streamCollector } from "@smithy/node-http-handler";
11
- import { HttpsProxyAgent } from "https-proxy-agent";
12
- import { joinURL } from "ufo";
13
- import { CliAIGNEHubChatModel } from "./cli-aigne-hub-model.js";
14
- const AIGNE_HUB_DID = "z8ia3xzq2tMq8CRHfaXj1BTYJyYnEcHbqP8cJ";
15
- export const AIGNE_HUB_URL = "https://hub.aigne.io/";
16
- export function availableModels() {
17
- const proxy = ["HTTPS_PROXY", "https_proxy", "HTTP_PROXY", "http_proxy", "ALL_PROXY", "all_proxy"]
18
- .map((i) => process.env[i])
19
- .filter(Boolean)[0];
20
- const httpAgent = proxy ? new HttpsProxyAgent(proxy) : undefined;
21
- const clientOptions = {
22
- fetchOptions: {
23
- // @ts-ignore
24
- agent: httpAgent,
25
- },
26
- };
27
- return [
28
- {
29
- name: OpenAIChatModel.name,
30
- apiKeyEnvName: "OPENAI_API_KEY",
31
- create: (params) => new OpenAIChatModel({ ...params, clientOptions }),
32
- },
33
- {
34
- name: AnthropicChatModel.name,
35
- apiKeyEnvName: "ANTHROPIC_API_KEY",
36
- create: (params) => new AnthropicChatModel({ ...params, clientOptions }),
37
- },
38
- {
39
- name: BedrockChatModel.name,
40
- apiKeyEnvName: "AWS_ACCESS_KEY_ID",
41
- create: (params) => new BedrockChatModel({
42
- ...params,
43
- clientOptions: {
44
- requestHandler: NodeHttpHandler.create({ httpAgent, httpsAgent: httpAgent }),
45
- streamCollector,
46
- },
47
- }),
48
- },
49
- {
50
- name: DeepSeekChatModel.name,
51
- apiKeyEnvName: "DEEPSEEK_API_KEY",
52
- create: (params) => new DeepSeekChatModel({ ...params, clientOptions }),
53
- },
54
- {
55
- name: [GeminiChatModel.name, "google"],
56
- apiKeyEnvName: ["GEMINI_API_KEY", "GOOGLE_API_KEY"],
57
- create: (params) => new GeminiChatModel({ ...params, clientOptions }),
58
- },
59
- {
60
- name: OllamaChatModel.name,
61
- apiKeyEnvName: "OLLAMA_API_KEY",
62
- create: (params) => new OllamaChatModel({ ...params, clientOptions }),
63
- },
64
- {
65
- name: OpenRouterChatModel.name,
66
- apiKeyEnvName: "OPEN_ROUTER_API_KEY",
67
- create: (params) => new OpenRouterChatModel({ ...params, clientOptions }),
68
- },
69
- {
70
- name: XAIChatModel.name,
71
- apiKeyEnvName: "XAI_API_KEY",
72
- create: (params) => new XAIChatModel({ ...params, clientOptions }),
73
- },
74
- {
75
- name: CliAIGNEHubChatModel.name,
76
- apiKeyEnvName: "AIGNE_HUB_API_KEY",
77
- create: (params) => new CliAIGNEHubChatModel({ ...params, clientOptions }),
78
- },
79
- ];
80
- }
81
- export function findModel(models, provider) {
82
- return models.find((m) => {
83
- if (typeof m.name === "string") {
84
- return m.name.toLowerCase().includes(provider.toLowerCase());
85
- }
86
- return m.name.some((n) => n.toLowerCase().includes(provider.toLowerCase()));
87
- });
88
- }
89
- const { MODEL_PROVIDER, MODEL_NAME } = nodejs.env;
90
- const DEFAULT_MODEL_PROVIDER = "openai";
91
- export async function loadModel(model, modelOptions, credential) {
92
- const params = {
93
- model: MODEL_NAME ?? model?.name ?? undefined,
94
- temperature: model?.temperature ?? undefined,
95
- topP: model?.topP ?? undefined,
96
- frequencyPenalty: model?.frequencyPenalty ?? undefined,
97
- presencePenalty: model?.presencePenalty ?? undefined,
98
- };
99
- const provider = (MODEL_PROVIDER ?? model?.provider ?? DEFAULT_MODEL_PROVIDER).replace(/-/g, "");
100
- const m = findModel(availableModels(), provider);
101
- if (!m)
102
- throw new Error(`Unsupported model: ${model?.provider} ${model?.name}`);
103
- return m.create({
104
- ...(credential || {}),
105
- model: params.model,
106
- modelOptions: { ...params, ...modelOptions },
107
- });
108
- }
109
- export async function getAIGNEHubMountPoint(url) {
110
- const { origin } = new URL(url);
111
- const BLOCKLET_JSON_PATH = "__blocklet__.js?type=json";
112
- const blockletInfo = await fetch(joinURL(origin, BLOCKLET_JSON_PATH));
113
- const blocklet = await blockletInfo.json();
114
- const aigneHubMount = (blocklet?.componentMountPoints || []).find((m) => m.did === AIGNE_HUB_DID);
115
- return joinURL(origin, aigneHubMount?.mountPoint || "");
116
- }