@easynet/agent-llm 1.0.15 → 1.0.16
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/api/create-agent-llm.d.ts +1 -19
- package/dist/api/create-agent-llm.d.ts.map +1 -1
- package/dist/{chunk-AZDHRCIM.js → chunk-CWU7GNFB.js} +38 -263
- package/dist/chunk-CWU7GNFB.js.map +1 -0
- package/dist/cli/index.d.ts +0 -7
- package/dist/cli/index.d.ts.map +1 -1
- package/dist/cli/index.js +3 -7
- package/dist/cli/index.js.map +1 -1
- package/dist/config/index.d.ts +0 -3
- package/dist/config/index.d.ts.map +1 -1
- package/dist/config/loader.d.ts +0 -14
- package/dist/config/loader.d.ts.map +1 -1
- package/dist/config/parser.d.ts +0 -3
- package/dist/config/parser.d.ts.map +1 -1
- package/dist/connectivity/check.d.ts +0 -23
- package/dist/connectivity/check.d.ts.map +1 -1
- package/dist/connectivity/index.d.ts +2 -4
- package/dist/connectivity/index.d.ts.map +1 -1
- package/dist/connectivity/types.d.ts +0 -11
- package/dist/connectivity/types.d.ts.map +1 -1
- package/dist/extensions/index.d.ts +2 -5
- package/dist/extensions/index.d.ts.map +1 -1
- package/dist/extensions/loader.d.ts +1 -17
- package/dist/extensions/loader.d.ts.map +1 -1
- package/dist/extensions/npm-protocol.d.ts +2 -33
- package/dist/extensions/npm-protocol.d.ts.map +1 -1
- package/dist/index.d.ts +9 -11
- package/dist/index.d.ts.map +1 -1
- package/dist/index.js +2 -41
- package/dist/index.js.map +1 -1
- package/dist/langchain/index.d.ts +13 -3
- package/dist/langchain/index.d.ts.map +1 -1
- package/dist/registry/chat-model.d.ts +1 -8
- package/dist/registry/chat-model.d.ts.map +1 -1
- package/dist/registry/index.d.ts +0 -5
- package/dist/registry/index.d.ts.map +1 -1
- package/dist/types.d.ts +8 -165
- package/dist/types.d.ts.map +1 -1
- package/package.json +6 -5
- package/dist/chunk-AZDHRCIM.js.map +0 -1
- package/dist/langchain/adapter.d.ts +0 -21
- package/dist/langchain/adapter.d.ts.map +0 -1
- package/dist/providers/index.d.ts +0 -7
- package/dist/providers/index.d.ts.map +0 -1
- package/dist/providers/openai.d.ts +0 -9
- package/dist/providers/openai.d.ts.map +0 -1
- package/dist/registry/client.d.ts +0 -13
- package/dist/registry/client.d.ts.map +0 -1
package/dist/index.js
CHANGED
|
@@ -1,4 +1,5 @@
|
|
|
1
1
|
import {
|
|
2
|
+
AGENT_LLM_PROVIDER_FIELD,
|
|
2
3
|
CIS_UNREACHABLE_REMINDER,
|
|
3
4
|
NPM_PROTOCOL_PREFIX,
|
|
4
5
|
buildUnreachableError,
|
|
@@ -6,10 +7,6 @@ import {
|
|
|
6
7
|
createAgentLlM,
|
|
7
8
|
createChatModelFromLlmConfig,
|
|
8
9
|
createChatModelFromLlmConfigWithNpm,
|
|
9
|
-
createClient,
|
|
10
|
-
createOpenAIChatClient,
|
|
11
|
-
createOpenAIClient,
|
|
12
|
-
createOpenAIImageClient,
|
|
13
10
|
discoverLLMExtensions,
|
|
14
11
|
ensureNpmPackageInstalled,
|
|
15
12
|
getChatModelFactory,
|
|
@@ -23,41 +20,11 @@ import {
|
|
|
23
20
|
parseLlmYaml,
|
|
24
21
|
parseNpmProviderSpec,
|
|
25
22
|
registerChatModelProvider,
|
|
26
|
-
registerProvider,
|
|
27
23
|
resolveLLMExtensionPackages,
|
|
28
24
|
resolveLlmSectionWithNpm,
|
|
29
25
|
resolveNpmProvider,
|
|
30
26
|
substituteEnv
|
|
31
|
-
} from "./chunk-
|
|
32
|
-
|
|
33
|
-
// src/registry/client.ts
|
|
34
|
-
function createLLMRegistry(options) {
|
|
35
|
-
const { defaultId, configs } = parseLlmSection(options.llmSection);
|
|
36
|
-
const map = /* @__PURE__ */ new Map();
|
|
37
|
-
for (const config of configs) {
|
|
38
|
-
try {
|
|
39
|
-
const client = createClient(config);
|
|
40
|
-
map.set(config.id, client);
|
|
41
|
-
} catch (err) {
|
|
42
|
-
console.warn(`[agent-llm] Skip LLM "${config.id}": ${err instanceof Error ? err.message : String(err)}`);
|
|
43
|
-
}
|
|
44
|
-
}
|
|
45
|
-
return {
|
|
46
|
-
get(id) {
|
|
47
|
-
return map.get(id);
|
|
48
|
-
},
|
|
49
|
-
defaultId() {
|
|
50
|
-
if (map.has(defaultId)) return defaultId;
|
|
51
|
-
return map.size > 0 ? [...map.keys()][0] : void 0;
|
|
52
|
-
},
|
|
53
|
-
ids() {
|
|
54
|
-
return [...map.keys()];
|
|
55
|
-
}
|
|
56
|
-
};
|
|
57
|
-
}
|
|
58
|
-
|
|
59
|
-
// src/types.ts
|
|
60
|
-
var AGENT_LLM_PROVIDER_FIELD = "agentLlmProvider";
|
|
27
|
+
} from "./chunk-CWU7GNFB.js";
|
|
61
28
|
export {
|
|
62
29
|
AGENT_LLM_PROVIDER_FIELD,
|
|
63
30
|
CIS_UNREACHABLE_REMINDER,
|
|
@@ -67,11 +34,6 @@ export {
|
|
|
67
34
|
createAgentLlM,
|
|
68
35
|
createChatModelFromLlmConfig,
|
|
69
36
|
createChatModelFromLlmConfigWithNpm,
|
|
70
|
-
createClient,
|
|
71
|
-
createLLMRegistry,
|
|
72
|
-
createOpenAIChatClient,
|
|
73
|
-
createOpenAIClient,
|
|
74
|
-
createOpenAIImageClient,
|
|
75
37
|
discoverLLMExtensions,
|
|
76
38
|
ensureNpmPackageInstalled,
|
|
77
39
|
getChatModelFactory,
|
|
@@ -85,7 +47,6 @@ export {
|
|
|
85
47
|
parseLlmYaml,
|
|
86
48
|
parseNpmProviderSpec,
|
|
87
49
|
registerChatModelProvider,
|
|
88
|
-
registerProvider,
|
|
89
50
|
resolveLLMExtensionPackages,
|
|
90
51
|
resolveLlmSectionWithNpm,
|
|
91
52
|
resolveNpmProvider,
|
package/dist/index.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"sources":[
|
|
1
|
+
{"version":3,"sources":[],"sourcesContent":[],"mappings":"","names":[]}
|
|
@@ -1,6 +1,16 @@
|
|
|
1
1
|
/**
|
|
2
|
-
* LangChain module:
|
|
2
|
+
* LangChain module: create ChatOpenAI from llm config.
|
|
3
|
+
* Uses @langchain/openai only; extensions can register custom ChatModel factories.
|
|
3
4
|
*/
|
|
4
|
-
|
|
5
|
-
export
|
|
5
|
+
import type { BaseChatModel } from "@langchain/core/language_models/chat_models";
|
|
6
|
+
export interface CreateChatModelFromLlmConfigOptions {
|
|
7
|
+
llmSection?: unknown;
|
|
8
|
+
modelEnv?: string;
|
|
9
|
+
apiKeyEnv?: string;
|
|
10
|
+
}
|
|
11
|
+
/**
|
|
12
|
+
* Create a LangChain ChatModel from agent config llm section.
|
|
13
|
+
* Uses extension-registered factory when available; otherwise ChatOpenAI from @langchain/openai.
|
|
14
|
+
*/
|
|
15
|
+
export declare function createChatModelFromLlmConfig(options: CreateChatModelFromLlmConfigOptions): BaseChatModel;
|
|
6
16
|
//# sourceMappingURL=index.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/langchain/index.ts"],"names":[],"mappings":"AAAA
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/langchain/index.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAGH,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,6CAA6C,CAAC;AAMjF,MAAM,WAAW,mCAAmC;IAClD,UAAU,CAAC,EAAE,OAAO,CAAC;IACrB,QAAQ,CAAC,EAAE,MAAM,CAAC;IAClB,SAAS,CAAC,EAAE,MAAM,CAAC;CACpB;AAOD;;;GAGG;AACH,wBAAgB,4BAA4B,CAC1C,OAAO,EAAE,mCAAmC,GAC3C,aAAa,CA0Ef"}
|
|
@@ -1,17 +1,10 @@
|
|
|
1
1
|
/**
|
|
2
2
|
* Registry for LangChain ChatModel by provider name.
|
|
3
|
-
* Extensions register via registerChatModelProvider
|
|
3
|
+
* Extensions register via registerChatModelProvider.
|
|
4
4
|
*/
|
|
5
5
|
import type { BaseChatModel } from "@langchain/core/language_models/chat_models";
|
|
6
6
|
import type { LLMConfig } from "../types.js";
|
|
7
7
|
export type ChatModelFactory = (config: LLMConfig) => BaseChatModel;
|
|
8
|
-
/**
|
|
9
|
-
* Register a ChatModel factory for a provider name.
|
|
10
|
-
* Called by extensions (e.g. wallee-llm) on load.
|
|
11
|
-
*/
|
|
12
8
|
export declare function registerChatModelProvider(providerName: string, factory: ChatModelFactory): void;
|
|
13
|
-
/**
|
|
14
|
-
* Get the ChatModel factory for a provider name, if registered.
|
|
15
|
-
*/
|
|
16
9
|
export declare function getChatModelFactory(providerName: string): ChatModelFactory | undefined;
|
|
17
10
|
//# sourceMappingURL=chat-model.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"chat-model.d.ts","sourceRoot":"","sources":["../../src/registry/chat-model.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAEH,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,6CAA6C,CAAC;AACjF,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,aAAa,CAAC;AAE7C,MAAM,MAAM,gBAAgB,GAAG,CAAC,MAAM,EAAE,SAAS,KAAK,aAAa,CAAC;AAIpE
|
|
1
|
+
{"version":3,"file":"chat-model.d.ts","sourceRoot":"","sources":["../../src/registry/chat-model.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAEH,OAAO,KAAK,EAAE,aAAa,EAAE,MAAM,6CAA6C,CAAC;AACjF,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,aAAa,CAAC;AAE7C,MAAM,MAAM,gBAAgB,GAAG,CAAC,MAAM,EAAE,SAAS,KAAK,aAAa,CAAC;AAIpE,wBAAgB,yBAAyB,CAAC,YAAY,EAAE,MAAM,EAAE,OAAO,EAAE,gBAAgB,GAAG,IAAI,CAE/F;AAED,wBAAgB,mBAAmB,CAAC,YAAY,EAAE,MAAM,GAAG,gBAAgB,GAAG,SAAS,CAEtF"}
|
package/dist/registry/index.d.ts
CHANGED
|
@@ -1,8 +1,3 @@
|
|
|
1
|
-
/**
|
|
2
|
-
* Registry module: manage LLM client and ChatModel registries.
|
|
3
|
-
*/
|
|
4
|
-
export { createLLMRegistry } from "./client.js";
|
|
5
|
-
export type { CreateLLMRegistryOptions } from "./client.js";
|
|
6
1
|
export { registerChatModelProvider, getChatModelFactory } from "./chat-model.js";
|
|
7
2
|
export type { ChatModelFactory } from "./chat-model.js";
|
|
8
3
|
//# sourceMappingURL=index.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/registry/index.ts"],"names":[],"mappings":"AAAA
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/registry/index.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,yBAAyB,EAAE,mBAAmB,EAAE,MAAM,iBAAiB,CAAC;AACjF,YAAY,EAAE,gBAAgB,EAAE,MAAM,iBAAiB,CAAC"}
|
package/dist/types.d.ts
CHANGED
|
@@ -1,37 +1,22 @@
|
|
|
1
1
|
/**
|
|
2
|
-
* Agent LLM: OpenAI-compatible
|
|
3
|
-
*
|
|
2
|
+
* Agent LLM: OpenAI-compatible config only. No custom client types.
|
|
3
|
+
* LangChain ChatOpenAI is the single implementation.
|
|
4
4
|
*/
|
|
5
|
-
/** LLM type: chat = conversation, image = image generation (OpenAI-compatible format) */
|
|
6
5
|
export type LLMType = "chat" | "image";
|
|
7
|
-
/** Single LLM config: id, type, model; OpenAI-compatible API only */
|
|
8
6
|
export interface LLMConfig {
|
|
9
|
-
/** Unique id for fetching the instance from the registry */
|
|
10
7
|
id: string;
|
|
11
|
-
/** chat | image */
|
|
12
8
|
type: LLMType;
|
|
13
|
-
/** Must be openai or openai-compatible; only this format is supported */
|
|
14
9
|
provider: string;
|
|
15
|
-
/** Model name, e.g. gpt-4o-mini, dall-e-3 */
|
|
16
10
|
model?: string;
|
|
17
|
-
/** Temperature etc.; commonly used for chat */
|
|
18
11
|
temperature?: number;
|
|
19
|
-
/** API key; can also be set via env */
|
|
20
12
|
apiKey?: string;
|
|
21
|
-
/** OpenAI-compatible endpoint baseURL (e.g. Azure, local proxy, other /v1-compatible vendors) */
|
|
22
13
|
baseURL?: string;
|
|
23
|
-
/** Other options (passed through) */
|
|
24
14
|
[key: string]: unknown;
|
|
25
15
|
}
|
|
26
|
-
/** agent.yaml llm section: flat (each model keyed by name), default+instances, or single object */
|
|
27
16
|
export interface AgentConfigLlmSection {
|
|
28
|
-
/** Default model name (id) to use */
|
|
29
17
|
default?: string;
|
|
30
|
-
/** npm package name or array for dynamic load; e.g. "wallee-llm" or ["wallee-llm"] */
|
|
31
18
|
type?: string | string[];
|
|
32
|
-
/** Array of LLM configs */
|
|
33
19
|
instances?: LLMConfig[];
|
|
34
|
-
/** Single-object form: provider, model etc., parsed as id=default chat */
|
|
35
20
|
provider?: string;
|
|
36
21
|
model?: string;
|
|
37
22
|
name?: string;
|
|
@@ -39,161 +24,19 @@ export interface AgentConfigLlmSection {
|
|
|
39
24
|
apiKey?: string;
|
|
40
25
|
baseURL?: string;
|
|
41
26
|
base_url?: string;
|
|
42
|
-
/** Flat: strong/medium/fast etc. name → config (provider, base_url, name, options) */
|
|
43
27
|
[key: string]: unknown;
|
|
44
28
|
}
|
|
45
|
-
/**
|
|
46
|
-
export
|
|
47
|
-
role: "system" | "user" | "assistant";
|
|
48
|
-
content: string;
|
|
49
|
-
}
|
|
50
|
-
/** Minimal chat result */
|
|
51
|
-
export interface ChatResult {
|
|
52
|
-
content: string;
|
|
53
|
-
usage?: {
|
|
54
|
-
promptTokens?: number;
|
|
55
|
-
completionTokens?: number;
|
|
56
|
-
};
|
|
57
|
-
}
|
|
58
|
-
/** Stream chunk for chat responses */
|
|
59
|
-
export interface ChatStreamChunk {
|
|
60
|
-
content: string;
|
|
61
|
-
done: boolean;
|
|
62
|
-
usage?: {
|
|
63
|
-
promptTokens?: number;
|
|
64
|
-
completionTokens?: number;
|
|
65
|
-
};
|
|
66
|
-
}
|
|
67
|
-
/** OpenAI-compatible tool definition (function) */
|
|
68
|
-
export interface ToolDefinition {
|
|
69
|
-
type: "function";
|
|
70
|
-
function: {
|
|
71
|
-
name: string;
|
|
72
|
-
description?: string;
|
|
73
|
-
parameters?: object;
|
|
74
|
-
};
|
|
75
|
-
}
|
|
76
|
-
/** Message with tool calls (assistant may include tool_calls; tool = tool result) */
|
|
77
|
-
export type ChatWithToolsMessage = ChatMessage | {
|
|
78
|
-
role: "tool";
|
|
79
|
-
content: string;
|
|
80
|
-
tool_call_id: string;
|
|
81
|
-
} | {
|
|
82
|
-
role: "assistant";
|
|
83
|
-
content?: string | null;
|
|
84
|
-
tool_calls?: Array<{
|
|
85
|
-
id: string;
|
|
86
|
-
type: "function";
|
|
87
|
-
function: {
|
|
88
|
-
name: string;
|
|
89
|
-
arguments: string;
|
|
90
|
-
};
|
|
91
|
-
}>;
|
|
92
|
-
};
|
|
93
|
-
/** Chat result with tool calls */
|
|
94
|
-
export interface ChatWithToolsResult {
|
|
95
|
-
message: {
|
|
96
|
-
role: "assistant";
|
|
97
|
-
content?: string | null;
|
|
98
|
-
tool_calls?: Array<{
|
|
99
|
-
id: string;
|
|
100
|
-
type: "function";
|
|
101
|
-
function: {
|
|
102
|
-
name: string;
|
|
103
|
-
arguments: string;
|
|
104
|
-
};
|
|
105
|
-
}>;
|
|
106
|
-
};
|
|
107
|
-
usage?: {
|
|
108
|
-
promptTokens?: number;
|
|
109
|
-
completionTokens?: number;
|
|
110
|
-
};
|
|
111
|
-
}
|
|
112
|
-
/** Stream chunk for chat with tools responses */
|
|
113
|
-
export interface ChatWithToolsStreamChunk {
|
|
114
|
-
delta: {
|
|
115
|
-
role?: "assistant";
|
|
116
|
-
content?: string | null;
|
|
117
|
-
tool_calls?: Array<{
|
|
118
|
-
index: number;
|
|
119
|
-
id?: string;
|
|
120
|
-
type?: "function";
|
|
121
|
-
function?: {
|
|
122
|
-
name?: string;
|
|
123
|
-
arguments?: string;
|
|
124
|
-
};
|
|
125
|
-
}>;
|
|
126
|
-
};
|
|
127
|
-
done: boolean;
|
|
128
|
-
usage?: {
|
|
129
|
-
promptTokens?: number;
|
|
130
|
-
completionTokens?: number;
|
|
131
|
-
};
|
|
132
|
-
}
|
|
133
|
-
/** Image generation result */
|
|
134
|
-
export interface ImageResult {
|
|
135
|
-
url?: string;
|
|
136
|
-
b64?: string;
|
|
137
|
-
}
|
|
138
|
-
/**
|
|
139
|
-
* Minimal interface for a single LLM instance.
|
|
140
|
-
* type=chat provides chat; type=image provides generateImage.
|
|
141
|
-
* chat type may optionally provide chatWithTools for ReAct/Agent tool calling.
|
|
142
|
-
*/
|
|
143
|
-
export interface ILLMClient {
|
|
144
|
-
readonly id: string;
|
|
145
|
-
readonly type: LLMType;
|
|
146
|
-
/** Chat (available when type=chat) */
|
|
147
|
-
chat(messages: ChatMessage[]): Promise<ChatResult>;
|
|
148
|
-
/**
|
|
149
|
-
* Chat with streaming (optional when type=chat).
|
|
150
|
-
* Returns an async generator that yields chunks as they arrive.
|
|
151
|
-
*/
|
|
152
|
-
chatStream?(messages: ChatMessage[]): AsyncGenerator<ChatStreamChunk, void, unknown>;
|
|
153
|
-
/**
|
|
154
|
-
* Chat with tools (optional when type=chat; for ReAct/Agent).
|
|
155
|
-
* If not implemented, caller may poll with chat or use another client.
|
|
156
|
-
*/
|
|
157
|
-
chatWithTools?(messages: ChatWithToolsMessage[], tools: ToolDefinition[], options?: {
|
|
158
|
-
timeoutMs?: number;
|
|
159
|
-
}): Promise<ChatWithToolsResult>;
|
|
160
|
-
/**
|
|
161
|
-
* Chat with tools and streaming (optional when type=chat).
|
|
162
|
-
* Returns an async generator that yields chunks as they arrive.
|
|
163
|
-
*/
|
|
164
|
-
chatWithToolsStream?(messages: ChatWithToolsMessage[], tools: ToolDefinition[], options?: {
|
|
165
|
-
timeoutMs?: number;
|
|
166
|
-
}): AsyncGenerator<ChatWithToolsStreamChunk, void, unknown>;
|
|
167
|
-
/** Image generation (when type=image); otherwise may throw or be ignored */
|
|
168
|
-
generateImage?(options: {
|
|
169
|
-
prompt: string;
|
|
170
|
-
size?: string;
|
|
171
|
-
n?: number;
|
|
172
|
-
}): Promise<ImageResult>;
|
|
173
|
-
}
|
|
174
|
-
/**
|
|
175
|
-
* LLM registry created from llm section: get instance by id.
|
|
176
|
-
*/
|
|
177
|
-
export interface ILLMRegistry {
|
|
178
|
-
get(id: string): ILLMClient | undefined;
|
|
179
|
-
defaultId(): string | undefined;
|
|
180
|
-
ids(): string[];
|
|
181
|
-
}
|
|
29
|
+
/** package.json field: when true, the package is an agent-llm provider (for discovery). */
|
|
30
|
+
export declare const AGENT_LLM_PROVIDER_FIELD = "agentLlmProvider";
|
|
182
31
|
/**
|
|
183
|
-
*
|
|
184
|
-
*
|
|
185
|
-
* Prefer createClient + createChatModel so the framework does registration (extension need not call register).
|
|
32
|
+
* Extension interface: register a ChatModel factory by provider name.
|
|
33
|
+
* Extensions implement getLLMExtension(); framework registers the factory when loading.
|
|
186
34
|
*/
|
|
187
35
|
export interface ILLMExtension {
|
|
188
|
-
/** Provider name (e.g. "openai" or extension-specific name). Framework registers under this name. */
|
|
189
36
|
readonly providerName: string;
|
|
190
|
-
/**
|
|
191
|
-
readonly createClient?: (config: LLMConfig) => ILLMClient;
|
|
192
|
-
/** LangChain ChatModel factory. If present with createClient, framework calls registerChatModelProvider(providerName, createChatModel). */
|
|
37
|
+
/** LangChain ChatModel factory. Framework calls registerChatModelProvider(providerName, createChatModel). */
|
|
193
38
|
readonly createChatModel?: (config: LLMConfig) => unknown;
|
|
194
|
-
/** Legacy: if extension provides register(), framework calls it.
|
|
39
|
+
/** Legacy: if extension provides register(), framework calls it. */
|
|
195
40
|
register?(): void;
|
|
196
41
|
}
|
|
197
|
-
/** package.json field: when true, the package is an agent-llm provider (for discovery). */
|
|
198
|
-
export declare const AGENT_LLM_PROVIDER_FIELD = "agentLlmProvider";
|
|
199
42
|
//# sourceMappingURL=types.d.ts.map
|
package/dist/types.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"types.d.ts","sourceRoot":"","sources":["../src/types.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAEH,
|
|
1
|
+
{"version":3,"file":"types.d.ts","sourceRoot":"","sources":["../src/types.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAEH,MAAM,MAAM,OAAO,GAAG,MAAM,GAAG,OAAO,CAAC;AAEvC,MAAM,WAAW,SAAS;IACxB,EAAE,EAAE,MAAM,CAAC;IACX,IAAI,EAAE,OAAO,CAAC;IACd,QAAQ,EAAE,MAAM,CAAC;IACjB,KAAK,CAAC,EAAE,MAAM,CAAC;IACf,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB,MAAM,CAAC,EAAE,MAAM,CAAC;IAChB,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC;CACxB;AAED,MAAM,WAAW,qBAAqB;IACpC,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB,IAAI,CAAC,EAAE,MAAM,GAAG,MAAM,EAAE,CAAC;IACzB,SAAS,CAAC,EAAE,SAAS,EAAE,CAAC;IACxB,QAAQ,CAAC,EAAE,MAAM,CAAC;IAClB,KAAK,CAAC,EAAE,MAAM,CAAC;IACf,IAAI,CAAC,EAAE,MAAM,CAAC;IACd,WAAW,CAAC,EAAE,MAAM,CAAC;IACrB,MAAM,CAAC,EAAE,MAAM,CAAC;IAChB,OAAO,CAAC,EAAE,MAAM,CAAC;IACjB,QAAQ,CAAC,EAAE,MAAM,CAAC;IAClB,CAAC,GAAG,EAAE,MAAM,GAAG,OAAO,CAAC;CACxB;AAED,2FAA2F;AAC3F,eAAO,MAAM,wBAAwB,qBAAqB,CAAC;AAE3D;;;GAGG;AACH,MAAM,WAAW,aAAa;IAC5B,QAAQ,CAAC,YAAY,EAAE,MAAM,CAAC;IAC9B,6GAA6G;IAC7G,QAAQ,CAAC,eAAe,CAAC,EAAE,CAAC,MAAM,EAAE,SAAS,KAAK,OAAO,CAAC;IAC1D,oEAAoE;IACpE,QAAQ,CAAC,IAAI,IAAI,CAAC;CACnB"}
|
package/package.json
CHANGED
|
@@ -1,11 +1,11 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@easynet/agent-llm",
|
|
3
|
-
"version": "1.0.
|
|
3
|
+
"version": "1.0.16",
|
|
4
4
|
"description": "Agent LLM: multi-provider, multi-model, simple chat/image API. Consumes agent.yaml llm section.",
|
|
5
5
|
"type": "module",
|
|
6
6
|
"bin": {
|
|
7
|
-
"agent-llm": "./dist/cli.js",
|
|
8
|
-
"@easynet/agent-llm": "./dist/cli.js"
|
|
7
|
+
"agent-llm": "./dist/cli/index.js",
|
|
8
|
+
"@easynet/agent-llm": "./dist/cli/index.js"
|
|
9
9
|
},
|
|
10
10
|
"main": "./dist/index.js",
|
|
11
11
|
"types": "./dist/index.d.ts",
|
|
@@ -14,8 +14,9 @@
|
|
|
14
14
|
"dev": "tsup --watch",
|
|
15
15
|
"test": "vitest run",
|
|
16
16
|
"test:watch": "vitest",
|
|
17
|
-
"test:
|
|
18
|
-
"test:
|
|
17
|
+
"test:unit": "vitest run test/unit",
|
|
18
|
+
"test:integration": "vitest run test/integration",
|
|
19
|
+
"test:live": "vitest run test/integration/llm-live.test.ts",
|
|
19
20
|
"example:agent": "cd examples && npm run start",
|
|
20
21
|
"typecheck": "tsc --noEmit"
|
|
21
22
|
},
|