opc-agent 1.1.0 → 1.1.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +51 -51
- package/CONTRIBUTING.md +75 -75
- package/README.md +341 -101
- package/README.zh-CN.md +382 -55
- package/dist/channels/web.js +256 -256
- package/dist/cli.js +1 -32
- package/dist/deploy/hermes.js +22 -22
- package/dist/deploy/openclaw.js +31 -31
- package/dist/i18n/index.js +60 -9
- package/dist/templates/code-reviewer.js +5 -5
- package/dist/templates/customer-service.js +2 -2
- package/dist/templates/data-analyst.js +5 -5
- package/dist/templates/knowledge-base.js +2 -2
- package/dist/templates/sales-assistant.js +4 -4
- package/dist/templates/teacher.js +6 -6
- package/docs/.vitepress/config.ts +103 -92
- package/docs/api/cli.md +48 -48
- package/docs/api/oad-schema.md +64 -64
- package/docs/api/sdk.md +80 -80
- package/docs/guide/concepts.md +51 -51
- package/docs/guide/configuration.md +79 -79
- package/docs/guide/deployment.md +42 -42
- package/docs/guide/getting-started.md +44 -44
- package/docs/guide/templates.md +28 -28
- package/docs/guide/testing.md +84 -84
- package/docs/index.md +27 -27
- package/docs/zh/api/cli.md +54 -0
- package/docs/zh/api/oad-schema.md +87 -3
- package/docs/zh/api/sdk.md +102 -0
- package/docs/zh/guide/concepts.md +104 -28
- package/docs/zh/guide/configuration.md +135 -39
- package/docs/zh/guide/deployment.md +81 -3
- package/docs/zh/guide/getting-started.md +82 -58
- package/docs/zh/guide/templates.md +84 -22
- package/docs/zh/guide/testing.md +88 -18
- package/docs/zh/index.md +27 -27
- package/examples/customer-service-demo/README.md +90 -90
- package/examples/customer-service-demo/oad.yaml +107 -107
- package/package.json +1 -1
- package/src/analytics/index.ts +66 -66
- package/src/channels/discord.ts +192 -192
- package/src/channels/email.ts +177 -177
- package/src/channels/feishu.ts +236 -236
- package/src/channels/index.ts +15 -15
- package/src/channels/slack.ts +160 -160
- package/src/channels/telegram.ts +90 -90
- package/src/channels/voice.ts +106 -106
- package/src/channels/web.ts +596 -596
- package/src/channels/webhook.ts +199 -199
- package/src/channels/websocket.ts +87 -87
- package/src/channels/wechat.ts +149 -149
- package/src/cli.ts +1 -35
- package/src/core/a2a.ts +143 -143
- package/src/core/agent.ts +152 -152
- package/src/core/analytics-engine.ts +186 -186
- package/src/core/auth.ts +57 -57
- package/src/core/cache.ts +141 -141
- package/src/core/compose.ts +77 -77
- package/src/core/config.ts +14 -14
- package/src/core/errors.ts +148 -148
- package/src/core/hitl.ts +138 -138
- package/src/core/knowledge.ts +210 -210
- package/src/core/logger.ts +57 -57
- package/src/core/orchestrator.ts +215 -215
- package/src/core/performance.ts +187 -187
- package/src/core/rate-limiter.ts +128 -128
- package/src/core/room.ts +109 -109
- package/src/core/runtime.ts +152 -152
- package/src/core/sandbox.ts +101 -101
- package/src/core/security.ts +171 -171
- package/src/core/types.ts +68 -68
- package/src/core/versioning.ts +106 -106
- package/src/core/watch.ts +178 -178
- package/src/core/workflow.ts +235 -235
- package/src/deploy/hermes.ts +156 -156
- package/src/deploy/openclaw.ts +200 -200
- package/src/dtv/data.ts +29 -29
- package/src/dtv/trust.ts +43 -43
- package/src/dtv/value.ts +47 -47
- package/src/i18n/index.ts +216 -165
- package/src/index.ts +110 -110
- package/src/marketplace/index.ts +223 -223
- package/src/memory/deepbrain.ts +108 -108
- package/src/memory/index.ts +34 -34
- package/src/plugins/index.ts +208 -208
- package/src/providers/index.ts +183 -183
- package/src/schema/oad.ts +155 -155
- package/src/skills/base.ts +16 -16
- package/src/skills/document.ts +100 -100
- package/src/skills/http.ts +35 -35
- package/src/skills/index.ts +27 -27
- package/src/skills/scheduler.ts +80 -80
- package/src/skills/webhook-trigger.ts +59 -59
- package/src/templates/code-reviewer.ts +34 -34
- package/src/templates/customer-service.ts +80 -80
- package/src/templates/data-analyst.ts +70 -70
- package/src/templates/executive-assistant.ts +71 -71
- package/src/templates/financial-advisor.ts +60 -60
- package/src/templates/knowledge-base.ts +31 -31
- package/src/templates/legal-assistant.ts +71 -71
- package/src/templates/sales-assistant.ts +79 -79
- package/src/templates/teacher.ts +79 -79
- package/src/testing/index.ts +181 -181
- package/src/tools/calculator.ts +73 -73
- package/src/tools/datetime.ts +149 -149
- package/src/tools/json-transform.ts +187 -187
- package/src/tools/mcp.ts +76 -76
- package/src/tools/text-analysis.ts +116 -116
- package/templates/Dockerfile +15 -15
- package/templates/code-reviewer/README.md +27 -27
- package/templates/code-reviewer/oad.yaml +41 -41
- package/templates/customer-service/README.md +22 -22
- package/templates/customer-service/oad.yaml +36 -36
- package/templates/docker-compose.yml +21 -21
- package/templates/knowledge-base/README.md +28 -28
- package/templates/knowledge-base/oad.yaml +38 -38
- package/templates/sales-assistant/README.md +26 -26
- package/templates/sales-assistant/oad.yaml +43 -43
- package/tests/a2a.test.ts +66 -66
- package/tests/agent.test.ts +72 -72
- package/tests/analytics.test.ts +50 -50
- package/tests/channel.test.ts +39 -39
- package/tests/e2e.test.ts +134 -134
- package/tests/errors.test.ts +83 -83
- package/tests/hitl.test.ts +71 -71
- package/tests/i18n.test.ts +41 -41
- package/tests/mcp.test.ts +54 -54
- package/tests/oad.test.ts +68 -68
- package/tests/performance.test.ts +115 -115
- package/tests/plugin.test.ts +74 -74
- package/tests/room.test.ts +106 -106
- package/tests/runtime.test.ts +42 -42
- package/tests/sandbox.test.ts +46 -46
- package/tests/security.test.ts +60 -60
- package/tests/templates.test.ts +77 -77
- package/tests/v070.test.ts +76 -76
- package/tests/versioning.test.ts +75 -75
- package/tests/voice.test.ts +61 -61
- package/tests/webhook.test.ts +29 -29
- package/tests/workflow.test.ts +143 -143
- package/tsconfig.json +19 -19
- package/vitest.config.ts +9 -9
package/src/core/agent.ts
CHANGED
|
@@ -1,152 +1,152 @@
|
|
|
1
|
-
import { EventEmitter } from 'events';
|
|
2
|
-
import type { AgentState, IAgent, IChannel, ISkill, Message, MemoryStore, AgentContext } from './types';
|
|
3
|
-
import { InMemoryStore } from '../memory';
|
|
4
|
-
import { createProvider, type LLMProvider } from '../providers';
|
|
5
|
-
|
|
6
|
-
export class BaseAgent extends EventEmitter implements IAgent {
|
|
7
|
-
readonly name: string;
|
|
8
|
-
private _state: AgentState = 'init';
|
|
9
|
-
private skills: Map<string, ISkill> = new Map();
|
|
10
|
-
private channels: IChannel[] = [];
|
|
11
|
-
private memory: MemoryStore;
|
|
12
|
-
private _provider: LLMProvider;
|
|
13
|
-
private systemPrompt: string;
|
|
14
|
-
private historyLimit: number;
|
|
15
|
-
|
|
16
|
-
constructor(options: {
|
|
17
|
-
name: string;
|
|
18
|
-
systemPrompt?: string;
|
|
19
|
-
provider?: string;
|
|
20
|
-
model?: string;
|
|
21
|
-
memory?: MemoryStore;
|
|
22
|
-
historyLimit?: number;
|
|
23
|
-
}) {
|
|
24
|
-
super();
|
|
25
|
-
this.name = options.name;
|
|
26
|
-
this.systemPrompt = options.systemPrompt ?? 'You are a helpful AI agent.';
|
|
27
|
-
this.memory = options.memory ?? new InMemoryStore();
|
|
28
|
-
this._provider = createProvider(options.provider ?? 'openai', options.model);
|
|
29
|
-
this.historyLimit = options.historyLimit ?? 50;
|
|
30
|
-
}
|
|
31
|
-
|
|
32
|
-
get state(): AgentState {
|
|
33
|
-
return this._state;
|
|
34
|
-
}
|
|
35
|
-
|
|
36
|
-
get provider(): LLMProvider {
|
|
37
|
-
return this._provider;
|
|
38
|
-
}
|
|
39
|
-
|
|
40
|
-
getSystemPrompt(): string {
|
|
41
|
-
return this.systemPrompt;
|
|
42
|
-
}
|
|
43
|
-
|
|
44
|
-
getMemory(): MemoryStore {
|
|
45
|
-
return this.memory;
|
|
46
|
-
}
|
|
47
|
-
|
|
48
|
-
private transition(to: AgentState): void {
|
|
49
|
-
const from = this._state;
|
|
50
|
-
this._state = to;
|
|
51
|
-
this.emit('state:change', from, to);
|
|
52
|
-
}
|
|
53
|
-
|
|
54
|
-
async init(): Promise<void> {
|
|
55
|
-
this.transition('ready');
|
|
56
|
-
}
|
|
57
|
-
|
|
58
|
-
async start(): Promise<void> {
|
|
59
|
-
if (this._state !== 'ready') {
|
|
60
|
-
throw new Error(`Cannot start agent in state: ${this._state}`);
|
|
61
|
-
}
|
|
62
|
-
for (const channel of this.channels) {
|
|
63
|
-
channel.onMessage((msg) => this.handleMessage(msg));
|
|
64
|
-
await channel.start();
|
|
65
|
-
}
|
|
66
|
-
this.transition('running');
|
|
67
|
-
}
|
|
68
|
-
|
|
69
|
-
async stop(): Promise<void> {
|
|
70
|
-
for (const channel of this.channels) {
|
|
71
|
-
await channel.stop();
|
|
72
|
-
}
|
|
73
|
-
this.transition('stopped');
|
|
74
|
-
}
|
|
75
|
-
|
|
76
|
-
registerSkill(skill: ISkill): void {
|
|
77
|
-
this.skills.set(skill.name, skill);
|
|
78
|
-
}
|
|
79
|
-
|
|
80
|
-
bindChannel(channel: IChannel): void {
|
|
81
|
-
this.channels.push(channel);
|
|
82
|
-
}
|
|
83
|
-
|
|
84
|
-
getChannels(): IChannel[] {
|
|
85
|
-
return this.channels;
|
|
86
|
-
}
|
|
87
|
-
|
|
88
|
-
async handleMessage(message: Message): Promise<Message> {
|
|
89
|
-
this.emit('message:in', message);
|
|
90
|
-
|
|
91
|
-
const sessionId = (message.metadata?.sessionId as string) ?? 'default';
|
|
92
|
-
await this.memory.addMessage(sessionId, message);
|
|
93
|
-
|
|
94
|
-
const context: AgentContext = {
|
|
95
|
-
agentName: this.name,
|
|
96
|
-
sessionId,
|
|
97
|
-
messages: (await this.memory.getConversation(sessionId)).slice(-this.historyLimit),
|
|
98
|
-
memory: this.memory,
|
|
99
|
-
metadata: {},
|
|
100
|
-
};
|
|
101
|
-
|
|
102
|
-
// Try skills first
|
|
103
|
-
for (const [name, skill] of this.skills) {
|
|
104
|
-
try {
|
|
105
|
-
const result = await skill.execute(context, message);
|
|
106
|
-
this.emit('skill:execute', name, result);
|
|
107
|
-
if (result.handled && result.response) {
|
|
108
|
-
const response = this.createResponse(result.response, message);
|
|
109
|
-
await this.memory.addMessage(sessionId, response);
|
|
110
|
-
this.emit('message:out', response);
|
|
111
|
-
return response;
|
|
112
|
-
}
|
|
113
|
-
} catch (err) {
|
|
114
|
-
this.emit('error', err instanceof Error ? err : new Error(String(err)));
|
|
115
|
-
}
|
|
116
|
-
}
|
|
117
|
-
|
|
118
|
-
// Fall back to LLM
|
|
119
|
-
const llmResponse = await this._provider.chat(context.messages, this.systemPrompt);
|
|
120
|
-
const response = this.createResponse(llmResponse, message);
|
|
121
|
-
await this.memory.addMessage(sessionId, response);
|
|
122
|
-
this.emit('message:out', response);
|
|
123
|
-
return response;
|
|
124
|
-
}
|
|
125
|
-
|
|
126
|
-
async *handleMessageStream(message: Message): AsyncIterable<string> {
|
|
127
|
-
const sessionId = (message.metadata?.sessionId as string) ?? 'default';
|
|
128
|
-
await this.memory.addMessage(sessionId, message);
|
|
129
|
-
|
|
130
|
-
const history = (await this.memory.getConversation(sessionId)).slice(-this.historyLimit);
|
|
131
|
-
|
|
132
|
-
let fullResponse = '';
|
|
133
|
-
for await (const chunk of this._provider.chatStream(history, this.systemPrompt)) {
|
|
134
|
-
fullResponse += chunk;
|
|
135
|
-
yield chunk;
|
|
136
|
-
}
|
|
137
|
-
|
|
138
|
-
const response = this.createResponse(fullResponse, message);
|
|
139
|
-
await this.memory.addMessage(sessionId, response);
|
|
140
|
-
this.emit('message:out', response);
|
|
141
|
-
}
|
|
142
|
-
|
|
143
|
-
private createResponse(content: string, inReplyTo: Message): Message {
|
|
144
|
-
return {
|
|
145
|
-
id: `msg_${Date.now()}_${Math.random().toString(36).slice(2, 8)}`,
|
|
146
|
-
role: 'assistant',
|
|
147
|
-
content,
|
|
148
|
-
timestamp: Date.now(),
|
|
149
|
-
metadata: { inReplyTo: inReplyTo.id },
|
|
150
|
-
};
|
|
151
|
-
}
|
|
152
|
-
}
|
|
1
|
+
import { EventEmitter } from 'events';
|
|
2
|
+
import type { AgentState, IAgent, IChannel, ISkill, Message, MemoryStore, AgentContext } from './types';
|
|
3
|
+
import { InMemoryStore } from '../memory';
|
|
4
|
+
import { createProvider, type LLMProvider } from '../providers';
|
|
5
|
+
|
|
6
|
+
export class BaseAgent extends EventEmitter implements IAgent {
|
|
7
|
+
readonly name: string;
|
|
8
|
+
private _state: AgentState = 'init';
|
|
9
|
+
private skills: Map<string, ISkill> = new Map();
|
|
10
|
+
private channels: IChannel[] = [];
|
|
11
|
+
private memory: MemoryStore;
|
|
12
|
+
private _provider: LLMProvider;
|
|
13
|
+
private systemPrompt: string;
|
|
14
|
+
private historyLimit: number;
|
|
15
|
+
|
|
16
|
+
constructor(options: {
|
|
17
|
+
name: string;
|
|
18
|
+
systemPrompt?: string;
|
|
19
|
+
provider?: string;
|
|
20
|
+
model?: string;
|
|
21
|
+
memory?: MemoryStore;
|
|
22
|
+
historyLimit?: number;
|
|
23
|
+
}) {
|
|
24
|
+
super();
|
|
25
|
+
this.name = options.name;
|
|
26
|
+
this.systemPrompt = options.systemPrompt ?? 'You are a helpful AI agent.';
|
|
27
|
+
this.memory = options.memory ?? new InMemoryStore();
|
|
28
|
+
this._provider = createProvider(options.provider ?? 'openai', options.model);
|
|
29
|
+
this.historyLimit = options.historyLimit ?? 50;
|
|
30
|
+
}
|
|
31
|
+
|
|
32
|
+
get state(): AgentState {
|
|
33
|
+
return this._state;
|
|
34
|
+
}
|
|
35
|
+
|
|
36
|
+
get provider(): LLMProvider {
|
|
37
|
+
return this._provider;
|
|
38
|
+
}
|
|
39
|
+
|
|
40
|
+
getSystemPrompt(): string {
|
|
41
|
+
return this.systemPrompt;
|
|
42
|
+
}
|
|
43
|
+
|
|
44
|
+
getMemory(): MemoryStore {
|
|
45
|
+
return this.memory;
|
|
46
|
+
}
|
|
47
|
+
|
|
48
|
+
private transition(to: AgentState): void {
|
|
49
|
+
const from = this._state;
|
|
50
|
+
this._state = to;
|
|
51
|
+
this.emit('state:change', from, to);
|
|
52
|
+
}
|
|
53
|
+
|
|
54
|
+
async init(): Promise<void> {
|
|
55
|
+
this.transition('ready');
|
|
56
|
+
}
|
|
57
|
+
|
|
58
|
+
async start(): Promise<void> {
|
|
59
|
+
if (this._state !== 'ready') {
|
|
60
|
+
throw new Error(`Cannot start agent in state: ${this._state}`);
|
|
61
|
+
}
|
|
62
|
+
for (const channel of this.channels) {
|
|
63
|
+
channel.onMessage((msg) => this.handleMessage(msg));
|
|
64
|
+
await channel.start();
|
|
65
|
+
}
|
|
66
|
+
this.transition('running');
|
|
67
|
+
}
|
|
68
|
+
|
|
69
|
+
async stop(): Promise<void> {
|
|
70
|
+
for (const channel of this.channels) {
|
|
71
|
+
await channel.stop();
|
|
72
|
+
}
|
|
73
|
+
this.transition('stopped');
|
|
74
|
+
}
|
|
75
|
+
|
|
76
|
+
registerSkill(skill: ISkill): void {
|
|
77
|
+
this.skills.set(skill.name, skill);
|
|
78
|
+
}
|
|
79
|
+
|
|
80
|
+
bindChannel(channel: IChannel): void {
|
|
81
|
+
this.channels.push(channel);
|
|
82
|
+
}
|
|
83
|
+
|
|
84
|
+
getChannels(): IChannel[] {
|
|
85
|
+
return this.channels;
|
|
86
|
+
}
|
|
87
|
+
|
|
88
|
+
async handleMessage(message: Message): Promise<Message> {
|
|
89
|
+
this.emit('message:in', message);
|
|
90
|
+
|
|
91
|
+
const sessionId = (message.metadata?.sessionId as string) ?? 'default';
|
|
92
|
+
await this.memory.addMessage(sessionId, message);
|
|
93
|
+
|
|
94
|
+
const context: AgentContext = {
|
|
95
|
+
agentName: this.name,
|
|
96
|
+
sessionId,
|
|
97
|
+
messages: (await this.memory.getConversation(sessionId)).slice(-this.historyLimit),
|
|
98
|
+
memory: this.memory,
|
|
99
|
+
metadata: {},
|
|
100
|
+
};
|
|
101
|
+
|
|
102
|
+
// Try skills first
|
|
103
|
+
for (const [name, skill] of this.skills) {
|
|
104
|
+
try {
|
|
105
|
+
const result = await skill.execute(context, message);
|
|
106
|
+
this.emit('skill:execute', name, result);
|
|
107
|
+
if (result.handled && result.response) {
|
|
108
|
+
const response = this.createResponse(result.response, message);
|
|
109
|
+
await this.memory.addMessage(sessionId, response);
|
|
110
|
+
this.emit('message:out', response);
|
|
111
|
+
return response;
|
|
112
|
+
}
|
|
113
|
+
} catch (err) {
|
|
114
|
+
this.emit('error', err instanceof Error ? err : new Error(String(err)));
|
|
115
|
+
}
|
|
116
|
+
}
|
|
117
|
+
|
|
118
|
+
// Fall back to LLM
|
|
119
|
+
const llmResponse = await this._provider.chat(context.messages, this.systemPrompt);
|
|
120
|
+
const response = this.createResponse(llmResponse, message);
|
|
121
|
+
await this.memory.addMessage(sessionId, response);
|
|
122
|
+
this.emit('message:out', response);
|
|
123
|
+
return response;
|
|
124
|
+
}
|
|
125
|
+
|
|
126
|
+
async *handleMessageStream(message: Message): AsyncIterable<string> {
|
|
127
|
+
const sessionId = (message.metadata?.sessionId as string) ?? 'default';
|
|
128
|
+
await this.memory.addMessage(sessionId, message);
|
|
129
|
+
|
|
130
|
+
const history = (await this.memory.getConversation(sessionId)).slice(-this.historyLimit);
|
|
131
|
+
|
|
132
|
+
let fullResponse = '';
|
|
133
|
+
for await (const chunk of this._provider.chatStream(history, this.systemPrompt)) {
|
|
134
|
+
fullResponse += chunk;
|
|
135
|
+
yield chunk;
|
|
136
|
+
}
|
|
137
|
+
|
|
138
|
+
const response = this.createResponse(fullResponse, message);
|
|
139
|
+
await this.memory.addMessage(sessionId, response);
|
|
140
|
+
this.emit('message:out', response);
|
|
141
|
+
}
|
|
142
|
+
|
|
143
|
+
private createResponse(content: string, inReplyTo: Message): Message {
|
|
144
|
+
return {
|
|
145
|
+
id: `msg_${Date.now()}_${Math.random().toString(36).slice(2, 8)}`,
|
|
146
|
+
role: 'assistant',
|
|
147
|
+
content,
|
|
148
|
+
timestamp: Date.now(),
|
|
149
|
+
metadata: { inReplyTo: inReplyTo.id },
|
|
150
|
+
};
|
|
151
|
+
}
|
|
152
|
+
}
|
|
@@ -1,186 +1,186 @@
|
|
|
1
|
-
/**
|
|
2
|
-
* Analytics Engine - Persistent analytics with JSON file storage.
|
|
3
|
-
* Tracks every message, LLM call, tool use, and error with timestamps.
|
|
4
|
-
*/
|
|
5
|
-
import * as fs from 'fs';
|
|
6
|
-
import * as path from 'path';
|
|
7
|
-
|
|
8
|
-
export interface AnalyticsEvent {
|
|
9
|
-
type: 'message' | 'llm_call' | 'tool_use' | 'error';
|
|
10
|
-
timestamp: number;
|
|
11
|
-
data: Record<string, any>;
|
|
12
|
-
}
|
|
13
|
-
|
|
14
|
-
export interface AnalyticsStats {
|
|
15
|
-
totalMessages: number;
|
|
16
|
-
totalLLMCalls: number;
|
|
17
|
-
totalToolUses: number;
|
|
18
|
-
totalErrors: number;
|
|
19
|
-
avgResponseTimeMs: number;
|
|
20
|
-
totalTokens: { input: number; output: number; total: number };
|
|
21
|
-
topSkills: { name: string; count: number }[];
|
|
22
|
-
topErrors: { message: string; count: number }[];
|
|
23
|
-
messagesPerDay: Record<string, number>;
|
|
24
|
-
period: { from: number; to: number };
|
|
25
|
-
}
|
|
26
|
-
|
|
27
|
-
export class AnalyticsEngine {
|
|
28
|
-
private dataDir: string;
|
|
29
|
-
private eventsFile: string;
|
|
30
|
-
private events: AnalyticsEvent[] = [];
|
|
31
|
-
|
|
32
|
-
constructor(dataDir: string = '.') {
|
|
33
|
-
this.dataDir = path.resolve(dataDir, 'data');
|
|
34
|
-
this.eventsFile = path.join(this.dataDir, 'analytics.json');
|
|
35
|
-
this.load();
|
|
36
|
-
}
|
|
37
|
-
|
|
38
|
-
private load(): void {
|
|
39
|
-
try {
|
|
40
|
-
if (fs.existsSync(this.eventsFile)) {
|
|
41
|
-
const raw = fs.readFileSync(this.eventsFile, 'utf-8');
|
|
42
|
-
this.events = JSON.parse(raw);
|
|
43
|
-
}
|
|
44
|
-
} catch {
|
|
45
|
-
this.events = [];
|
|
46
|
-
}
|
|
47
|
-
}
|
|
48
|
-
|
|
49
|
-
private save(): void {
|
|
50
|
-
if (!fs.existsSync(this.dataDir)) {
|
|
51
|
-
fs.mkdirSync(this.dataDir, { recursive: true });
|
|
52
|
-
}
|
|
53
|
-
// Keep last 10000 events to prevent unbounded growth
|
|
54
|
-
if (this.events.length > 10000) {
|
|
55
|
-
this.events = this.events.slice(-10000);
|
|
56
|
-
}
|
|
57
|
-
fs.writeFileSync(this.eventsFile, JSON.stringify(this.events, null, 2));
|
|
58
|
-
}
|
|
59
|
-
|
|
60
|
-
track(type: AnalyticsEvent['type'], data: Record<string, any>): void {
|
|
61
|
-
this.events.push({ type, timestamp: Date.now(), data });
|
|
62
|
-
this.save();
|
|
63
|
-
}
|
|
64
|
-
|
|
65
|
-
trackMessage(userId: string, responseTimeMs: number, tokensIn: number, tokensOut: number): void {
|
|
66
|
-
this.track('message', { userId, responseTimeMs, tokensIn, tokensOut });
|
|
67
|
-
}
|
|
68
|
-
|
|
69
|
-
trackLLMCall(provider: string, model: string, tokensIn: number, tokensOut: number, latencyMs: number): void {
|
|
70
|
-
this.track('llm_call', { provider, model, tokensIn, tokensOut, latencyMs });
|
|
71
|
-
}
|
|
72
|
-
|
|
73
|
-
trackToolUse(toolName: string, success: boolean, latencyMs: number): void {
|
|
74
|
-
this.track('tool_use', { toolName, success, latencyMs });
|
|
75
|
-
}
|
|
76
|
-
|
|
77
|
-
trackError(error: string, context?: string): void {
|
|
78
|
-
this.track('error', { error, context });
|
|
79
|
-
}
|
|
80
|
-
|
|
81
|
-
getStats(fromTs?: number, toTs?: number): AnalyticsStats {
|
|
82
|
-
const now = Date.now();
|
|
83
|
-
const from = fromTs ?? 0;
|
|
84
|
-
const to = toTs ?? now;
|
|
85
|
-
const filtered = this.events.filter(e => e.timestamp >= from && e.timestamp <= to);
|
|
86
|
-
|
|
87
|
-
const messages = filtered.filter(e => e.type === 'message');
|
|
88
|
-
const llmCalls = filtered.filter(e => e.type === 'llm_call');
|
|
89
|
-
const toolUses = filtered.filter(e => e.type === 'tool_use');
|
|
90
|
-
const errors = filtered.filter(e => e.type === 'error');
|
|
91
|
-
|
|
92
|
-
// Avg response time
|
|
93
|
-
const totalResponseTime = messages.reduce((sum, e) => sum + (e.data.responseTimeMs ?? 0), 0);
|
|
94
|
-
const avgResponseTimeMs = messages.length > 0 ? Math.round(totalResponseTime / messages.length) : 0;
|
|
95
|
-
|
|
96
|
-
// Total tokens
|
|
97
|
-
const totalTokensIn = llmCalls.reduce((sum, e) => sum + (e.data.tokensIn ?? 0), 0);
|
|
98
|
-
const totalTokensOut = llmCalls.reduce((sum, e) => sum + (e.data.tokensOut ?? 0), 0);
|
|
99
|
-
|
|
100
|
-
// Top skills (from tool_use)
|
|
101
|
-
const skillCounts: Record<string, number> = {};
|
|
102
|
-
for (const e of toolUses) {
|
|
103
|
-
const name = e.data.toolName ?? 'unknown';
|
|
104
|
-
skillCounts[name] = (skillCounts[name] ?? 0) + 1;
|
|
105
|
-
}
|
|
106
|
-
const topSkills = Object.entries(skillCounts)
|
|
107
|
-
.sort((a, b) => b[1] - a[1])
|
|
108
|
-
.slice(0, 10)
|
|
109
|
-
.map(([name, count]) => ({ name, count }));
|
|
110
|
-
|
|
111
|
-
// Top errors
|
|
112
|
-
const errorCounts: Record<string, number> = {};
|
|
113
|
-
for (const e of errors) {
|
|
114
|
-
const msg = e.data.error ?? 'unknown';
|
|
115
|
-
errorCounts[msg] = (errorCounts[msg] ?? 0) + 1;
|
|
116
|
-
}
|
|
117
|
-
const topErrors = Object.entries(errorCounts)
|
|
118
|
-
.sort((a, b) => b[1] - a[1])
|
|
119
|
-
.slice(0, 10)
|
|
120
|
-
.map(([message, count]) => ({ message, count }));
|
|
121
|
-
|
|
122
|
-
// Messages per day
|
|
123
|
-
const messagesPerDay: Record<string, number> = {};
|
|
124
|
-
for (const e of messages) {
|
|
125
|
-
const day = new Date(e.timestamp).toISOString().slice(0, 10);
|
|
126
|
-
messagesPerDay[day] = (messagesPerDay[day] ?? 0) + 1;
|
|
127
|
-
}
|
|
128
|
-
|
|
129
|
-
return {
|
|
130
|
-
totalMessages: messages.length,
|
|
131
|
-
totalLLMCalls: llmCalls.length,
|
|
132
|
-
totalToolUses: toolUses.length,
|
|
133
|
-
totalErrors: errors.length,
|
|
134
|
-
avgResponseTimeMs,
|
|
135
|
-
totalTokens: { input: totalTokensIn, output: totalTokensOut, total: totalTokensIn + totalTokensOut },
|
|
136
|
-
topSkills,
|
|
137
|
-
topErrors,
|
|
138
|
-
messagesPerDay,
|
|
139
|
-
period: { from, to },
|
|
140
|
-
};
|
|
141
|
-
}
|
|
142
|
-
|
|
143
|
-
getRecentEvents(limit: number = 50): AnalyticsEvent[] {
|
|
144
|
-
return this.events.slice(-limit);
|
|
145
|
-
}
|
|
146
|
-
|
|
147
|
-
clear(): void {
|
|
148
|
-
this.events = [];
|
|
149
|
-
this.save();
|
|
150
|
-
}
|
|
151
|
-
|
|
152
|
-
/**
|
|
153
|
-
* Format stats for CLI display.
|
|
154
|
-
*/
|
|
155
|
-
static formatStats(stats: AnalyticsStats): string {
|
|
156
|
-
const lines: string[] = [];
|
|
157
|
-
lines.push('');
|
|
158
|
-
lines.push('══════════════════════════════════════════');
|
|
159
|
-
lines.push(' OPC Agent Analytics');
|
|
160
|
-
lines.push('══════════════════════════════════════════');
|
|
161
|
-
lines.push('');
|
|
162
|
-
lines.push(` 📨 Messages: ${stats.totalMessages}`);
|
|
163
|
-
lines.push(` 🤖 LLM Calls: ${stats.totalLLMCalls}`);
|
|
164
|
-
lines.push(` 🔧 Tool Uses: ${stats.totalToolUses}`);
|
|
165
|
-
lines.push(` ❌ Errors: ${stats.totalErrors}`);
|
|
166
|
-
lines.push(` ⏱ Avg Response: ${stats.avgResponseTimeMs}ms`);
|
|
167
|
-
lines.push(` 🪙 Tokens: ${stats.totalTokens.total} (in: ${stats.totalTokens.input}, out: ${stats.totalTokens.output})`);
|
|
168
|
-
lines.push('');
|
|
169
|
-
if (stats.topSkills.length > 0) {
|
|
170
|
-
lines.push(' Top Skills:');
|
|
171
|
-
for (const s of stats.topSkills.slice(0, 5)) {
|
|
172
|
-
lines.push(` • ${s.name}: ${s.count}`);
|
|
173
|
-
}
|
|
174
|
-
lines.push('');
|
|
175
|
-
}
|
|
176
|
-
if (stats.topErrors.length > 0) {
|
|
177
|
-
lines.push(' Top Errors:');
|
|
178
|
-
for (const e of stats.topErrors.slice(0, 3)) {
|
|
179
|
-
lines.push(` • ${e.message}: ${e.count}`);
|
|
180
|
-
}
|
|
181
|
-
lines.push('');
|
|
182
|
-
}
|
|
183
|
-
lines.push('──────────────────────────────────────────');
|
|
184
|
-
return lines.join('\n');
|
|
185
|
-
}
|
|
186
|
-
}
|
|
1
|
+
/**
|
|
2
|
+
* Analytics Engine - Persistent analytics with JSON file storage.
|
|
3
|
+
* Tracks every message, LLM call, tool use, and error with timestamps.
|
|
4
|
+
*/
|
|
5
|
+
import * as fs from 'fs';
|
|
6
|
+
import * as path from 'path';
|
|
7
|
+
|
|
8
|
+
export interface AnalyticsEvent {
|
|
9
|
+
type: 'message' | 'llm_call' | 'tool_use' | 'error';
|
|
10
|
+
timestamp: number;
|
|
11
|
+
data: Record<string, any>;
|
|
12
|
+
}
|
|
13
|
+
|
|
14
|
+
export interface AnalyticsStats {
|
|
15
|
+
totalMessages: number;
|
|
16
|
+
totalLLMCalls: number;
|
|
17
|
+
totalToolUses: number;
|
|
18
|
+
totalErrors: number;
|
|
19
|
+
avgResponseTimeMs: number;
|
|
20
|
+
totalTokens: { input: number; output: number; total: number };
|
|
21
|
+
topSkills: { name: string; count: number }[];
|
|
22
|
+
topErrors: { message: string; count: number }[];
|
|
23
|
+
messagesPerDay: Record<string, number>;
|
|
24
|
+
period: { from: number; to: number };
|
|
25
|
+
}
|
|
26
|
+
|
|
27
|
+
export class AnalyticsEngine {
|
|
28
|
+
private dataDir: string;
|
|
29
|
+
private eventsFile: string;
|
|
30
|
+
private events: AnalyticsEvent[] = [];
|
|
31
|
+
|
|
32
|
+
constructor(dataDir: string = '.') {
|
|
33
|
+
this.dataDir = path.resolve(dataDir, 'data');
|
|
34
|
+
this.eventsFile = path.join(this.dataDir, 'analytics.json');
|
|
35
|
+
this.load();
|
|
36
|
+
}
|
|
37
|
+
|
|
38
|
+
private load(): void {
|
|
39
|
+
try {
|
|
40
|
+
if (fs.existsSync(this.eventsFile)) {
|
|
41
|
+
const raw = fs.readFileSync(this.eventsFile, 'utf-8');
|
|
42
|
+
this.events = JSON.parse(raw);
|
|
43
|
+
}
|
|
44
|
+
} catch {
|
|
45
|
+
this.events = [];
|
|
46
|
+
}
|
|
47
|
+
}
|
|
48
|
+
|
|
49
|
+
private save(): void {
|
|
50
|
+
if (!fs.existsSync(this.dataDir)) {
|
|
51
|
+
fs.mkdirSync(this.dataDir, { recursive: true });
|
|
52
|
+
}
|
|
53
|
+
// Keep last 10000 events to prevent unbounded growth
|
|
54
|
+
if (this.events.length > 10000) {
|
|
55
|
+
this.events = this.events.slice(-10000);
|
|
56
|
+
}
|
|
57
|
+
fs.writeFileSync(this.eventsFile, JSON.stringify(this.events, null, 2));
|
|
58
|
+
}
|
|
59
|
+
|
|
60
|
+
track(type: AnalyticsEvent['type'], data: Record<string, any>): void {
|
|
61
|
+
this.events.push({ type, timestamp: Date.now(), data });
|
|
62
|
+
this.save();
|
|
63
|
+
}
|
|
64
|
+
|
|
65
|
+
trackMessage(userId: string, responseTimeMs: number, tokensIn: number, tokensOut: number): void {
|
|
66
|
+
this.track('message', { userId, responseTimeMs, tokensIn, tokensOut });
|
|
67
|
+
}
|
|
68
|
+
|
|
69
|
+
trackLLMCall(provider: string, model: string, tokensIn: number, tokensOut: number, latencyMs: number): void {
|
|
70
|
+
this.track('llm_call', { provider, model, tokensIn, tokensOut, latencyMs });
|
|
71
|
+
}
|
|
72
|
+
|
|
73
|
+
trackToolUse(toolName: string, success: boolean, latencyMs: number): void {
|
|
74
|
+
this.track('tool_use', { toolName, success, latencyMs });
|
|
75
|
+
}
|
|
76
|
+
|
|
77
|
+
trackError(error: string, context?: string): void {
|
|
78
|
+
this.track('error', { error, context });
|
|
79
|
+
}
|
|
80
|
+
|
|
81
|
+
getStats(fromTs?: number, toTs?: number): AnalyticsStats {
|
|
82
|
+
const now = Date.now();
|
|
83
|
+
const from = fromTs ?? 0;
|
|
84
|
+
const to = toTs ?? now;
|
|
85
|
+
const filtered = this.events.filter(e => e.timestamp >= from && e.timestamp <= to);
|
|
86
|
+
|
|
87
|
+
const messages = filtered.filter(e => e.type === 'message');
|
|
88
|
+
const llmCalls = filtered.filter(e => e.type === 'llm_call');
|
|
89
|
+
const toolUses = filtered.filter(e => e.type === 'tool_use');
|
|
90
|
+
const errors = filtered.filter(e => e.type === 'error');
|
|
91
|
+
|
|
92
|
+
// Avg response time
|
|
93
|
+
const totalResponseTime = messages.reduce((sum, e) => sum + (e.data.responseTimeMs ?? 0), 0);
|
|
94
|
+
const avgResponseTimeMs = messages.length > 0 ? Math.round(totalResponseTime / messages.length) : 0;
|
|
95
|
+
|
|
96
|
+
// Total tokens
|
|
97
|
+
const totalTokensIn = llmCalls.reduce((sum, e) => sum + (e.data.tokensIn ?? 0), 0);
|
|
98
|
+
const totalTokensOut = llmCalls.reduce((sum, e) => sum + (e.data.tokensOut ?? 0), 0);
|
|
99
|
+
|
|
100
|
+
// Top skills (from tool_use)
|
|
101
|
+
const skillCounts: Record<string, number> = {};
|
|
102
|
+
for (const e of toolUses) {
|
|
103
|
+
const name = e.data.toolName ?? 'unknown';
|
|
104
|
+
skillCounts[name] = (skillCounts[name] ?? 0) + 1;
|
|
105
|
+
}
|
|
106
|
+
const topSkills = Object.entries(skillCounts)
|
|
107
|
+
.sort((a, b) => b[1] - a[1])
|
|
108
|
+
.slice(0, 10)
|
|
109
|
+
.map(([name, count]) => ({ name, count }));
|
|
110
|
+
|
|
111
|
+
// Top errors
|
|
112
|
+
const errorCounts: Record<string, number> = {};
|
|
113
|
+
for (const e of errors) {
|
|
114
|
+
const msg = e.data.error ?? 'unknown';
|
|
115
|
+
errorCounts[msg] = (errorCounts[msg] ?? 0) + 1;
|
|
116
|
+
}
|
|
117
|
+
const topErrors = Object.entries(errorCounts)
|
|
118
|
+
.sort((a, b) => b[1] - a[1])
|
|
119
|
+
.slice(0, 10)
|
|
120
|
+
.map(([message, count]) => ({ message, count }));
|
|
121
|
+
|
|
122
|
+
// Messages per day
|
|
123
|
+
const messagesPerDay: Record<string, number> = {};
|
|
124
|
+
for (const e of messages) {
|
|
125
|
+
const day = new Date(e.timestamp).toISOString().slice(0, 10);
|
|
126
|
+
messagesPerDay[day] = (messagesPerDay[day] ?? 0) + 1;
|
|
127
|
+
}
|
|
128
|
+
|
|
129
|
+
return {
|
|
130
|
+
totalMessages: messages.length,
|
|
131
|
+
totalLLMCalls: llmCalls.length,
|
|
132
|
+
totalToolUses: toolUses.length,
|
|
133
|
+
totalErrors: errors.length,
|
|
134
|
+
avgResponseTimeMs,
|
|
135
|
+
totalTokens: { input: totalTokensIn, output: totalTokensOut, total: totalTokensIn + totalTokensOut },
|
|
136
|
+
topSkills,
|
|
137
|
+
topErrors,
|
|
138
|
+
messagesPerDay,
|
|
139
|
+
period: { from, to },
|
|
140
|
+
};
|
|
141
|
+
}
|
|
142
|
+
|
|
143
|
+
getRecentEvents(limit: number = 50): AnalyticsEvent[] {
|
|
144
|
+
return this.events.slice(-limit);
|
|
145
|
+
}
|
|
146
|
+
|
|
147
|
+
clear(): void {
|
|
148
|
+
this.events = [];
|
|
149
|
+
this.save();
|
|
150
|
+
}
|
|
151
|
+
|
|
152
|
+
/**
|
|
153
|
+
* Format stats for CLI display.
|
|
154
|
+
*/
|
|
155
|
+
static formatStats(stats: AnalyticsStats): string {
|
|
156
|
+
const lines: string[] = [];
|
|
157
|
+
lines.push('');
|
|
158
|
+
lines.push('══════════════════════════════════════════');
|
|
159
|
+
lines.push(' OPC Agent Analytics');
|
|
160
|
+
lines.push('══════════════════════════════════════════');
|
|
161
|
+
lines.push('');
|
|
162
|
+
lines.push(` 📨 Messages: ${stats.totalMessages}`);
|
|
163
|
+
lines.push(` 🤖 LLM Calls: ${stats.totalLLMCalls}`);
|
|
164
|
+
lines.push(` 🔧 Tool Uses: ${stats.totalToolUses}`);
|
|
165
|
+
lines.push(` ❌ Errors: ${stats.totalErrors}`);
|
|
166
|
+
lines.push(` ⏱ Avg Response: ${stats.avgResponseTimeMs}ms`);
|
|
167
|
+
lines.push(` 🪙 Tokens: ${stats.totalTokens.total} (in: ${stats.totalTokens.input}, out: ${stats.totalTokens.output})`);
|
|
168
|
+
lines.push('');
|
|
169
|
+
if (stats.topSkills.length > 0) {
|
|
170
|
+
lines.push(' Top Skills:');
|
|
171
|
+
for (const s of stats.topSkills.slice(0, 5)) {
|
|
172
|
+
lines.push(` • ${s.name}: ${s.count}`);
|
|
173
|
+
}
|
|
174
|
+
lines.push('');
|
|
175
|
+
}
|
|
176
|
+
if (stats.topErrors.length > 0) {
|
|
177
|
+
lines.push(' Top Errors:');
|
|
178
|
+
for (const e of stats.topErrors.slice(0, 3)) {
|
|
179
|
+
lines.push(` • ${e.message}: ${e.count}`);
|
|
180
|
+
}
|
|
181
|
+
lines.push('');
|
|
182
|
+
}
|
|
183
|
+
lines.push('──────────────────────────────────────────');
|
|
184
|
+
return lines.join('\n');
|
|
185
|
+
}
|
|
186
|
+
}
|