@donggui/core 1.5.14 → 1.6.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (36) hide show
  1. package/dist/es/agent/agent.mjs +187 -10
  2. package/dist/es/agent/agent.mjs.map +1 -1
  3. package/dist/es/agent/task-builder.mjs +46 -2
  4. package/dist/es/agent/task-builder.mjs.map +1 -1
  5. package/dist/es/agent/utils.mjs +1 -1
  6. package/dist/es/ai-model/assert.mjs +415 -0
  7. package/dist/es/ai-model/assert.mjs.map +1 -0
  8. package/dist/es/ai-model/index.mjs +2 -1
  9. package/dist/es/ai-model/prompt/cache-verify.mjs +34 -0
  10. package/dist/es/ai-model/prompt/cache-verify.mjs.map +1 -0
  11. package/dist/es/service/index.mjs +81 -6
  12. package/dist/es/service/index.mjs.map +1 -1
  13. package/dist/es/types.mjs.map +1 -1
  14. package/dist/es/utils.mjs +2 -2
  15. package/dist/lib/agent/agent.js +185 -8
  16. package/dist/lib/agent/agent.js.map +1 -1
  17. package/dist/lib/agent/task-builder.js +46 -2
  18. package/dist/lib/agent/task-builder.js.map +1 -1
  19. package/dist/lib/agent/utils.js +1 -1
  20. package/dist/lib/ai-model/assert.js +455 -0
  21. package/dist/lib/ai-model/assert.js.map +1 -0
  22. package/dist/lib/ai-model/index.js +18 -11
  23. package/dist/lib/ai-model/prompt/cache-verify.js +68 -0
  24. package/dist/lib/ai-model/prompt/cache-verify.js.map +1 -0
  25. package/dist/lib/service/index.js +81 -6
  26. package/dist/lib/service/index.js.map +1 -1
  27. package/dist/lib/types.js +3 -3
  28. package/dist/lib/types.js.map +1 -1
  29. package/dist/lib/utils.js +2 -2
  30. package/dist/types/agent/agent.d.ts +38 -2
  31. package/dist/types/ai-model/assert.d.ts +66 -0
  32. package/dist/types/ai-model/index.d.ts +2 -0
  33. package/dist/types/ai-model/prompt/cache-verify.d.ts +1 -0
  34. package/dist/types/service/index.d.ts +3 -2
  35. package/dist/types/types.d.ts +166 -0
  36. package/package.json +25 -44
@@ -152,7 +152,7 @@ async function matchElementFromCache(context, cacheEntry, cachePrompt, cacheable
152
152
  return;
153
153
  }
154
154
  }
155
- const getMidsceneVersion = ()=>"1.5.14";
155
+ const getMidsceneVersion = ()=>"1.6.0";
156
156
  const parsePrompt = (prompt)=>{
157
157
  if ('string' == typeof prompt) return {
158
158
  textPrompt: prompt,
@@ -0,0 +1,415 @@
1
+ import { getDebug } from "@midscene/shared/logger";
2
+ import { callAIWithObjectResponse } from "./service-caller/index.mjs";
3
+ const debug = getDebug('ai:assert');
4
+ const DEFAULT_SYSTEM_CHECK_PROMPT = `请检查以下系统级问题:
5
+
6
+ 1. **白屏检测**:页面是否完全白屏或大面积空白
7
+ 2. **布局遮挡**:重要元素是否存在组件间重叠遮挡
8
+ 3. **未加载内容**:页面是否有加载中的占位符、骨架屏
9
+ 4. **错误提示**:页面是否显示错误信息、异常提示
10
+ 5. **后端错误**:是否有后端请求失败的提示,如"活动太火爆了"、报错 Toast、错误码 等
11
+
12
+ 如果发现以上问题,请在 systemCheckResults 中标注对应字段为 true。`;
13
+ function buildAssertionPrompt(options) {
14
+ const { assertion, businessContext, systemCheckPrompt } = options;
15
+ let prompt = `## 用户断言描述
16
+ ${assertion}`;
17
+ if (businessContext) prompt += `
18
+
19
+ ## 业务知识上下文
20
+ ${businessContext}`;
21
+ if (systemCheckPrompt) prompt += `
22
+
23
+ ## 系统校验规则
24
+ ${systemCheckPrompt}`;
25
+ prompt += `
26
+
27
+ ## 输出要求
28
+ 请以 JSON 格式输出断言结果:
29
+ {
30
+ "pass": boolean, // 断言是否通过
31
+ "thought": string, // 思考过程
32
+ "reason": string, // 失败原因(如果失败)
33
+ "systemCheckResults": { // 系统校验结果(可选)
34
+ "whiteScreen": boolean,
35
+ "layoutBlocked": boolean,
36
+ "loadingContent": boolean,
37
+ "errorPrompt": boolean,
38
+ "backendError": boolean
39
+ }
40
+ }`;
41
+ return prompt;
42
+ }
43
+ async function AiAssertElement(options) {
44
+ const { beforeScreenshot, afterScreenshot, assertion, businessContext, enableSystemCheck = true, customSystemCheckRules, modelConfig, abortSignal } = options;
45
+ const systemPrompt = `你是一个自动化测试断言专家。请根据以下信息判断断言是否通过。
46
+
47
+ 你需要:
48
+ 1. 分析操作前后的截图变化(如果提供了操作前截图)
49
+ 2. 验证用户的断言描述是否成立
50
+ 3. 检查是否存在系统级问题(如果启用)
51
+ 4. 给出详细的思考过程和判断结果`;
52
+ const systemCheckPrompt = enableSystemCheck ? customSystemCheckRules || DEFAULT_SYSTEM_CHECK_PROMPT : '';
53
+ const userContent = [];
54
+ if (beforeScreenshot) {
55
+ userContent.push({
56
+ type: 'text',
57
+ text: '## 操作前截图(执行操作前的页面状态)'
58
+ });
59
+ userContent.push({
60
+ type: 'image_url',
61
+ image_url: {
62
+ url: beforeScreenshot,
63
+ detail: 'high'
64
+ }
65
+ });
66
+ userContent.push({
67
+ type: 'text',
68
+ text: '## 操作后截图(执行操作后的页面状态)'
69
+ });
70
+ } else userContent.push({
71
+ type: 'text',
72
+ text: '## 当前页面截图'
73
+ });
74
+ userContent.push({
75
+ type: 'image_url',
76
+ image_url: {
77
+ url: afterScreenshot,
78
+ detail: 'high'
79
+ }
80
+ });
81
+ userContent.push({
82
+ type: 'text',
83
+ text: buildAssertionPrompt({
84
+ assertion,
85
+ businessContext,
86
+ systemCheckPrompt
87
+ })
88
+ });
89
+ const msgs = [
90
+ {
91
+ role: 'system',
92
+ content: systemPrompt
93
+ },
94
+ {
95
+ role: 'user',
96
+ content: userContent
97
+ }
98
+ ];
99
+ debug('calling AI for assertion:', assertion);
100
+ try {
101
+ const result = await callAIWithObjectResponse(msgs, modelConfig, {
102
+ abortSignal
103
+ });
104
+ debug('assertion result:', result.content);
105
+ return {
106
+ pass: result.content.pass,
107
+ thought: result.content.thought || '',
108
+ reason: result.content.reason,
109
+ usage: result.usage,
110
+ systemCheckResults: result.content.systemCheckResults,
111
+ rawResponse: JSON.stringify(result.content)
112
+ };
113
+ } catch (error) {
114
+ debug('assertion error:', error);
115
+ throw error;
116
+ }
117
+ }
118
+ const DIFF_SYSTEM_PROMPT = `你是一个自动化测试的图像对比专家。你的任务是对比当前页面截图与基准图片,判断页面样式是否符合预期。
119
+
120
+ 你需要:
121
+ 1. 仔细对比两张图片的布局、颜色、元素位置
122
+ 2. 分析差异是否在可接受范围内
123
+ 3. 识别动态内容(如时间、日期)导致的差异
124
+ 4. 给出详细的对比分析和判断结果`;
125
+ function buildDiffPrompt(options) {
126
+ const { assertion, businessContext, diffThreshold = 0.1, ignoreRegions, ignoreDynamicContent, strictMode } = options;
127
+ let prompt = `## 图像对比任务
128
+
129
+ 请对比基准图片(预期样式)和当前截图(实际样式),判断页面是否符合预期。
130
+
131
+ ## 用户断言描述
132
+ ${assertion}`;
133
+ if (businessContext) prompt += `
134
+
135
+ ## 业务知识上下文
136
+ ${businessContext}`;
137
+ prompt += `
138
+
139
+ ## 对比要求
140
+
141
+ 请从以下维度进行对比分析:
142
+
143
+ 1. **布局一致性**:页面布局是否与基准图片一致
144
+ 2. **颜色一致性**:主要颜色是否与基准图片一致
145
+ 3. **元素位置**:关键元素的位置是否与基准图片一致
146
+ 4. **文字内容**:文字内容是否与基准图片一致
147
+ 5. **图片资源**:图片、图标等资源是否正确加载
148
+
149
+ ## 差异阈值
150
+ 允许的差异阈值:${100 * diffThreshold}%`;
151
+ if (ignoreRegions && ignoreRegions.length > 0) {
152
+ prompt += `
153
+
154
+ ## 忽略对比的区域
155
+ 以下区域不参与对比:`;
156
+ ignoreRegions.forEach((region, index)=>{
157
+ prompt += `
158
+ ${index + 1}. 位置 (${region.x}, ${region.y}),尺寸 ${region.width}x${region.height}`;
159
+ });
160
+ }
161
+ if (ignoreDynamicContent) prompt += `
162
+
163
+ ## 动态内容处理
164
+ 请忽略以下动态内容导致的差异:
165
+ - 时间显示
166
+ - 日期显示
167
+ - 随机验证码
168
+ - 动画效果`;
169
+ if (strictMode) prompt += `
170
+
171
+ ## 严格模式
172
+ 当前为严格模式,任何差异都视为失败。`;
173
+ prompt += `
174
+
175
+ ## 输出要求
176
+ 请以 JSON 格式输出对比结果:
177
+ {
178
+ "pass": boolean, // 断言是否通过
179
+ "thought": string, // 对比分析过程
180
+ "reason": string, // 失败原因(如果失败)
181
+ "diffDetails": { // 差异详情
182
+ "layoutMatch": boolean, // 布局是否匹配
183
+ "colorMatch": boolean, // 颜色是否匹配
184
+ "elementPositionMatch": boolean, // 元素位置是否匹配
185
+ "textContentMatch": boolean, // 文字内容是否匹配
186
+ "resourceMatch": boolean, // 资源是否匹配
187
+ "acceptableDifferences": string[], // 可接受的差异列表
188
+ "unacceptableDifferences": string[] // 不可接受的差异列表
189
+ }
190
+ }`;
191
+ return prompt;
192
+ }
193
+ async function AiAssertDiff(options) {
194
+ const { currentScreenshot, referenceImages, assertion, businessContext, diffThreshold, ignoreRegions, ignoreDynamicContent, strictMode, modelConfig, abortSignal } = options;
195
+ const userContent = [];
196
+ for (const refImage of referenceImages){
197
+ userContent.push({
198
+ type: 'text',
199
+ text: `## 基准图片:${refImage.name}`
200
+ });
201
+ userContent.push({
202
+ type: 'image_url',
203
+ image_url: {
204
+ url: refImage.url,
205
+ detail: 'high'
206
+ }
207
+ });
208
+ }
209
+ userContent.push({
210
+ type: 'text',
211
+ text: '## 当前截图(实际样式)'
212
+ });
213
+ userContent.push({
214
+ type: 'image_url',
215
+ image_url: {
216
+ url: currentScreenshot,
217
+ detail: 'high'
218
+ }
219
+ });
220
+ userContent.push({
221
+ type: 'text',
222
+ text: buildDiffPrompt({
223
+ assertion,
224
+ businessContext,
225
+ diffThreshold,
226
+ ignoreRegions,
227
+ ignoreDynamicContent,
228
+ strictMode
229
+ })
230
+ });
231
+ const msgs = [
232
+ {
233
+ role: 'system',
234
+ content: DIFF_SYSTEM_PROMPT
235
+ },
236
+ {
237
+ role: 'user',
238
+ content: userContent
239
+ }
240
+ ];
241
+ debug('calling AI for diff assertion:', assertion);
242
+ try {
243
+ const result = await callAIWithObjectResponse(msgs, modelConfig, {
244
+ abortSignal
245
+ });
246
+ debug('diff assertion result:', result.content);
247
+ return {
248
+ pass: result.content.pass,
249
+ thought: result.content.thought || '',
250
+ reason: result.content.reason,
251
+ diffDetails: result.content.diffDetails,
252
+ usage: result.usage,
253
+ rawResponse: JSON.stringify(result.content)
254
+ };
255
+ } catch (error) {
256
+ debug('diff assertion error:', error);
257
+ throw error;
258
+ }
259
+ }
260
+ const VIDEO_SYSTEM_PROMPT = `你是一个自动化测试的视频/动画分析专家。你的任务是对比当前录制的视频与基准视频,判断动画效果是否符合预期。
261
+
262
+ 你需要:
263
+ 1. 分析视频中的动画流畅度、时长、完整性
264
+ 2. 对比基准视频与当前视频的差异
265
+ 3. 识别动画类型(过渡动画、加载动画、交互动画等)
266
+ 4. 检测动画质量问题(卡顿、跳帧等)
267
+ 5. 给出详细的分析过程和判断结果`;
268
+ function buildVideoPrompt(options) {
269
+ const { assertion, businessContext, videoOptions } = options;
270
+ let prompt = `## 视频/动画对比任务
271
+
272
+ 请对比基准视频(预期动画效果)和当前录制的视频(实际动画效果),判断动画是否符合预期。
273
+
274
+ ## 用户断言描述
275
+ ${assertion}`;
276
+ if (businessContext) prompt += `
277
+
278
+ ## 业务知识上下文
279
+ ${businessContext}`;
280
+ prompt += `
281
+
282
+ ## 分析要求
283
+
284
+ 请从以下维度进行动画分析:
285
+
286
+ 1. **动画流畅度**:动画是否流畅,有无卡顿、跳帧
287
+ 2. **动画时长**:动画时长是否在预期范围内
288
+ 3. **动画完整性**:动画是否完整执行,有无中断
289
+ 4. **动画类型**:识别动画类型(过渡、加载、交互等)
290
+ 5. **关键帧匹配**:关键时间点的画面是否符合预期`;
291
+ if (videoOptions) {
292
+ if (videoOptions.checkSmoothness) prompt += `
293
+
294
+ ## 流畅度检测
295
+ 流畅度阈值:${videoOptions.smoothnessThreshold || 60}/100
296
+ 请评估动画流畅度并给出评分。`;
297
+ if (videoOptions.checkDuration && videoOptions.expectedDuration) prompt += `
298
+
299
+ ## 时长检测
300
+ 预期时长范围:${videoOptions.expectedDuration.min || 0}秒 - ${videoOptions.expectedDuration.max || '无限制'}秒`;
301
+ if (videoOptions.keyframes && videoOptions.keyframes.timestamps.length > 0) {
302
+ prompt += `
303
+
304
+ ## 关键帧验证
305
+ 需要在以下时间点验证画面:`;
306
+ videoOptions.keyframes.timestamps.forEach((ts, index)=>{
307
+ const desc = videoOptions.keyframes?.descriptions?.[index] || '未描述';
308
+ prompt += `
309
+ - ${ts}秒:${desc}`;
310
+ });
311
+ }
312
+ if (videoOptions.animationType && 'auto' !== videoOptions.animationType) prompt += `
313
+
314
+ ## 动画类型
315
+ 预期动画类型:${videoOptions.animationType}`;
316
+ }
317
+ prompt += `
318
+
319
+ ## 输出要求
320
+ 请以 JSON 格式输出对比结果:
321
+ {
322
+ "pass": boolean, // 断言是否通过
323
+ "thought": string, // 分析过程
324
+ "reason": string, // 失败原因(如果失败)
325
+ "videoDetails": { // 视频详情
326
+ "smoothnessScore": number, // 流畅度评分(0-100)
327
+ "duration": number, // 动画时长(秒)
328
+ "isComplete": boolean, // 动画是否完整
329
+ "keyframeMatches": [ // 关键帧匹配结果
330
+ {
331
+ "timestamp": number,
332
+ "matched": boolean,
333
+ "description": string
334
+ }
335
+ ],
336
+ "detectedAnimationType": string, // 检测到的动画类型
337
+ "qualityAssessment": { // 质量评估
338
+ "hasStuttering": boolean, // 是否有卡顿
339
+ "hasFrameDropping": boolean, // 是否有跳帧
340
+ "averageFrameInterval": number, // 平均帧间隔(ms)
341
+ "frameIntervalStdDev": number // 帧间隔标准差
342
+ },
343
+ "acceptableDifferences": string[], // 可接受的差异列表
344
+ "unacceptableDifferences": string[] // 不可接受的差异列表
345
+ }
346
+ }`;
347
+ return prompt;
348
+ }
349
+ async function AiAssertVideo(options) {
350
+ const { currentVideoFrames, assertion, businessContext, videoOptions, modelConfig, abortSignal } = options;
351
+ const MAX_DURATION = 5;
352
+ const DEFAULT_FPS = 30;
353
+ const MAX_FRAMES = MAX_DURATION * DEFAULT_FPS;
354
+ if (currentVideoFrames.length > MAX_FRAMES) throw new Error(`Video frames exceed maximum limit. Maximum allowed: ${MAX_FRAMES} frames (${MAX_DURATION}s at ${DEFAULT_FPS}fps), got: ${currentVideoFrames.length} frames`);
355
+ const userContent = [];
356
+ userContent.push({
357
+ type: 'text',
358
+ text: `## 当前视频(共 ${currentVideoFrames.length} 帧,最大限制 ${MAX_FRAMES} 帧)
359
+
360
+ 以下为当前录制视频的关键帧截图:`
361
+ });
362
+ const frameInterval = Math.max(1, Math.floor(currentVideoFrames.length / 10));
363
+ for(let i = 0; i < currentVideoFrames.length; i += frameInterval){
364
+ userContent.push({
365
+ type: 'text',
366
+ text: `帧 ${i + 1}/${currentVideoFrames.length}`
367
+ });
368
+ userContent.push({
369
+ type: 'image_url',
370
+ image_url: {
371
+ url: currentVideoFrames[i],
372
+ detail: 'low'
373
+ }
374
+ });
375
+ }
376
+ userContent.push({
377
+ type: 'text',
378
+ text: buildVideoPrompt({
379
+ assertion,
380
+ businessContext,
381
+ videoOptions
382
+ })
383
+ });
384
+ const msgs = [
385
+ {
386
+ role: 'system',
387
+ content: VIDEO_SYSTEM_PROMPT
388
+ },
389
+ {
390
+ role: 'user',
391
+ content: userContent
392
+ }
393
+ ];
394
+ debug('calling AI for video assertion:', assertion);
395
+ try {
396
+ const result = await callAIWithObjectResponse(msgs, modelConfig, {
397
+ abortSignal
398
+ });
399
+ debug('video assertion result:', result.content);
400
+ return {
401
+ pass: result.content.pass,
402
+ thought: result.content.thought || '',
403
+ reason: result.content.reason,
404
+ videoDetails: result.content.videoDetails,
405
+ usage: result.usage,
406
+ rawResponse: JSON.stringify(result.content)
407
+ };
408
+ } catch (error) {
409
+ debug('video assertion error:', error);
410
+ throw error;
411
+ }
412
+ }
413
+ export { AiAssertDiff, AiAssertElement, AiAssertVideo };
414
+
415
+ //# sourceMappingURL=assert.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"ai-model/assert.mjs","sources":["../../../src/ai-model/assert.ts"],"sourcesContent":["import type {\n AIUsageInfo,\n DiffDetails,\n IgnoreRegion,\n ReferenceImage,\n SystemCheckResults,\n VideoAssertionOptions,\n VideoDetails,\n} from '@/types';\nimport type { IModelConfig } from '@midscene/shared/env';\nimport { getDebug } from '@midscene/shared/logger';\nimport type { ChatCompletionUserMessageParam } from 'openai/resources/index';\nimport { callAIWithObjectResponse } from './service-caller/index';\n\nconst debug = getDebug('ai:assert');\n\nexport interface AiAssertOptions {\n beforeScreenshot?: string;\n afterScreenshot: string;\n assertion: string;\n businessContext?: string;\n enableSystemCheck?: boolean;\n customSystemCheckRules?: string;\n modelConfig: IModelConfig;\n abortSignal?: AbortSignal;\n referenceImages?: ReferenceImage[];\n diffThreshold?: number;\n ignoreRegions?: IgnoreRegion[];\n ignoreDynamicContent?: boolean;\n strictMode?: boolean;\n}\n\nexport interface AIAssertionResponse {\n pass: boolean;\n thought: string;\n reason?: string;\n systemCheckResults?: SystemCheckResults;\n diffDetails?: DiffDetails;\n}\n\nconst DEFAULT_SYSTEM_CHECK_PROMPT = `请检查以下系统级问题:\n\n1. **白屏检测**:页面是否完全白屏或大面积空白\n2. **布局遮挡**:重要元素是否存在组件间重叠遮挡\n3. **未加载内容**:页面是否有加载中的占位符、骨架屏\n4. **错误提示**:页面是否显示错误信息、异常提示\n5. **后端错误**:是否有后端请求失败的提示,如\"活动太火爆了\"、报错 Toast、错误码 等\n\n如果发现以上问题,请在 systemCheckResults 中标注对应字段为 true。`;\n\nfunction buildAssertionPrompt(options: {\n assertion: string;\n businessContext?: string;\n systemCheckPrompt?: string;\n}): string {\n const { assertion, businessContext, systemCheckPrompt } = options;\n\n let prompt = `## 用户断言描述\n${assertion}`;\n\n if (businessContext) {\n prompt += `\n\n## 业务知识上下文\n${businessContext}`;\n }\n\n if (systemCheckPrompt) {\n prompt += `\n\n## 系统校验规则\n${systemCheckPrompt}`;\n }\n\n prompt += `\n\n## 输出要求\n请以 JSON 格式输出断言结果:\n{\n \"pass\": boolean, // 断言是否通过\n \"thought\": string, // 思考过程\n \"reason\": string, // 失败原因(如果失败)\n \"systemCheckResults\": { // 系统校验结果(可选)\n \"whiteScreen\": boolean,\n \"layoutBlocked\": boolean,\n \"loadingContent\": boolean,\n \"errorPrompt\": boolean,\n \"backendError\": boolean\n }\n}`;\n\n return prompt;\n}\n\nexport async function AiAssertElement(options: AiAssertOptions): Promise<{\n pass: boolean;\n thought: string;\n reason?: string;\n usage?: AIUsageInfo;\n systemCheckResults?: SystemCheckResults;\n rawResponse?: string;\n}> {\n const {\n beforeScreenshot,\n afterScreenshot,\n assertion,\n businessContext,\n enableSystemCheck = true,\n customSystemCheckRules,\n modelConfig,\n abortSignal,\n } = options;\n\n const systemPrompt = `你是一个自动化测试断言专家。请根据以下信息判断断言是否通过。\n\n你需要:\n1. 分析操作前后的截图变化(如果提供了操作前截图)\n2. 验证用户的断言描述是否成立\n3. 检查是否存在系统级问题(如果启用)\n4. 给出详细的思考过程和判断结果`;\n\n const systemCheckPrompt = enableSystemCheck\n ? customSystemCheckRules || DEFAULT_SYSTEM_CHECK_PROMPT\n : '';\n\n const userContent: ChatCompletionUserMessageParam['content'] = [];\n\n if (beforeScreenshot) {\n userContent.push({\n type: 'text',\n text: '## 操作前截图(执行操作前的页面状态)',\n });\n userContent.push({\n type: 'image_url',\n image_url: { url: beforeScreenshot, detail: 'high' },\n });\n userContent.push({\n type: 'text',\n text: '## 操作后截图(执行操作后的页面状态)',\n });\n } else {\n userContent.push({\n type: 'text',\n text: '## 当前页面截图',\n });\n }\n\n userContent.push({\n type: 'image_url',\n image_url: { url: afterScreenshot, detail: 'high' },\n });\n\n userContent.push({\n type: 'text',\n text: buildAssertionPrompt({\n assertion,\n businessContext,\n systemCheckPrompt,\n }),\n });\n\n const msgs = [\n { role: 'system' as const, content: systemPrompt },\n { role: 'user' as const, content: userContent },\n ];\n\n debug('calling AI for assertion:', assertion);\n\n try {\n const result = await callAIWithObjectResponse<AIAssertionResponse>(\n msgs,\n modelConfig,\n { abortSignal },\n );\n\n debug('assertion result:', result.content);\n\n return {\n pass: result.content.pass,\n thought: result.content.thought || '',\n reason: result.content.reason,\n usage: result.usage,\n systemCheckResults: result.content.systemCheckResults,\n rawResponse: JSON.stringify(result.content),\n };\n } catch (error) {\n debug('assertion error:', error);\n throw error;\n }\n}\n\nconst DIFF_SYSTEM_PROMPT = `你是一个自动化测试的图像对比专家。你的任务是对比当前页面截图与基准图片,判断页面样式是否符合预期。\n\n你需要:\n1. 仔细对比两张图片的布局、颜色、元素位置\n2. 分析差异是否在可接受范围内\n3. 识别动态内容(如时间、日期)导致的差异\n4. 给出详细的对比分析和判断结果`;\n\nfunction buildDiffPrompt(options: {\n assertion: string;\n businessContext?: string;\n diffThreshold?: number;\n ignoreRegions?: IgnoreRegion[];\n ignoreDynamicContent?: boolean;\n strictMode?: boolean;\n}): string {\n const {\n assertion,\n businessContext,\n diffThreshold = 0.1,\n ignoreRegions,\n ignoreDynamicContent,\n strictMode,\n } = options;\n\n let prompt = `## 图像对比任务\n\n请对比基准图片(预期样式)和当前截图(实际样式),判断页面是否符合预期。\n\n## 用户断言描述\n${assertion}`;\n\n if (businessContext) {\n prompt += `\n\n## 业务知识上下文\n${businessContext}`;\n }\n\n prompt += `\n\n## 对比要求\n\n请从以下维度进行对比分析:\n\n1. **布局一致性**:页面布局是否与基准图片一致\n2. **颜色一致性**:主要颜色是否与基准图片一致\n3. **元素位置**:关键元素的位置是否与基准图片一致\n4. **文字内容**:文字内容是否与基准图片一致\n5. **图片资源**:图片、图标等资源是否正确加载\n\n## 差异阈值\n允许的差异阈值:${diffThreshold * 100}%`;\n\n if (ignoreRegions && ignoreRegions.length > 0) {\n prompt += `\n\n## 忽略对比的区域\n以下区域不参与对比:`;\n ignoreRegions.forEach((region, index) => {\n prompt += `\n${index + 1}. 位置 (${region.x}, ${region.y}),尺寸 ${region.width}x${region.height}`;\n });\n }\n\n if (ignoreDynamicContent) {\n prompt += `\n\n## 动态内容处理\n请忽略以下动态内容导致的差异:\n- 时间显示\n- 日期显示\n- 随机验证码\n- 动画效果`;\n }\n\n if (strictMode) {\n prompt += `\n\n## 严格模式\n当前为严格模式,任何差异都视为失败。`;\n }\n\n prompt += `\n\n## 输出要求\n请以 JSON 格式输出对比结果:\n{\n \"pass\": boolean, // 断言是否通过\n \"thought\": string, // 对比分析过程\n \"reason\": string, // 失败原因(如果失败)\n \"diffDetails\": { // 差异详情\n \"layoutMatch\": boolean, // 布局是否匹配\n \"colorMatch\": boolean, // 颜色是否匹配\n \"elementPositionMatch\": boolean, // 元素位置是否匹配\n \"textContentMatch\": boolean, // 文字内容是否匹配\n \"resourceMatch\": boolean, // 资源是否匹配\n \"acceptableDifferences\": string[], // 可接受的差异列表\n \"unacceptableDifferences\": string[] // 不可接受的差异列表\n }\n}`;\n\n return prompt;\n}\n\nexport async function AiAssertDiff(options: {\n currentScreenshot: string;\n referenceImages: ReferenceImage[];\n assertion: string;\n businessContext?: string;\n diffThreshold?: number;\n ignoreRegions?: IgnoreRegion[];\n ignoreDynamicContent?: boolean;\n strictMode?: boolean;\n modelConfig: IModelConfig;\n abortSignal?: AbortSignal;\n}): Promise<{\n pass: boolean;\n thought: string;\n reason?: string;\n diffDetails?: DiffDetails;\n usage?: AIUsageInfo;\n rawResponse?: string;\n}> {\n const {\n currentScreenshot,\n referenceImages,\n assertion,\n businessContext,\n diffThreshold,\n ignoreRegions,\n ignoreDynamicContent,\n strictMode,\n modelConfig,\n abortSignal,\n } = options;\n\n const userContent: ChatCompletionUserMessageParam['content'] = [];\n\n // 添加基准图片\n for (const refImage of referenceImages) {\n userContent.push({\n type: 'text',\n text: `## 基准图片:${refImage.name}`,\n });\n userContent.push({\n type: 'image_url',\n image_url: { url: refImage.url, detail: 'high' },\n });\n }\n\n // 添加当前截图\n userContent.push({\n type: 'text',\n text: '## 当前截图(实际样式)',\n });\n userContent.push({\n type: 'image_url',\n image_url: { url: currentScreenshot, detail: 'high' },\n });\n\n // 添加对比提示\n userContent.push({\n type: 'text',\n text: buildDiffPrompt({\n assertion,\n businessContext,\n diffThreshold,\n ignoreRegions,\n ignoreDynamicContent,\n strictMode,\n }),\n });\n\n const msgs = [\n { role: 'system' as const, content: DIFF_SYSTEM_PROMPT },\n { role: 'user' as const, content: userContent },\n ];\n\n debug('calling AI for diff assertion:', assertion);\n\n try {\n const result = await callAIWithObjectResponse<AIAssertionResponse>(\n msgs,\n modelConfig,\n { abortSignal },\n );\n\n debug('diff assertion result:', result.content);\n\n return {\n pass: result.content.pass,\n thought: result.content.thought || '',\n reason: result.content.reason,\n diffDetails: result.content.diffDetails,\n usage: result.usage,\n rawResponse: JSON.stringify(result.content),\n };\n } catch (error) {\n debug('diff assertion error:', error);\n throw error;\n }\n}\n\nconst VIDEO_SYSTEM_PROMPT = `你是一个自动化测试的视频/动画分析专家。你的任务是对比当前录制的视频与基准视频,判断动画效果是否符合预期。\n\n你需要:\n1. 分析视频中的动画流畅度、时长、完整性\n2. 对比基准视频与当前视频的差异\n3. 识别动画类型(过渡动画、加载动画、交互动画等)\n4. 检测动画质量问题(卡顿、跳帧等)\n5. 给出详细的分析过程和判断结果`;\n\nfunction buildVideoPrompt(options: {\n assertion: string;\n businessContext?: string;\n videoOptions?: VideoAssertionOptions;\n}): string {\n const { assertion, businessContext, videoOptions } = options;\n\n let prompt = `## 视频/动画对比任务\n\n请对比基准视频(预期动画效果)和当前录制的视频(实际动画效果),判断动画是否符合预期。\n\n## 用户断言描述\n${assertion}`;\n\n if (businessContext) {\n prompt += `\n\n## 业务知识上下文\n${businessContext}`;\n }\n\n prompt += `\n\n## 分析要求\n\n请从以下维度进行动画分析:\n\n1. **动画流畅度**:动画是否流畅,有无卡顿、跳帧\n2. **动画时长**:动画时长是否在预期范围内\n3. **动画完整性**:动画是否完整执行,有无中断\n4. **动画类型**:识别动画类型(过渡、加载、交互等)\n5. **关键帧匹配**:关键时间点的画面是否符合预期`;\n\n if (videoOptions) {\n if (videoOptions.checkSmoothness) {\n prompt += `\n\n## 流畅度检测\n流畅度阈值:${videoOptions.smoothnessThreshold || 60}/100\n请评估动画流畅度并给出评分。`;\n }\n\n if (videoOptions.checkDuration && videoOptions.expectedDuration) {\n prompt += `\n\n## 时长检测\n预期时长范围:${videoOptions.expectedDuration.min || 0}秒 - ${videoOptions.expectedDuration.max || '无限制'}秒`;\n }\n\n if (\n videoOptions.keyframes &&\n videoOptions.keyframes.timestamps.length > 0\n ) {\n prompt += `\n\n## 关键帧验证\n需要在以下时间点验证画面:`;\n videoOptions.keyframes.timestamps.forEach((ts, index) => {\n const desc = videoOptions.keyframes?.descriptions?.[index] || '未描述';\n prompt += `\n- ${ts}秒:${desc}`;\n });\n }\n\n if (videoOptions.animationType && videoOptions.animationType !== 'auto') {\n prompt += `\n\n## 动画类型\n预期动画类型:${videoOptions.animationType}`;\n }\n }\n\n prompt += `\n\n## 输出要求\n请以 JSON 格式输出对比结果:\n{\n \"pass\": boolean, // 断言是否通过\n \"thought\": string, // 分析过程\n \"reason\": string, // 失败原因(如果失败)\n \"videoDetails\": { // 视频详情\n \"smoothnessScore\": number, // 流畅度评分(0-100)\n \"duration\": number, // 动画时长(秒)\n \"isComplete\": boolean, // 动画是否完整\n \"keyframeMatches\": [ // 关键帧匹配结果\n {\n \"timestamp\": number,\n \"matched\": boolean,\n \"description\": string\n }\n ],\n \"detectedAnimationType\": string, // 检测到的动画类型\n \"qualityAssessment\": { // 质量评估\n \"hasStuttering\": boolean, // 是否有卡顿\n \"hasFrameDropping\": boolean, // 是否有跳帧\n \"averageFrameInterval\": number, // 平均帧间隔(ms)\n \"frameIntervalStdDev\": number // 帧间隔标准差\n },\n \"acceptableDifferences\": string[], // 可接受的差异列表\n \"unacceptableDifferences\": string[] // 不可接受的差异列表\n }\n}`;\n\n return prompt;\n}\n\nexport async function AiAssertVideo(options: {\n currentVideoFrames: string[];\n assertion: string;\n businessContext?: string;\n videoOptions?: VideoAssertionOptions;\n modelConfig: IModelConfig;\n abortSignal?: AbortSignal;\n}): Promise<{\n pass: boolean;\n thought: string;\n reason?: string;\n videoDetails?: VideoDetails;\n usage?: AIUsageInfo;\n rawResponse?: string;\n}> {\n const {\n currentVideoFrames,\n assertion,\n businessContext,\n videoOptions,\n modelConfig,\n abortSignal,\n } = options;\n\n const MAX_DURATION = 5;\n const DEFAULT_FPS = 30;\n const MAX_FRAMES = MAX_DURATION * DEFAULT_FPS;\n\n if (currentVideoFrames.length > MAX_FRAMES) {\n throw new Error(\n `Video frames exceed maximum limit. Maximum allowed: ${MAX_FRAMES} frames (${MAX_DURATION}s at ${DEFAULT_FPS}fps), got: ${currentVideoFrames.length} frames`,\n );\n }\n\n const userContent: ChatCompletionUserMessageParam['content'] = [];\n\n userContent.push({\n type: 'text',\n text: `## 当前视频(共 ${currentVideoFrames.length} 帧,最大限制 ${MAX_FRAMES} 帧)\n\n以下为当前录制视频的关键帧截图:`,\n });\n\n const frameInterval = Math.max(1, Math.floor(currentVideoFrames.length / 10));\n for (let i = 0; i < currentVideoFrames.length; i += frameInterval) {\n userContent.push({\n type: 'text',\n text: `帧 ${i + 1}/${currentVideoFrames.length}`,\n });\n userContent.push({\n type: 'image_url',\n image_url: { url: currentVideoFrames[i], detail: 'low' },\n });\n }\n\n userContent.push({\n type: 'text',\n text: buildVideoPrompt({\n assertion,\n businessContext,\n videoOptions,\n }),\n });\n\n const msgs = [\n { role: 'system' as const, content: VIDEO_SYSTEM_PROMPT },\n { role: 'user' as const, content: userContent },\n ];\n\n debug('calling AI for video assertion:', assertion);\n\n try {\n const result = await callAIWithObjectResponse<{\n pass: boolean;\n thought: string;\n reason?: string;\n videoDetails?: VideoDetails;\n }>(msgs, modelConfig, { abortSignal });\n\n debug('video assertion result:', result.content);\n\n return {\n pass: result.content.pass,\n thought: result.content.thought || '',\n reason: result.content.reason,\n videoDetails: result.content.videoDetails,\n usage: result.usage,\n rawResponse: JSON.stringify(result.content),\n };\n } catch (error) {\n debug('video assertion error:', error);\n throw error;\n }\n}\n"],"names":["debug","getDebug","DEFAULT_SYSTEM_CHECK_PROMPT","buildAssertionPrompt","options","assertion","businessContext","systemCheckPrompt","prompt","AiAssertElement","beforeScreenshot","afterScreenshot","enableSystemCheck","customSystemCheckRules","modelConfig","abortSignal","systemPrompt","userContent","msgs","result","callAIWithObjectResponse","JSON","error","DIFF_SYSTEM_PROMPT","buildDiffPrompt","diffThreshold","ignoreRegions","ignoreDynamicContent","strictMode","region","index","AiAssertDiff","currentScreenshot","referenceImages","refImage","VIDEO_SYSTEM_PROMPT","buildVideoPrompt","videoOptions","ts","desc","AiAssertVideo","currentVideoFrames","MAX_DURATION","DEFAULT_FPS","MAX_FRAMES","Error","frameInterval","Math","i"],"mappings":";;AAcA,MAAMA,QAAQC,SAAS;AA0BvB,MAAMC,8BAA8B,CAAC;;;;;;;;6CAQQ,CAAC;AAE9C,SAASC,qBAAqBC,OAI7B;IACC,MAAM,EAAEC,SAAS,EAAEC,eAAe,EAAEC,iBAAiB,EAAE,GAAGH;IAE1D,IAAII,SAAS,CAAC;AAChB,EAAEH,WAAW;IAEX,IAAIC,iBACFE,UAAU,CAAC;;;AAGf,EAAEF,iBAAiB;IAGjB,IAAIC,mBACFC,UAAU,CAAC;;;AAGf,EAAED,mBAAmB;IAGnBC,UAAU,CAAC;;;;;;;;;;;;;;;CAeZ,CAAC;IAEA,OAAOA;AACT;AAEO,eAAeC,gBAAgBL,OAAwB;IAQ5D,MAAM,EACJM,gBAAgB,EAChBC,eAAe,EACfN,SAAS,EACTC,eAAe,EACfM,oBAAoB,IAAI,EACxBC,sBAAsB,EACtBC,WAAW,EACXC,WAAW,EACZ,GAAGX;IAEJ,MAAMY,eAAe,CAAC;;;;;;iBAMP,CAAC;IAEhB,MAAMT,oBAAoBK,oBACtBC,0BAA0BX,8BAC1B;IAEJ,MAAMe,cAAyD,EAAE;IAEjE,IAAIP,kBAAkB;QACpBO,YAAY,IAAI,CAAC;YACf,MAAM;YACN,MAAM;QACR;QACAA,YAAY,IAAI,CAAC;YACf,MAAM;YACN,WAAW;gBAAE,KAAKP;gBAAkB,QAAQ;YAAO;QACrD;QACAO,YAAY,IAAI,CAAC;YACf,MAAM;YACN,MAAM;QACR;IACF,OACEA,YAAY,IAAI,CAAC;QACf,MAAM;QACN,MAAM;IACR;IAGFA,YAAY,IAAI,CAAC;QACf,MAAM;QACN,WAAW;YAAE,KAAKN;YAAiB,QAAQ;QAAO;IACpD;IAEAM,YAAY,IAAI,CAAC;QACf,MAAM;QACN,MAAMd,qBAAqB;YACzBE;YACAC;YACAC;QACF;IACF;IAEA,MAAMW,OAAO;QACX;YAAE,MAAM;YAAmB,SAASF;QAAa;QACjD;YAAE,MAAM;YAAiB,SAASC;QAAY;KAC/C;IAEDjB,MAAM,6BAA6BK;IAEnC,IAAI;QACF,MAAMc,SAAS,MAAMC,yBACnBF,MACAJ,aACA;YAAEC;QAAY;QAGhBf,MAAM,qBAAqBmB,OAAO,OAAO;QAEzC,OAAO;YACL,MAAMA,OAAO,OAAO,CAAC,IAAI;YACzB,SAASA,OAAO,OAAO,CAAC,OAAO,IAAI;YACnC,QAAQA,OAAO,OAAO,CAAC,MAAM;YAC7B,OAAOA,OAAO,KAAK;YACnB,oBAAoBA,OAAO,OAAO,CAAC,kBAAkB;YACrD,aAAaE,KAAK,SAAS,CAACF,OAAO,OAAO;QAC5C;IACF,EAAE,OAAOG,OAAO;QACdtB,MAAM,oBAAoBsB;QAC1B,MAAMA;IACR;AACF;AAEA,MAAMC,qBAAqB,CAAC;;;;;;iBAMX,CAAC;AAElB,SAASC,gBAAgBpB,OAOxB;IACC,MAAM,EACJC,SAAS,EACTC,eAAe,EACfmB,gBAAgB,GAAG,EACnBC,aAAa,EACbC,oBAAoB,EACpBC,UAAU,EACX,GAAGxB;IAEJ,IAAII,SAAS,CAAC;;;;;AAKhB,EAAEH,WAAW;IAEX,IAAIC,iBACFE,UAAU,CAAC;;;AAGf,EAAEF,iBAAiB;IAGjBE,UAAU,CAAC;;;;;;;;;;;;;QAaL,EAAEiB,AAAgB,MAAhBA,cAAoB,CAAC,CAAC;IAE9B,IAAIC,iBAAiBA,cAAc,MAAM,GAAG,GAAG;QAC7ClB,UAAU,CAAC;;;UAGL,CAAC;QACPkB,cAAc,OAAO,CAAC,CAACG,QAAQC;YAC7BtB,UAAU,CAAC;AACjB,EAAEsB,QAAQ,EAAE,MAAM,EAAED,OAAO,CAAC,CAAC,EAAE,EAAEA,OAAO,CAAC,CAAC,KAAK,EAAEA,OAAO,KAAK,CAAC,CAAC,EAAEA,OAAO,MAAM,EAAE;QAC5E;IACF;IAEA,IAAIF,sBACFnB,UAAU,CAAC;;;;;;;MAOT,CAAC;IAGL,IAAIoB,YACFpB,UAAU,CAAC;;;kBAGG,CAAC;IAGjBA,UAAU,CAAC;;;;;;;;;;;;;;;;;CAiBZ,CAAC;IAEA,OAAOA;AACT;AAEO,eAAeuB,aAAa3B,OAWlC;IAQC,MAAM,EACJ4B,iBAAiB,EACjBC,eAAe,EACf5B,SAAS,EACTC,eAAe,EACfmB,aAAa,EACbC,aAAa,EACbC,oBAAoB,EACpBC,UAAU,EACVd,WAAW,EACXC,WAAW,EACZ,GAAGX;IAEJ,MAAMa,cAAyD,EAAE;IAGjE,KAAK,MAAMiB,YAAYD,gBAAiB;QACtChB,YAAY,IAAI,CAAC;YACf,MAAM;YACN,MAAM,CAAC,QAAQ,EAAEiB,SAAS,IAAI,EAAE;QAClC;QACAjB,YAAY,IAAI,CAAC;YACf,MAAM;YACN,WAAW;gBAAE,KAAKiB,SAAS,GAAG;gBAAE,QAAQ;YAAO;QACjD;IACF;IAGAjB,YAAY,IAAI,CAAC;QACf,MAAM;QACN,MAAM;IACR;IACAA,YAAY,IAAI,CAAC;QACf,MAAM;QACN,WAAW;YAAE,KAAKe;YAAmB,QAAQ;QAAO;IACtD;IAGAf,YAAY,IAAI,CAAC;QACf,MAAM;QACN,MAAMO,gBAAgB;YACpBnB;YACAC;YACAmB;YACAC;YACAC;YACAC;QACF;IACF;IAEA,MAAMV,OAAO;QACX;YAAE,MAAM;YAAmB,SAASK;QAAmB;QACvD;YAAE,MAAM;YAAiB,SAASN;QAAY;KAC/C;IAEDjB,MAAM,kCAAkCK;IAExC,IAAI;QACF,MAAMc,SAAS,MAAMC,yBACnBF,MACAJ,aACA;YAAEC;QAAY;QAGhBf,MAAM,0BAA0BmB,OAAO,OAAO;QAE9C,OAAO;YACL,MAAMA,OAAO,OAAO,CAAC,IAAI;YACzB,SAASA,OAAO,OAAO,CAAC,OAAO,IAAI;YACnC,QAAQA,OAAO,OAAO,CAAC,MAAM;YAC7B,aAAaA,OAAO,OAAO,CAAC,WAAW;YACvC,OAAOA,OAAO,KAAK;YACnB,aAAaE,KAAK,SAAS,CAACF,OAAO,OAAO;QAC5C;IACF,EAAE,OAAOG,OAAO;QACdtB,MAAM,yBAAyBsB;QAC/B,MAAMA;IACR;AACF;AAEA,MAAMa,sBAAsB,CAAC;;;;;;;iBAOZ,CAAC;AAElB,SAASC,iBAAiBhC,OAIzB;IACC,MAAM,EAAEC,SAAS,EAAEC,eAAe,EAAE+B,YAAY,EAAE,GAAGjC;IAErD,IAAII,SAAS,CAAC;;;;;AAKhB,EAAEH,WAAW;IAEX,IAAIC,iBACFE,UAAU,CAAC;;;AAGf,EAAEF,iBAAiB;IAGjBE,UAAU,CAAC;;;;;;;;;;2BAUc,CAAC;IAE1B,IAAI6B,cAAc;QAChB,IAAIA,aAAa,eAAe,EAC9B7B,UAAU,CAAC;;;MAGX,EAAE6B,aAAa,mBAAmB,IAAI,GAAG;cACjC,CAAC;QAGX,IAAIA,aAAa,aAAa,IAAIA,aAAa,gBAAgB,EAC7D7B,UAAU,CAAC;;;OAGV,EAAE6B,aAAa,gBAAgB,CAAC,GAAG,IAAI,EAAE,IAAI,EAAEA,aAAa,gBAAgB,CAAC,GAAG,IAAI,MAAM,CAAC,CAAC;QAG/F,IACEA,aAAa,SAAS,IACtBA,aAAa,SAAS,CAAC,UAAU,CAAC,MAAM,GAAG,GAC3C;YACA7B,UAAU,CAAC;;;aAGJ,CAAC;YACR6B,aAAa,SAAS,CAAC,UAAU,CAAC,OAAO,CAAC,CAACC,IAAIR;gBAC7C,MAAMS,OAAOF,aAAa,SAAS,EAAE,cAAc,CAACP,MAAM,IAAI;gBAC9DtB,UAAU,CAAC;EACjB,EAAE8B,GAAG,EAAE,EAAEC,MAAM;YACX;QACF;QAEA,IAAIF,aAAa,aAAa,IAAIA,AAA+B,WAA/BA,aAAa,aAAa,EAC1D7B,UAAU,CAAC;;;OAGV,EAAE6B,aAAa,aAAa,EAAE;IAEnC;IAEA7B,UAAU,CAAC;;;;;;;;;;;;;;;;;;;;;;;;;;;;;CA6BZ,CAAC;IAEA,OAAOA;AACT;AAEO,eAAegC,cAAcpC,OAOnC;IAQC,MAAM,EACJqC,kBAAkB,EAClBpC,SAAS,EACTC,eAAe,EACf+B,YAAY,EACZvB,WAAW,EACXC,WAAW,EACZ,GAAGX;IAEJ,MAAMsC,eAAe;IACrB,MAAMC,cAAc;IACpB,MAAMC,aAAaF,eAAeC;IAElC,IAAIF,mBAAmB,MAAM,GAAGG,YAC9B,MAAM,IAAIC,MACR,CAAC,oDAAoD,EAAED,WAAW,SAAS,EAAEF,aAAa,KAAK,EAAEC,YAAY,WAAW,EAAEF,mBAAmB,MAAM,CAAC,OAAO,CAAC;IAIhK,MAAMxB,cAAyD,EAAE;IAEjEA,YAAY,IAAI,CAAC;QACf,MAAM;QACN,MAAM,CAAC,UAAU,EAAEwB,mBAAmB,MAAM,CAAC,QAAQ,EAAEG,WAAW;;gBAEtD,CAAC;IACf;IAEA,MAAME,gBAAgBC,KAAK,GAAG,CAAC,GAAGA,KAAK,KAAK,CAACN,mBAAmB,MAAM,GAAG;IACzE,IAAK,IAAIO,IAAI,GAAGA,IAAIP,mBAAmB,MAAM,EAAEO,KAAKF,cAAe;QACjE7B,YAAY,IAAI,CAAC;YACf,MAAM;YACN,MAAM,CAAC,EAAE,EAAE+B,IAAI,EAAE,CAAC,EAAEP,mBAAmB,MAAM,EAAE;QACjD;QACAxB,YAAY,IAAI,CAAC;YACf,MAAM;YACN,WAAW;gBAAE,KAAKwB,kBAAkB,CAACO,EAAE;gBAAE,QAAQ;YAAM;QACzD;IACF;IAEA/B,YAAY,IAAI,CAAC;QACf,MAAM;QACN,MAAMmB,iBAAiB;YACrB/B;YACAC;YACA+B;QACF;IACF;IAEA,MAAMnB,OAAO;QACX;YAAE,MAAM;YAAmB,SAASiB;QAAoB;QACxD;YAAE,MAAM;YAAiB,SAASlB;QAAY;KAC/C;IAEDjB,MAAM,mCAAmCK;IAEzC,IAAI;QACF,MAAMc,SAAS,MAAMC,yBAKlBF,MAAMJ,aAAa;YAAEC;QAAY;QAEpCf,MAAM,2BAA2BmB,OAAO,OAAO;QAE/C,OAAO;YACL,MAAMA,OAAO,OAAO,CAAC,IAAI;YACzB,SAASA,OAAO,OAAO,CAAC,OAAO,IAAI;YACnC,QAAQA,OAAO,OAAO,CAAC,MAAM;YAC7B,cAAcA,OAAO,OAAO,CAAC,YAAY;YACzC,OAAOA,OAAO,KAAK;YACnB,aAAaE,KAAK,SAAS,CAACF,OAAO,OAAO;QAC5C;IACF,EAAE,OAAOG,OAAO;QACdtB,MAAM,0BAA0BsB;QAChC,MAAMA;IACR;AACF"}
@@ -3,9 +3,10 @@ import { systemPromptToLocateElement } from "./prompt/llm-locator.mjs";
3
3
  import { generatePlaywrightTest, generatePlaywrightTestStream } from "./prompt/playwright-generator.mjs";
4
4
  import { generateYamlTest, generateYamlTestStream } from "./prompt/yaml-generator.mjs";
5
5
  import { AiExtractElementInfo, AiJudgeOrderSensitive, AiLocateElement, AiLocateSection } from "./inspect.mjs";
6
+ import { AiAssertDiff, AiAssertElement } from "./assert.mjs";
6
7
  import { plan } from "./llm-planning.mjs";
7
8
  import { autoGLMPlanning } from "./auto-glm/planning.mjs";
8
9
  import { PointSchema, RectSchema, SizeSchema, TMultimodalPromptSchema, TUserPromptSchema, adaptBboxToRect, dumpActionParam, findAllMidsceneLocatorField, getMidsceneLocationSchema, parseActionParam } from "../common.mjs";
9
10
  import { uiTarsPlanning } from "./ui-tars-planning.mjs";
10
11
  import { ConversationHistory } from "./conversation-history.mjs";
11
- export { AIResponseParseError, AiExtractElementInfo, AiJudgeOrderSensitive, AiLocateElement, AiLocateSection, ConversationHistory, PointSchema, RectSchema, SizeSchema, TMultimodalPromptSchema, TUserPromptSchema, adaptBboxToRect, autoGLMPlanning, callAI, callAIWithObjectResponse, callAIWithStringResponse, dumpActionParam, findAllMidsceneLocatorField, generatePlaywrightTest, generatePlaywrightTestStream, generateYamlTest, generateYamlTestStream, getMidsceneLocationSchema, parseActionParam, plan, systemPromptToLocateElement, uiTarsPlanning };
12
+ export { AIResponseParseError, AiAssertDiff, AiAssertElement, AiExtractElementInfo, AiJudgeOrderSensitive, AiLocateElement, AiLocateSection, ConversationHistory, PointSchema, RectSchema, SizeSchema, TMultimodalPromptSchema, TUserPromptSchema, adaptBboxToRect, autoGLMPlanning, callAI, callAIWithObjectResponse, callAIWithStringResponse, dumpActionParam, findAllMidsceneLocatorField, generatePlaywrightTest, generatePlaywrightTestStream, generateYamlTest, generateYamlTestStream, getMidsceneLocationSchema, parseActionParam, plan, systemPromptToLocateElement, uiTarsPlanning };
@@ -0,0 +1,34 @@
1
+ import { getPreferredLanguage } from "@midscene/shared/env";
2
+ const cacheVerifyInstruction = ()=>{
3
+ const preferredLanguage = getPreferredLanguage();
4
+ return `
5
+ You are verifying whether a cached element location is still correct.
6
+ The image shows a SMALL CROPPED AREA around the cached coordinates, with the target element marked by a red rectangle.
7
+
8
+ TASK: Determine if the element in the red rectangle matches the user's original description.
9
+
10
+ USER'S ORIGINAL DESCRIPTION: "{{targetPrompt}}"
11
+
12
+ RULES:
13
+ 1. Look at the element inside the red rectangle carefully
14
+ 2. Compare it with the user's description: "{{targetPrompt}}"
15
+ 3. Consider:
16
+ - Is it the same TYPE of element? (button, input, link, text, etc.)
17
+ - Does it have matching TEXT or LABEL?
18
+ - Is it in a reasonable CONTEXT for that description?
19
+ 4. If the description is vague (e.g., "click the button"), be more lenient
20
+ 5. If the description is specific (e.g., "Login button", "Submit button"), be strict
21
+
22
+ RESPONSE FORMAT (JSON):
23
+ {
24
+ "match": true/false,
25
+ "description": "brief description of what you see in the red rectangle",
26
+ "reason": "short explanation of why it matches or not"
27
+ }
28
+
29
+ IMPORTANT: Write descriptions and reasons in ${preferredLanguage}.
30
+ `;
31
+ };
32
+ export { cacheVerifyInstruction };
33
+
34
+ //# sourceMappingURL=cache-verify.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"ai-model/prompt/cache-verify.mjs","sources":["../../../../src/ai-model/prompt/cache-verify.ts"],"sourcesContent":["import { getPreferredLanguage } from '@midscene/shared/env';\n\nexport const cacheVerifyInstruction = () => {\n const preferredLanguage = getPreferredLanguage();\n\n return `\nYou are verifying whether a cached element location is still correct.\nThe image shows a SMALL CROPPED AREA around the cached coordinates, with the target element marked by a red rectangle.\n\nTASK: Determine if the element in the red rectangle matches the user's original description.\n\nUSER'S ORIGINAL DESCRIPTION: \"{{targetPrompt}}\"\n\nRULES:\n1. Look at the element inside the red rectangle carefully\n2. Compare it with the user's description: \"{{targetPrompt}}\"\n3. Consider:\n - Is it the same TYPE of element? (button, input, link, text, etc.)\n - Does it have matching TEXT or LABEL?\n - Is it in a reasonable CONTEXT for that description?\n4. If the description is vague (e.g., \"click the button\"), be more lenient\n5. If the description is specific (e.g., \"Login button\", \"Submit button\"), be strict\n\nRESPONSE FORMAT (JSON):\n{\n \"match\": true/false,\n \"description\": \"brief description of what you see in the red rectangle\",\n \"reason\": \"short explanation of why it matches or not\"\n}\n\nIMPORTANT: Write descriptions and reasons in ${preferredLanguage}.\n`;\n};\n"],"names":["cacheVerifyInstruction","preferredLanguage","getPreferredLanguage"],"mappings":";AAEO,MAAMA,yBAAyB;IACpC,MAAMC,oBAAoBC;IAE1B,OAAO,CAAC;;;;;;;;;;;;;;;;;;;;;;;;;6CAyBmC,EAAED,kBAAkB;AACjE,CAAC;AACD"}
@@ -1,6 +1,7 @@
1
1
  import { isAutoGLM } from "../ai-model/auto-glm/util.mjs";
2
2
  import { AIResponseParseError, AiExtractElementInfo, AiLocateElement, callAIWithObjectResponse } from "../ai-model/index.mjs";
3
3
  import { AiLocateSection } from "../ai-model/inspect.mjs";
4
+ import { cacheVerifyInstruction } from "../ai-model/prompt/cache-verify.mjs";
4
5
  import { elementDescriberInstruction } from "../ai-model/prompt/describe.mjs";
5
6
  import { expandSearchArea } from "../common.mjs";
6
7
  import { ServiceError } from "../types.mjs";
@@ -38,7 +39,6 @@ class Service {
38
39
  const context = opt?.context || await this.contextRetrieverFn();
39
40
  let searchArea;
40
41
  let searchAreaRawResponse;
41
- let searchAreaUsage;
42
42
  let searchAreaResponse;
43
43
  if (searchAreaPrompt) {
44
44
  searchAreaResponse = await AiLocateSection({
@@ -49,7 +49,7 @@ class Service {
49
49
  });
50
50
  assert(searchAreaResponse.rect, `cannot find search area for "${searchAreaPrompt}"${searchAreaResponse.error ? `: ${searchAreaResponse.error}` : ''}`);
51
51
  searchAreaRawResponse = searchAreaResponse.rawResponse;
52
- searchAreaUsage = searchAreaResponse.usage;
52
+ searchAreaResponse.usage;
53
53
  searchArea = searchAreaResponse.rect;
54
54
  }
55
55
  const startTime = Date.now();
@@ -69,7 +69,6 @@ class Service {
69
69
  usage,
70
70
  searchArea,
71
71
  searchAreaRawResponse,
72
- searchAreaUsage,
73
72
  reasoning_content
74
73
  };
75
74
  let errorLog;
@@ -82,9 +81,9 @@ class Service {
82
81
  matchedElement: [],
83
82
  matchedRect: rect,
84
83
  data: null,
85
- taskInfo,
86
- deepLocate: !!searchArea,
87
- error: errorLog
84
+ deepLocate: !!searchArea || !!searchAreaPrompt,
85
+ error: errorLog,
86
+ taskInfo
88
87
  };
89
88
  const elements = parseResult.elements || [];
90
89
  const dump = createServiceDump({
@@ -242,6 +241,82 @@ class Service {
242
241
  assert(content.description, 'failed to describe the element');
243
242
  return content;
244
243
  }
244
+ async verifyCachedElement(center, targetPrompt, modelConfig, uiContext) {
245
+ const context = uiContext || await this.contextRetrieverFn();
246
+ const { shotSize } = context;
247
+ const screenshotBase64 = context.screenshot.base64;
248
+ assert(screenshotBase64, 'screenshot is required for verifyCachedElement');
249
+ const defaultRectSize = 60;
250
+ const targetRect = {
251
+ left: Math.floor(center[0] - defaultRectSize / 2),
252
+ top: Math.floor(center[1] - defaultRectSize / 2),
253
+ width: defaultRectSize,
254
+ height: defaultRectSize
255
+ };
256
+ const cropArea = expandSearchArea(targetRect, shotSize);
257
+ let imagePayload = await compositeElementInfoImg({
258
+ inputImgBase64: screenshotBase64,
259
+ size: shotSize,
260
+ elementsPositionInfo: [
261
+ {
262
+ rect: targetRect
263
+ }
264
+ ],
265
+ borderThickness: 3
266
+ });
267
+ debug('verifyCachedElement: cropping to area', cropArea, 'from center', center);
268
+ const croppedResult = await cropByRect(imagePayload, cropArea, 'qwen2.5-vl' === modelConfig.modelFamily);
269
+ imagePayload = croppedResult.imageBase64;
270
+ const promptText = 'string' == typeof targetPrompt ? targetPrompt : targetPrompt.prompt || '';
271
+ const systemPrompt = cacheVerifyInstruction().replace(/{{targetPrompt}}/g, promptText);
272
+ const msgs = [
273
+ {
274
+ role: 'system',
275
+ content: systemPrompt
276
+ },
277
+ {
278
+ role: 'user',
279
+ content: [
280
+ {
281
+ type: 'image_url',
282
+ image_url: {
283
+ url: imagePayload,
284
+ detail: 'low'
285
+ }
286
+ }
287
+ ]
288
+ }
289
+ ];
290
+ try {
291
+ const res = await callAIWithObjectResponse(msgs, modelConfig);
292
+ const content = res.content;
293
+ if (content.error) {
294
+ debug('verifyCachedElement AI returned error:', content.error);
295
+ return {
296
+ pass: false,
297
+ reason: content.error
298
+ };
299
+ }
300
+ const result = {
301
+ pass: !!content.match,
302
+ description: content.description,
303
+ reason: content.reason
304
+ };
305
+ debug('verifyCachedElement result:', {
306
+ pass: result.pass,
307
+ description: result.description,
308
+ reason: result.reason,
309
+ targetPrompt: promptText
310
+ });
311
+ return result;
312
+ } catch (error) {
313
+ debug('verifyCachedElement failed:', error);
314
+ return {
315
+ pass: false,
316
+ reason: String(error)
317
+ };
318
+ }
319
+ }
245
320
  constructor(context, opt){
246
321
  _define_property(this, "contextRetrieverFn", void 0);
247
322
  _define_property(this, "taskInfo", void 0);