@mastra/libsql 1.7.1 → 1.7.2-alpha.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +13 -0
- package/dist/docs/SKILL.md +2 -4
- package/dist/docs/assets/SOURCE_MAP.json +1 -1
- package/dist/docs/references/docs-agents-agent-approval.md +114 -193
- package/dist/docs/references/docs-agents-networks.md +88 -205
- package/dist/docs/references/docs-memory-overview.md +219 -24
- package/dist/docs/references/docs-memory-semantic-recall.md +1 -1
- package/dist/docs/references/docs-memory-storage.md +4 -4
- package/dist/docs/references/docs-memory-working-memory.md +1 -1
- package/dist/docs/references/docs-observability-overview.md +1 -1
- package/dist/docs/references/reference-core-getMemory.md +1 -2
- package/dist/docs/references/reference-core-listMemory.md +1 -2
- package/dist/index.cjs +112 -5
- package/dist/index.cjs.map +1 -1
- package/dist/index.js +112 -5
- package/dist/index.js.map +1 -1
- package/dist/storage/domains/datasets/index.d.ts.map +1 -1
- package/dist/storage/domains/experiments/index.d.ts +2 -1
- package/dist/storage/domains/experiments/index.d.ts.map +1 -1
- package/package.json +5 -5
- package/dist/docs/references/docs-agents-agent-memory.md +0 -209
- package/dist/docs/references/docs-agents-network-approval.md +0 -278
package/dist/index.js
CHANGED
|
@@ -2985,7 +2985,11 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
2985
2985
|
await this.#db.createTable({ tableName: TABLE_DATASET_ITEMS, schema: DATASET_ITEMS_SCHEMA });
|
|
2986
2986
|
await this.#db.createTable({ tableName: TABLE_DATASET_VERSIONS, schema: DATASET_VERSIONS_SCHEMA });
|
|
2987
2987
|
await this.#addColumnIfNotExists(TABLE_DATASETS, "requestContextSchema", "TEXT");
|
|
2988
|
+
await this.#addColumnIfNotExists(TABLE_DATASETS, "tags", "TEXT");
|
|
2989
|
+
await this.#addColumnIfNotExists(TABLE_DATASETS, "targetType", "TEXT");
|
|
2990
|
+
await this.#addColumnIfNotExists(TABLE_DATASETS, "targetIds", "TEXT");
|
|
2988
2991
|
await this.#addColumnIfNotExists(TABLE_DATASET_ITEMS, "requestContext", "TEXT");
|
|
2992
|
+
await this.#addColumnIfNotExists(TABLE_DATASET_ITEMS, "source", "TEXT");
|
|
2989
2993
|
await this.#client.execute({
|
|
2990
2994
|
sql: `CREATE INDEX IF NOT EXISTS idx_dataset_items_dataset_validto ON "${TABLE_DATASET_ITEMS}" ("datasetId", "validTo")`,
|
|
2991
2995
|
args: []
|
|
@@ -3028,6 +3032,9 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3028
3032
|
inputSchema: row.inputSchema ? safelyParseJSON(row.inputSchema) : void 0,
|
|
3029
3033
|
groundTruthSchema: row.groundTruthSchema ? safelyParseJSON(row.groundTruthSchema) : void 0,
|
|
3030
3034
|
requestContextSchema: row.requestContextSchema ? safelyParseJSON(row.requestContextSchema) : void 0,
|
|
3035
|
+
tags: row.tags ? safelyParseJSON(row.tags) : void 0,
|
|
3036
|
+
targetType: row.targetType || void 0,
|
|
3037
|
+
targetIds: row.targetIds ? safelyParseJSON(row.targetIds) : void 0,
|
|
3031
3038
|
version: row.version,
|
|
3032
3039
|
createdAt: ensureDate(row.createdAt),
|
|
3033
3040
|
updatedAt: ensureDate(row.updatedAt)
|
|
@@ -3042,6 +3049,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3042
3049
|
groundTruth: row.groundTruth ? safelyParseJSON(row.groundTruth) : void 0,
|
|
3043
3050
|
requestContext: row.requestContext ? safelyParseJSON(row.requestContext) : void 0,
|
|
3044
3051
|
metadata: row.metadata ? safelyParseJSON(row.metadata) : void 0,
|
|
3052
|
+
source: row.source ? safelyParseJSON(row.source) : void 0,
|
|
3045
3053
|
createdAt: ensureDate(row.createdAt),
|
|
3046
3054
|
updatedAt: ensureDate(row.updatedAt)
|
|
3047
3055
|
};
|
|
@@ -3057,6 +3065,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3057
3065
|
groundTruth: row.groundTruth ? safelyParseJSON(row.groundTruth) : void 0,
|
|
3058
3066
|
requestContext: row.requestContext ? safelyParseJSON(row.requestContext) : void 0,
|
|
3059
3067
|
metadata: row.metadata ? safelyParseJSON(row.metadata) : void 0,
|
|
3068
|
+
source: row.source ? safelyParseJSON(row.source) : void 0,
|
|
3060
3069
|
createdAt: ensureDate(row.createdAt),
|
|
3061
3070
|
updatedAt: ensureDate(row.updatedAt)
|
|
3062
3071
|
};
|
|
@@ -3085,6 +3094,8 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3085
3094
|
inputSchema: input.inputSchema ?? null,
|
|
3086
3095
|
groundTruthSchema: input.groundTruthSchema ?? null,
|
|
3087
3096
|
requestContextSchema: input.requestContextSchema ?? null,
|
|
3097
|
+
targetType: input.targetType ?? null,
|
|
3098
|
+
targetIds: input.targetIds ? JSON.stringify(input.targetIds) : null,
|
|
3088
3099
|
version: 0,
|
|
3089
3100
|
createdAt: nowIso,
|
|
3090
3101
|
updatedAt: nowIso
|
|
@@ -3098,6 +3109,8 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3098
3109
|
inputSchema: input.inputSchema ?? void 0,
|
|
3099
3110
|
groundTruthSchema: input.groundTruthSchema ?? void 0,
|
|
3100
3111
|
requestContextSchema: input.requestContextSchema ?? void 0,
|
|
3112
|
+
targetType: input.targetType ?? void 0,
|
|
3113
|
+
targetIds: input.targetIds ?? void 0,
|
|
3101
3114
|
version: 0,
|
|
3102
3115
|
createdAt: now,
|
|
3103
3116
|
updatedAt: now
|
|
@@ -3169,6 +3182,18 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3169
3182
|
updates.push("requestContextSchema = ?");
|
|
3170
3183
|
values.push(args.requestContextSchema === null ? null : JSON.stringify(args.requestContextSchema));
|
|
3171
3184
|
}
|
|
3185
|
+
if (args.tags !== void 0) {
|
|
3186
|
+
updates.push("tags = ?");
|
|
3187
|
+
values.push(args.tags === null ? null : JSON.stringify(args.tags));
|
|
3188
|
+
}
|
|
3189
|
+
if (args.targetType !== void 0) {
|
|
3190
|
+
updates.push("targetType = ?");
|
|
3191
|
+
values.push(args.targetType === null ? null : args.targetType);
|
|
3192
|
+
}
|
|
3193
|
+
if (args.targetIds !== void 0) {
|
|
3194
|
+
updates.push("targetIds = ?");
|
|
3195
|
+
values.push(args.targetIds === null ? null : JSON.stringify(args.targetIds));
|
|
3196
|
+
}
|
|
3172
3197
|
values.push(args.id);
|
|
3173
3198
|
await this.#client.execute({
|
|
3174
3199
|
sql: `UPDATE ${TABLE_DATASETS} SET ${updates.join(", ")} WHERE id = ?`,
|
|
@@ -3182,6 +3207,9 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3182
3207
|
inputSchema: (args.inputSchema !== void 0 ? args.inputSchema : existing.inputSchema) ?? void 0,
|
|
3183
3208
|
groundTruthSchema: (args.groundTruthSchema !== void 0 ? args.groundTruthSchema : existing.groundTruthSchema) ?? void 0,
|
|
3184
3209
|
requestContextSchema: (args.requestContextSchema !== void 0 ? args.requestContextSchema : existing.requestContextSchema) ?? void 0,
|
|
3210
|
+
tags: (args.tags !== void 0 ? args.tags : existing.tags) ?? void 0,
|
|
3211
|
+
targetType: (args.targetType !== void 0 ? args.targetType : existing.targetType) ?? void 0,
|
|
3212
|
+
targetIds: (args.targetIds !== void 0 ? args.targetIds : existing.targetIds) ?? void 0,
|
|
3185
3213
|
updatedAt: new Date(now)
|
|
3186
3214
|
};
|
|
3187
3215
|
} catch (error) {
|
|
@@ -3287,7 +3315,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3287
3315
|
args: [args.datasetId]
|
|
3288
3316
|
},
|
|
3289
3317
|
{
|
|
3290
|
-
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3318
|
+
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3291
3319
|
args: [
|
|
3292
3320
|
id,
|
|
3293
3321
|
args.datasetId,
|
|
@@ -3296,6 +3324,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3296
3324
|
jsonbArg(args.groundTruth),
|
|
3297
3325
|
jsonbArg(args.requestContext),
|
|
3298
3326
|
jsonbArg(args.metadata),
|
|
3327
|
+
jsonbArg(args.source),
|
|
3299
3328
|
nowIso,
|
|
3300
3329
|
nowIso
|
|
3301
3330
|
]
|
|
@@ -3316,6 +3345,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3316
3345
|
groundTruth: args.groundTruth,
|
|
3317
3346
|
requestContext: args.requestContext,
|
|
3318
3347
|
metadata: args.metadata,
|
|
3348
|
+
source: args.source,
|
|
3319
3349
|
createdAt: now,
|
|
3320
3350
|
updatedAt: now
|
|
3321
3351
|
};
|
|
@@ -3357,6 +3387,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3357
3387
|
const mergedGroundTruth = args.groundTruth ?? existing.groundTruth;
|
|
3358
3388
|
const mergedRequestContext = args.requestContext ?? existing.requestContext;
|
|
3359
3389
|
const mergedMetadata = args.metadata ?? existing.metadata;
|
|
3390
|
+
const mergedSource = args.source ?? existing.source;
|
|
3360
3391
|
const results = await this.#client.batch(
|
|
3361
3392
|
[
|
|
3362
3393
|
{
|
|
@@ -3368,7 +3399,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3368
3399
|
args: [args.datasetId, args.id]
|
|
3369
3400
|
},
|
|
3370
3401
|
{
|
|
3371
|
-
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3402
|
+
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3372
3403
|
args: [
|
|
3373
3404
|
args.id,
|
|
3374
3405
|
args.datasetId,
|
|
@@ -3377,6 +3408,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3377
3408
|
jsonbArg(mergedGroundTruth),
|
|
3378
3409
|
jsonbArg(mergedRequestContext),
|
|
3379
3410
|
jsonbArg(mergedMetadata),
|
|
3411
|
+
jsonbArg(mergedSource),
|
|
3380
3412
|
existing.createdAt.toISOString(),
|
|
3381
3413
|
nowIso
|
|
3382
3414
|
]
|
|
@@ -3396,6 +3428,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3396
3428
|
groundTruth: mergedGroundTruth,
|
|
3397
3429
|
requestContext: mergedRequestContext,
|
|
3398
3430
|
metadata: mergedMetadata,
|
|
3431
|
+
source: mergedSource,
|
|
3399
3432
|
updatedAt: now
|
|
3400
3433
|
};
|
|
3401
3434
|
} catch (error) {
|
|
@@ -3435,7 +3468,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3435
3468
|
args: [datasetId, id]
|
|
3436
3469
|
},
|
|
3437
3470
|
{
|
|
3438
|
-
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 1, jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3471
|
+
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 1, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3439
3472
|
args: [
|
|
3440
3473
|
id,
|
|
3441
3474
|
datasetId,
|
|
@@ -3444,6 +3477,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3444
3477
|
jsonbArg(existing.groundTruth),
|
|
3445
3478
|
jsonbArg(existing.requestContext),
|
|
3446
3479
|
jsonbArg(existing.metadata),
|
|
3480
|
+
jsonbArg(existing.source),
|
|
3447
3481
|
existing.createdAt.toISOString(),
|
|
3448
3482
|
nowIso
|
|
3449
3483
|
]
|
|
@@ -3723,7 +3757,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3723
3757
|
const id = crypto.randomUUID();
|
|
3724
3758
|
items.push({ id, input: itemInput });
|
|
3725
3759
|
statements.push({
|
|
3726
|
-
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3760
|
+
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3727
3761
|
args: [
|
|
3728
3762
|
id,
|
|
3729
3763
|
input.datasetId,
|
|
@@ -3732,6 +3766,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3732
3766
|
jsonbArg(itemInput.groundTruth),
|
|
3733
3767
|
jsonbArg(itemInput.requestContext),
|
|
3734
3768
|
jsonbArg(itemInput.metadata),
|
|
3769
|
+
jsonbArg(itemInput.source),
|
|
3735
3770
|
nowIso,
|
|
3736
3771
|
nowIso
|
|
3737
3772
|
]
|
|
@@ -3751,6 +3786,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3751
3786
|
groundTruth: itemInput.groundTruth,
|
|
3752
3787
|
requestContext: itemInput.requestContext,
|
|
3753
3788
|
metadata: itemInput.metadata,
|
|
3789
|
+
source: itemInput.source,
|
|
3754
3790
|
createdAt: now,
|
|
3755
3791
|
updatedAt: now
|
|
3756
3792
|
}));
|
|
@@ -3799,7 +3835,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3799
3835
|
args: [input.datasetId, item.id]
|
|
3800
3836
|
});
|
|
3801
3837
|
statements.push({
|
|
3802
|
-
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 1, jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3838
|
+
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 1, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3803
3839
|
args: [
|
|
3804
3840
|
item.id,
|
|
3805
3841
|
input.datasetId,
|
|
@@ -3808,6 +3844,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3808
3844
|
jsonbArg(item.groundTruth),
|
|
3809
3845
|
jsonbArg(item.requestContext),
|
|
3810
3846
|
jsonbArg(item.metadata),
|
|
3847
|
+
jsonbArg(item.source),
|
|
3811
3848
|
item.createdAt.toISOString(),
|
|
3812
3849
|
nowIso
|
|
3813
3850
|
]
|
|
@@ -3900,6 +3937,8 @@ var ExperimentsLibSQL = class extends ExperimentsStorage {
|
|
|
3900
3937
|
completedAt: ensureDate(row.completedAt),
|
|
3901
3938
|
retryCount: row.retryCount,
|
|
3902
3939
|
traceId: row.traceId ?? null,
|
|
3940
|
+
status: row.status ?? null,
|
|
3941
|
+
tags: row.tags ? safelyParseJSON(row.tags) : null,
|
|
3903
3942
|
createdAt: ensureDate(row.createdAt)
|
|
3904
3943
|
};
|
|
3905
3944
|
}
|
|
@@ -4143,6 +4182,8 @@ var ExperimentsLibSQL = class extends ExperimentsStorage {
|
|
|
4143
4182
|
completedAt: input.completedAt.toISOString(),
|
|
4144
4183
|
retryCount: input.retryCount,
|
|
4145
4184
|
traceId: input.traceId ?? null,
|
|
4185
|
+
status: input.status ?? null,
|
|
4186
|
+
tags: input.tags !== void 0 && input.tags !== null ? JSON.stringify(input.tags) : null,
|
|
4146
4187
|
createdAt: nowIso
|
|
4147
4188
|
}
|
|
4148
4189
|
});
|
|
@@ -4159,6 +4200,8 @@ var ExperimentsLibSQL = class extends ExperimentsStorage {
|
|
|
4159
4200
|
completedAt: input.completedAt,
|
|
4160
4201
|
retryCount: input.retryCount,
|
|
4161
4202
|
traceId: input.traceId ?? null,
|
|
4203
|
+
status: input.status ?? null,
|
|
4204
|
+
tags: input.tags ?? null,
|
|
4162
4205
|
createdAt: now
|
|
4163
4206
|
};
|
|
4164
4207
|
} catch (error) {
|
|
@@ -4172,6 +4215,70 @@ var ExperimentsLibSQL = class extends ExperimentsStorage {
|
|
|
4172
4215
|
);
|
|
4173
4216
|
}
|
|
4174
4217
|
}
|
|
4218
|
+
async updateExperimentResult(input) {
|
|
4219
|
+
try {
|
|
4220
|
+
const setClauses = [];
|
|
4221
|
+
const values = [];
|
|
4222
|
+
if (input.status !== void 0) {
|
|
4223
|
+
setClauses.push(`"status" = ?`);
|
|
4224
|
+
values.push(input.status);
|
|
4225
|
+
}
|
|
4226
|
+
if (input.tags !== void 0) {
|
|
4227
|
+
setClauses.push(`"tags" = ?`);
|
|
4228
|
+
values.push(JSON.stringify(input.tags));
|
|
4229
|
+
}
|
|
4230
|
+
if (setClauses.length === 0) {
|
|
4231
|
+
const existing = await this.getExperimentResultById({ id: input.id });
|
|
4232
|
+
if (!existing) {
|
|
4233
|
+
throw new MastraError({
|
|
4234
|
+
id: createStorageErrorId("LIBSQL", "UPDATE_EXPERIMENT_RESULT", "NOT_FOUND"),
|
|
4235
|
+
domain: ErrorDomain.STORAGE,
|
|
4236
|
+
category: ErrorCategory.USER,
|
|
4237
|
+
details: { resultId: input.id }
|
|
4238
|
+
});
|
|
4239
|
+
}
|
|
4240
|
+
return existing;
|
|
4241
|
+
}
|
|
4242
|
+
values.push(input.id);
|
|
4243
|
+
let whereClause = `"id" = ?`;
|
|
4244
|
+
if (input.experimentId) {
|
|
4245
|
+
values.push(input.experimentId);
|
|
4246
|
+
whereClause += ` AND "experimentId" = ?`;
|
|
4247
|
+
}
|
|
4248
|
+
const updateResult = await this.#client.execute({
|
|
4249
|
+
sql: `UPDATE ${TABLE_EXPERIMENT_RESULTS} SET ${setClauses.join(", ")} WHERE ${whereClause}`,
|
|
4250
|
+
args: values
|
|
4251
|
+
});
|
|
4252
|
+
if (updateResult.rowsAffected === 0) {
|
|
4253
|
+
throw new MastraError({
|
|
4254
|
+
id: createStorageErrorId("LIBSQL", "UPDATE_EXPERIMENT_RESULT", "NOT_FOUND"),
|
|
4255
|
+
domain: ErrorDomain.STORAGE,
|
|
4256
|
+
category: ErrorCategory.USER,
|
|
4257
|
+
details: { resultId: input.id, ...input.experimentId ? { experimentId: input.experimentId } : {} }
|
|
4258
|
+
});
|
|
4259
|
+
}
|
|
4260
|
+
const result = await this.getExperimentResultById({ id: input.id });
|
|
4261
|
+
if (!result) {
|
|
4262
|
+
throw new MastraError({
|
|
4263
|
+
id: createStorageErrorId("LIBSQL", "UPDATE_EXPERIMENT_RESULT", "NOT_FOUND"),
|
|
4264
|
+
domain: ErrorDomain.STORAGE,
|
|
4265
|
+
category: ErrorCategory.USER,
|
|
4266
|
+
details: { resultId: input.id }
|
|
4267
|
+
});
|
|
4268
|
+
}
|
|
4269
|
+
return result;
|
|
4270
|
+
} catch (error) {
|
|
4271
|
+
if (error instanceof MastraError) throw error;
|
|
4272
|
+
throw new MastraError(
|
|
4273
|
+
{
|
|
4274
|
+
id: createStorageErrorId("LIBSQL", "UPDATE_EXPERIMENT_RESULT", "FAILED"),
|
|
4275
|
+
domain: ErrorDomain.STORAGE,
|
|
4276
|
+
category: ErrorCategory.THIRD_PARTY
|
|
4277
|
+
},
|
|
4278
|
+
error
|
|
4279
|
+
);
|
|
4280
|
+
}
|
|
4281
|
+
}
|
|
4175
4282
|
async getExperimentResultById(args) {
|
|
4176
4283
|
try {
|
|
4177
4284
|
const result = await this.#client.execute({
|