@mastra/libsql 1.7.3 → 1.7.4-alpha.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +9 -0
- package/dist/docs/SKILL.md +1 -3
- package/dist/docs/assets/SOURCE_MAP.json +1 -1
- package/dist/docs/references/docs-memory-message-history.md +6 -4
- package/dist/docs/references/docs-memory-overview.md +4 -4
- package/dist/docs/references/docs-memory-semantic-recall.md +28 -19
- package/dist/docs/references/docs-memory-storage.md +4 -4
- package/dist/docs/references/reference-storage-dynamodb.md +1 -1
- package/dist/index.cjs +20 -9
- package/dist/index.cjs.map +1 -1
- package/dist/index.js +20 -9
- package/dist/index.js.map +1 -1
- package/dist/storage/domains/datasets/index.d.ts.map +1 -1
- package/dist/storage/domains/memory/index.d.ts.map +1 -1
- package/package.json +5 -5
- package/dist/docs/references/docs-observability-overview.md +0 -70
- package/dist/docs/references/docs-observability-tracing-exporters-default.md +0 -209
package/dist/index.js
CHANGED
|
@@ -2991,6 +2991,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
2991
2991
|
await this.#addColumnIfNotExists(TABLE_DATASETS, "scorerIds", "TEXT");
|
|
2992
2992
|
await this.#addColumnIfNotExists(TABLE_DATASET_ITEMS, "requestContext", "TEXT");
|
|
2993
2993
|
await this.#addColumnIfNotExists(TABLE_DATASET_ITEMS, "source", "TEXT");
|
|
2994
|
+
await this.#addColumnIfNotExists(TABLE_DATASET_ITEMS, "expectedTrajectory", "TEXT");
|
|
2994
2995
|
await this.#client.execute({
|
|
2995
2996
|
sql: `CREATE INDEX IF NOT EXISTS idx_dataset_items_dataset_validto ON "${TABLE_DATASET_ITEMS}" ("datasetId", "validTo")`,
|
|
2996
2997
|
args: []
|
|
@@ -3049,6 +3050,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3049
3050
|
datasetVersion: row.datasetVersion,
|
|
3050
3051
|
input: safelyParseJSON(row.input),
|
|
3051
3052
|
groundTruth: row.groundTruth ? safelyParseJSON(row.groundTruth) : void 0,
|
|
3053
|
+
expectedTrajectory: row.expectedTrajectory ? safelyParseJSON(row.expectedTrajectory) : void 0,
|
|
3052
3054
|
requestContext: row.requestContext ? safelyParseJSON(row.requestContext) : void 0,
|
|
3053
3055
|
metadata: row.metadata ? safelyParseJSON(row.metadata) : void 0,
|
|
3054
3056
|
source: row.source ? safelyParseJSON(row.source) : void 0,
|
|
@@ -3065,6 +3067,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3065
3067
|
isDeleted: Boolean(row.isDeleted),
|
|
3066
3068
|
input: safelyParseJSON(row.input),
|
|
3067
3069
|
groundTruth: row.groundTruth ? safelyParseJSON(row.groundTruth) : void 0,
|
|
3070
|
+
expectedTrajectory: row.expectedTrajectory ? safelyParseJSON(row.expectedTrajectory) : void 0,
|
|
3068
3071
|
requestContext: row.requestContext ? safelyParseJSON(row.requestContext) : void 0,
|
|
3069
3072
|
metadata: row.metadata ? safelyParseJSON(row.metadata) : void 0,
|
|
3070
3073
|
source: row.source ? safelyParseJSON(row.source) : void 0,
|
|
@@ -3324,13 +3327,14 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3324
3327
|
args: [args.datasetId]
|
|
3325
3328
|
},
|
|
3326
3329
|
{
|
|
3327
|
-
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3330
|
+
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, expectedTrajectory, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3328
3331
|
args: [
|
|
3329
3332
|
id,
|
|
3330
3333
|
args.datasetId,
|
|
3331
3334
|
args.datasetId,
|
|
3332
3335
|
jsonbArg(args.input),
|
|
3333
3336
|
jsonbArg(args.groundTruth),
|
|
3337
|
+
jsonbArg(args.expectedTrajectory),
|
|
3334
3338
|
jsonbArg(args.requestContext),
|
|
3335
3339
|
jsonbArg(args.metadata),
|
|
3336
3340
|
jsonbArg(args.source),
|
|
@@ -3352,6 +3356,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3352
3356
|
datasetVersion: newVersion,
|
|
3353
3357
|
input: args.input,
|
|
3354
3358
|
groundTruth: args.groundTruth,
|
|
3359
|
+
expectedTrajectory: args.expectedTrajectory,
|
|
3355
3360
|
requestContext: args.requestContext,
|
|
3356
3361
|
metadata: args.metadata,
|
|
3357
3362
|
source: args.source,
|
|
@@ -3392,11 +3397,12 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3392
3397
|
const versionId = crypto.randomUUID();
|
|
3393
3398
|
const now = /* @__PURE__ */ new Date();
|
|
3394
3399
|
const nowIso = now.toISOString();
|
|
3395
|
-
const mergedInput = args.input
|
|
3396
|
-
const mergedGroundTruth = args.groundTruth
|
|
3397
|
-
const
|
|
3398
|
-
const
|
|
3399
|
-
const
|
|
3400
|
+
const mergedInput = args.input !== void 0 ? args.input : existing.input;
|
|
3401
|
+
const mergedGroundTruth = args.groundTruth !== void 0 ? args.groundTruth : existing.groundTruth;
|
|
3402
|
+
const mergedExpectedTrajectory = args.expectedTrajectory !== void 0 ? args.expectedTrajectory : existing.expectedTrajectory;
|
|
3403
|
+
const mergedRequestContext = args.requestContext !== void 0 ? args.requestContext : existing.requestContext;
|
|
3404
|
+
const mergedMetadata = args.metadata !== void 0 ? args.metadata : existing.metadata;
|
|
3405
|
+
const mergedSource = args.source !== void 0 ? args.source : existing.source;
|
|
3400
3406
|
const results = await this.#client.batch(
|
|
3401
3407
|
[
|
|
3402
3408
|
{
|
|
@@ -3408,13 +3414,14 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3408
3414
|
args: [args.datasetId, args.id]
|
|
3409
3415
|
},
|
|
3410
3416
|
{
|
|
3411
|
-
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3417
|
+
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, expectedTrajectory, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3412
3418
|
args: [
|
|
3413
3419
|
args.id,
|
|
3414
3420
|
args.datasetId,
|
|
3415
3421
|
args.datasetId,
|
|
3416
3422
|
jsonbArg(mergedInput),
|
|
3417
3423
|
jsonbArg(mergedGroundTruth),
|
|
3424
|
+
jsonbArg(mergedExpectedTrajectory),
|
|
3418
3425
|
jsonbArg(mergedRequestContext),
|
|
3419
3426
|
jsonbArg(mergedMetadata),
|
|
3420
3427
|
jsonbArg(mergedSource),
|
|
@@ -3435,6 +3442,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3435
3442
|
datasetVersion: newVersion,
|
|
3436
3443
|
input: mergedInput,
|
|
3437
3444
|
groundTruth: mergedGroundTruth,
|
|
3445
|
+
expectedTrajectory: mergedExpectedTrajectory,
|
|
3438
3446
|
requestContext: mergedRequestContext,
|
|
3439
3447
|
metadata: mergedMetadata,
|
|
3440
3448
|
source: mergedSource,
|
|
@@ -3766,13 +3774,14 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3766
3774
|
const id = crypto.randomUUID();
|
|
3767
3775
|
items.push({ id, input: itemInput });
|
|
3768
3776
|
statements.push({
|
|
3769
|
-
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3777
|
+
sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, expectedTrajectory, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
|
|
3770
3778
|
args: [
|
|
3771
3779
|
id,
|
|
3772
3780
|
input.datasetId,
|
|
3773
3781
|
input.datasetId,
|
|
3774
3782
|
jsonbArg(itemInput.input),
|
|
3775
3783
|
jsonbArg(itemInput.groundTruth),
|
|
3784
|
+
jsonbArg(itemInput.expectedTrajectory),
|
|
3776
3785
|
jsonbArg(itemInput.requestContext),
|
|
3777
3786
|
jsonbArg(itemInput.metadata),
|
|
3778
3787
|
jsonbArg(itemInput.source),
|
|
@@ -3793,6 +3802,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
|
|
|
3793
3802
|
datasetVersion: newVersion,
|
|
3794
3803
|
input: itemInput.input,
|
|
3795
3804
|
groundTruth: itemInput.groundTruth,
|
|
3805
|
+
expectedTrajectory: itemInput.expectedTrajectory,
|
|
3796
3806
|
requestContext: itemInput.requestContext,
|
|
3797
3807
|
metadata: itemInput.metadata,
|
|
3798
3808
|
source: itemInput.source,
|
|
@@ -7061,7 +7071,8 @@ var MemoryLibSQL = class extends MemoryStorage {
|
|
|
7061
7071
|
lastObservedAt: input.chunk.lastObservedAt,
|
|
7062
7072
|
createdAt: /* @__PURE__ */ new Date(),
|
|
7063
7073
|
suggestedContinuation: input.chunk.suggestedContinuation,
|
|
7064
|
-
currentTask: input.chunk.currentTask
|
|
7074
|
+
currentTask: input.chunk.currentTask,
|
|
7075
|
+
threadTitle: input.chunk.threadTitle
|
|
7065
7076
|
};
|
|
7066
7077
|
const newChunks = [...existingChunks, newChunk];
|
|
7067
7078
|
const lastBufferedAtTime = input.lastBufferedAtTime ? input.lastBufferedAtTime.toISOString() : null;
|