@mastra/libsql 1.7.3 → 1.7.4-alpha.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -2991,6 +2991,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
2991
2991
  await this.#addColumnIfNotExists(TABLE_DATASETS, "scorerIds", "TEXT");
2992
2992
  await this.#addColumnIfNotExists(TABLE_DATASET_ITEMS, "requestContext", "TEXT");
2993
2993
  await this.#addColumnIfNotExists(TABLE_DATASET_ITEMS, "source", "TEXT");
2994
+ await this.#addColumnIfNotExists(TABLE_DATASET_ITEMS, "expectedTrajectory", "TEXT");
2994
2995
  await this.#client.execute({
2995
2996
  sql: `CREATE INDEX IF NOT EXISTS idx_dataset_items_dataset_validto ON "${TABLE_DATASET_ITEMS}" ("datasetId", "validTo")`,
2996
2997
  args: []
@@ -3049,6 +3050,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
3049
3050
  datasetVersion: row.datasetVersion,
3050
3051
  input: safelyParseJSON(row.input),
3051
3052
  groundTruth: row.groundTruth ? safelyParseJSON(row.groundTruth) : void 0,
3053
+ expectedTrajectory: row.expectedTrajectory ? safelyParseJSON(row.expectedTrajectory) : void 0,
3052
3054
  requestContext: row.requestContext ? safelyParseJSON(row.requestContext) : void 0,
3053
3055
  metadata: row.metadata ? safelyParseJSON(row.metadata) : void 0,
3054
3056
  source: row.source ? safelyParseJSON(row.source) : void 0,
@@ -3065,6 +3067,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
3065
3067
  isDeleted: Boolean(row.isDeleted),
3066
3068
  input: safelyParseJSON(row.input),
3067
3069
  groundTruth: row.groundTruth ? safelyParseJSON(row.groundTruth) : void 0,
3070
+ expectedTrajectory: row.expectedTrajectory ? safelyParseJSON(row.expectedTrajectory) : void 0,
3068
3071
  requestContext: row.requestContext ? safelyParseJSON(row.requestContext) : void 0,
3069
3072
  metadata: row.metadata ? safelyParseJSON(row.metadata) : void 0,
3070
3073
  source: row.source ? safelyParseJSON(row.source) : void 0,
@@ -3324,13 +3327,14 @@ var DatasetsLibSQL = class extends DatasetsStorage {
3324
3327
  args: [args.datasetId]
3325
3328
  },
3326
3329
  {
3327
- sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
3330
+ sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, expectedTrajectory, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
3328
3331
  args: [
3329
3332
  id,
3330
3333
  args.datasetId,
3331
3334
  args.datasetId,
3332
3335
  jsonbArg(args.input),
3333
3336
  jsonbArg(args.groundTruth),
3337
+ jsonbArg(args.expectedTrajectory),
3334
3338
  jsonbArg(args.requestContext),
3335
3339
  jsonbArg(args.metadata),
3336
3340
  jsonbArg(args.source),
@@ -3352,6 +3356,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
3352
3356
  datasetVersion: newVersion,
3353
3357
  input: args.input,
3354
3358
  groundTruth: args.groundTruth,
3359
+ expectedTrajectory: args.expectedTrajectory,
3355
3360
  requestContext: args.requestContext,
3356
3361
  metadata: args.metadata,
3357
3362
  source: args.source,
@@ -3392,11 +3397,12 @@ var DatasetsLibSQL = class extends DatasetsStorage {
3392
3397
  const versionId = crypto.randomUUID();
3393
3398
  const now = /* @__PURE__ */ new Date();
3394
3399
  const nowIso = now.toISOString();
3395
- const mergedInput = args.input ?? existing.input;
3396
- const mergedGroundTruth = args.groundTruth ?? existing.groundTruth;
3397
- const mergedRequestContext = args.requestContext ?? existing.requestContext;
3398
- const mergedMetadata = args.metadata ?? existing.metadata;
3399
- const mergedSource = args.source ?? existing.source;
3400
+ const mergedInput = args.input !== void 0 ? args.input : existing.input;
3401
+ const mergedGroundTruth = args.groundTruth !== void 0 ? args.groundTruth : existing.groundTruth;
3402
+ const mergedExpectedTrajectory = args.expectedTrajectory !== void 0 ? args.expectedTrajectory : existing.expectedTrajectory;
3403
+ const mergedRequestContext = args.requestContext !== void 0 ? args.requestContext : existing.requestContext;
3404
+ const mergedMetadata = args.metadata !== void 0 ? args.metadata : existing.metadata;
3405
+ const mergedSource = args.source !== void 0 ? args.source : existing.source;
3400
3406
  const results = await this.#client.batch(
3401
3407
  [
3402
3408
  {
@@ -3408,13 +3414,14 @@ var DatasetsLibSQL = class extends DatasetsStorage {
3408
3414
  args: [args.datasetId, args.id]
3409
3415
  },
3410
3416
  {
3411
- sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
3417
+ sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, expectedTrajectory, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
3412
3418
  args: [
3413
3419
  args.id,
3414
3420
  args.datasetId,
3415
3421
  args.datasetId,
3416
3422
  jsonbArg(mergedInput),
3417
3423
  jsonbArg(mergedGroundTruth),
3424
+ jsonbArg(mergedExpectedTrajectory),
3418
3425
  jsonbArg(mergedRequestContext),
3419
3426
  jsonbArg(mergedMetadata),
3420
3427
  jsonbArg(mergedSource),
@@ -3435,6 +3442,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
3435
3442
  datasetVersion: newVersion,
3436
3443
  input: mergedInput,
3437
3444
  groundTruth: mergedGroundTruth,
3445
+ expectedTrajectory: mergedExpectedTrajectory,
3438
3446
  requestContext: mergedRequestContext,
3439
3447
  metadata: mergedMetadata,
3440
3448
  source: mergedSource,
@@ -3766,13 +3774,14 @@ var DatasetsLibSQL = class extends DatasetsStorage {
3766
3774
  const id = crypto.randomUUID();
3767
3775
  items.push({ id, input: itemInput });
3768
3776
  statements.push({
3769
- sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
3777
+ sql: `INSERT INTO ${TABLE_DATASET_ITEMS} (id, datasetId, datasetVersion, validTo, isDeleted, input, groundTruth, expectedTrajectory, requestContext, metadata, source, createdAt, updatedAt) VALUES (?, ?, (SELECT version FROM ${TABLE_DATASETS} WHERE id = ?), NULL, 0, jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), jsonb(?), ?, ?)`,
3770
3778
  args: [
3771
3779
  id,
3772
3780
  input.datasetId,
3773
3781
  input.datasetId,
3774
3782
  jsonbArg(itemInput.input),
3775
3783
  jsonbArg(itemInput.groundTruth),
3784
+ jsonbArg(itemInput.expectedTrajectory),
3776
3785
  jsonbArg(itemInput.requestContext),
3777
3786
  jsonbArg(itemInput.metadata),
3778
3787
  jsonbArg(itemInput.source),
@@ -3793,6 +3802,7 @@ var DatasetsLibSQL = class extends DatasetsStorage {
3793
3802
  datasetVersion: newVersion,
3794
3803
  input: itemInput.input,
3795
3804
  groundTruth: itemInput.groundTruth,
3805
+ expectedTrajectory: itemInput.expectedTrajectory,
3796
3806
  requestContext: itemInput.requestContext,
3797
3807
  metadata: itemInput.metadata,
3798
3808
  source: itemInput.source,
@@ -7061,7 +7071,8 @@ var MemoryLibSQL = class extends MemoryStorage {
7061
7071
  lastObservedAt: input.chunk.lastObservedAt,
7062
7072
  createdAt: /* @__PURE__ */ new Date(),
7063
7073
  suggestedContinuation: input.chunk.suggestedContinuation,
7064
- currentTask: input.chunk.currentTask
7074
+ currentTask: input.chunk.currentTask,
7075
+ threadTitle: input.chunk.threadTitle
7065
7076
  };
7066
7077
  const newChunks = [...existingChunks, newChunk];
7067
7078
  const lastBufferedAtTime = input.lastBufferedAtTime ? input.lastBufferedAtTime.toISOString() : null;