@mastra/pg 1.8.2-alpha.0 → 1.8.3-alpha.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -4327,7 +4327,11 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4327
4327
  });
4328
4328
  await this.#db.createTable({ tableName: TABLE_DATASET_VERSIONS, schema: DATASET_VERSIONS_SCHEMA });
4329
4329
  await this.#addColumnIfNotExists(TABLE_DATASETS, "requestContextSchema", "JSONB");
4330
+ await this.#addColumnIfNotExists(TABLE_DATASETS, "tags", "JSONB");
4331
+ await this.#addColumnIfNotExists(TABLE_DATASETS, "targetType", "TEXT");
4332
+ await this.#addColumnIfNotExists(TABLE_DATASETS, "targetIds", "JSONB");
4330
4333
  await this.#addColumnIfNotExists(TABLE_DATASET_ITEMS, "requestContext", "JSONB");
4334
+ await this.#addColumnIfNotExists(TABLE_DATASET_ITEMS, "source", "JSONB");
4331
4335
  await this.createDefaultIndexes();
4332
4336
  await this.createCustomIndexes();
4333
4337
  }
@@ -4394,6 +4398,9 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4394
4398
  inputSchema: row.inputSchema ? safelyParseJSON(row.inputSchema) : void 0,
4395
4399
  groundTruthSchema: row.groundTruthSchema ? safelyParseJSON(row.groundTruthSchema) : void 0,
4396
4400
  requestContextSchema: row.requestContextSchema ? safelyParseJSON(row.requestContextSchema) : void 0,
4401
+ tags: row.tags ? safelyParseJSON(row.tags) : void 0,
4402
+ targetType: row.targetType || null,
4403
+ targetIds: row.targetIds || null,
4397
4404
  version: row.version,
4398
4405
  createdAt: ensureDate(row.createdAtZ || row.createdAt),
4399
4406
  updatedAt: ensureDate(row.updatedAtZ || row.updatedAt)
@@ -4408,6 +4415,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4408
4415
  groundTruth: row.groundTruth ? safelyParseJSON(row.groundTruth) : void 0,
4409
4416
  requestContext: row.requestContext ? safelyParseJSON(row.requestContext) : void 0,
4410
4417
  metadata: row.metadata ? safelyParseJSON(row.metadata) : void 0,
4418
+ source: row.source ? safelyParseJSON(row.source) : void 0,
4411
4419
  createdAt: ensureDate(row.createdAtZ || row.createdAt),
4412
4420
  updatedAt: ensureDate(row.updatedAtZ || row.updatedAt)
4413
4421
  };
@@ -4423,6 +4431,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4423
4431
  groundTruth: row.groundTruth ? safelyParseJSON(row.groundTruth) : void 0,
4424
4432
  requestContext: row.requestContext ? safelyParseJSON(row.requestContext) : void 0,
4425
4433
  metadata: row.metadata ? safelyParseJSON(row.metadata) : void 0,
4434
+ source: row.source ? safelyParseJSON(row.source) : void 0,
4426
4435
  createdAt: ensureDate(row.createdAtZ || row.createdAt),
4427
4436
  updatedAt: ensureDate(row.updatedAtZ || row.updatedAt)
4428
4437
  };
@@ -4451,6 +4460,8 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4451
4460
  inputSchema: input.inputSchema ?? null,
4452
4461
  groundTruthSchema: input.groundTruthSchema ?? null,
4453
4462
  requestContextSchema: input.requestContextSchema ?? null,
4463
+ targetType: input.targetType ?? null,
4464
+ targetIds: input.targetIds !== void 0 ? JSON.stringify(input.targetIds) : null,
4454
4465
  version: 0,
4455
4466
  createdAt: nowIso,
4456
4467
  updatedAt: nowIso
@@ -4464,6 +4475,8 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4464
4475
  inputSchema: input.inputSchema ?? void 0,
4465
4476
  groundTruthSchema: input.groundTruthSchema ?? void 0,
4466
4477
  requestContextSchema: input.requestContextSchema ?? void 0,
4478
+ targetType: input.targetType ?? null,
4479
+ targetIds: input.targetIds ?? null,
4467
4480
  version: 0,
4468
4481
  createdAt: now,
4469
4482
  updatedAt: now
@@ -4535,6 +4548,18 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4535
4548
  setClauses.push(`"requestContextSchema" = $${paramIndex++}`);
4536
4549
  values.push(args.requestContextSchema === null ? null : JSON.stringify(args.requestContextSchema));
4537
4550
  }
4551
+ if (args.tags !== void 0) {
4552
+ setClauses.push(`"tags" = $${paramIndex++}`);
4553
+ values.push(args.tags === null ? null : JSON.stringify(args.tags));
4554
+ }
4555
+ if (args.targetType !== void 0) {
4556
+ setClauses.push(`"targetType" = $${paramIndex++}`);
4557
+ values.push(args.targetType);
4558
+ }
4559
+ if (args.targetIds !== void 0) {
4560
+ setClauses.push(`"targetIds" = $${paramIndex++}`);
4561
+ values.push(args.targetIds === null ? null : JSON.stringify(args.targetIds));
4562
+ }
4538
4563
  values.push(args.id);
4539
4564
  await this.#db.client.none(
4540
4565
  `UPDATE ${tableName} SET ${setClauses.join(", ")} WHERE "id" = $${paramIndex}`,
@@ -4548,6 +4573,9 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4548
4573
  inputSchema: (args.inputSchema !== void 0 ? args.inputSchema : existing.inputSchema) ?? void 0,
4549
4574
  groundTruthSchema: (args.groundTruthSchema !== void 0 ? args.groundTruthSchema : existing.groundTruthSchema) ?? void 0,
4550
4575
  requestContextSchema: (args.requestContextSchema !== void 0 ? args.requestContextSchema : existing.requestContextSchema) ?? void 0,
4576
+ tags: (args.tags !== void 0 ? args.tags : existing.tags) ?? void 0,
4577
+ targetType: (args.targetType !== void 0 ? args.targetType : existing.targetType) ?? null,
4578
+ targetIds: (args.targetIds !== void 0 ? args.targetIds : existing.targetIds) ?? null,
4551
4579
  updatedAt: new Date(now)
4552
4580
  };
4553
4581
  } catch (error) {
@@ -4663,7 +4691,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4663
4691
  );
4664
4692
  newVersion = row.version;
4665
4693
  await t.none(
4666
- `INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,false,$4,$5,$6,$7,$8,$9,$10,$11)`,
4694
+ `INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,false,$4,$5,$6,$7,$8,$9,$10,$11,$12)`,
4667
4695
  [
4668
4696
  id,
4669
4697
  args.datasetId,
@@ -4672,6 +4700,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4672
4700
  jsonbArg(args.groundTruth),
4673
4701
  jsonbArg(args.requestContext),
4674
4702
  jsonbArg(args.metadata),
4703
+ jsonbArg(args.source),
4675
4704
  nowIso,
4676
4705
  nowIso,
4677
4706
  nowIso,
@@ -4691,6 +4720,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4691
4720
  groundTruth: args.groundTruth,
4692
4721
  requestContext: args.requestContext,
4693
4722
  metadata: args.metadata,
4723
+ source: args.source,
4694
4724
  createdAt: now,
4695
4725
  updatedAt: now
4696
4726
  };
@@ -4738,6 +4768,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4738
4768
  const mergedGroundTruth = args.groundTruth ?? existing.groundTruth;
4739
4769
  const mergedRequestContext = args.requestContext ?? existing.requestContext;
4740
4770
  const mergedMetadata = args.metadata ?? existing.metadata;
4771
+ const mergedSource = args.source ?? existing.source;
4741
4772
  let newVersion;
4742
4773
  await this.#db.client.tx(async (t) => {
4743
4774
  const row = await t.one(
@@ -4750,7 +4781,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4750
4781
  [newVersion, args.id]
4751
4782
  );
4752
4783
  await t.none(
4753
- `INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,false,$4,$5,$6,$7,$8,$9,$10,$11)`,
4784
+ `INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,false,$4,$5,$6,$7,$8,$9,$10,$11,$12)`,
4754
4785
  [
4755
4786
  args.id,
4756
4787
  args.datasetId,
@@ -4759,6 +4790,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4759
4790
  jsonbArg(mergedGroundTruth),
4760
4791
  jsonbArg(mergedRequestContext),
4761
4792
  jsonbArg(mergedMetadata),
4793
+ jsonbArg(mergedSource),
4762
4794
  existing.createdAt.toISOString(),
4763
4795
  existing.createdAt.toISOString(),
4764
4796
  nowIso,
@@ -4777,6 +4809,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4777
4809
  groundTruth: mergedGroundTruth,
4778
4810
  requestContext: mergedRequestContext,
4779
4811
  metadata: mergedMetadata,
4812
+ source: mergedSource,
4780
4813
  updatedAt: now
4781
4814
  };
4782
4815
  } catch (error) {
@@ -4822,7 +4855,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4822
4855
  [newVersion, id]
4823
4856
  );
4824
4857
  await t.none(
4825
- `INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,true,$4,$5,$6,$7,$8,$9,$10,$11)`,
4858
+ `INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,true,$4,$5,$6,$7,$8,$9,$10,$11,$12)`,
4826
4859
  [
4827
4860
  id,
4828
4861
  datasetId,
@@ -4831,6 +4864,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4831
4864
  jsonbArg(existing.groundTruth),
4832
4865
  jsonbArg(existing.requestContext),
4833
4866
  jsonbArg(existing.metadata),
4867
+ jsonbArg(existing.source),
4834
4868
  existing.createdAt.toISOString(),
4835
4869
  existing.createdAt.toISOString(),
4836
4870
  nowIso,
@@ -4884,7 +4918,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4884
4918
  newVersion = row.version;
4885
4919
  for (const { id, input: itemInput } of itemsWithIds) {
4886
4920
  await t.none(
4887
- `INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,false,$4,$5,$6,$7,$8,$9,$10,$11)`,
4921
+ `INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,false,$4,$5,$6,$7,$8,$9,$10,$11,$12)`,
4888
4922
  [
4889
4923
  id,
4890
4924
  input.datasetId,
@@ -4893,6 +4927,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4893
4927
  jsonbArg(itemInput.groundTruth),
4894
4928
  jsonbArg(itemInput.requestContext),
4895
4929
  jsonbArg(itemInput.metadata),
4930
+ jsonbArg(itemInput.source),
4896
4931
  nowIso,
4897
4932
  nowIso,
4898
4933
  nowIso,
@@ -4913,6 +4948,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4913
4948
  groundTruth: itemInput.groundTruth,
4914
4949
  requestContext: itemInput.requestContext,
4915
4950
  metadata: itemInput.metadata,
4951
+ source: itemInput.source,
4916
4952
  createdAt: now,
4917
4953
  updatedAt: now
4918
4954
  }));
@@ -4967,7 +5003,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4967
5003
  [newVersion, item.id]
4968
5004
  );
4969
5005
  await t.none(
4970
- `INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,true,$4,$5,$6,$7,$8,$9,$10,$11)`,
5006
+ `INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,true,$4,$5,$6,$7,$8,$9,$10,$11,$12)`,
4971
5007
  [
4972
5008
  item.id,
4973
5009
  input.datasetId,
@@ -4976,6 +5012,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
4976
5012
  jsonbArg(item.groundTruth),
4977
5013
  jsonbArg(item.requestContext),
4978
5014
  jsonbArg(item.metadata),
5015
+ jsonbArg(item.source),
4979
5016
  item.createdAt.toISOString(),
4980
5017
  item.createdAt.toISOString(),
4981
5018
  nowIso,
@@ -5296,6 +5333,8 @@ var ExperimentsPG = class _ExperimentsPG extends ExperimentsStorage {
5296
5333
  completedAt: ensureDate(row.completedAtZ || row.completedAt),
5297
5334
  retryCount: row.retryCount,
5298
5335
  traceId: row.traceId ?? null,
5336
+ status: row.status ?? null,
5337
+ tags: row.tags ? safelyParseJSON(row.tags) : null,
5299
5338
  createdAt: ensureDate(row.createdAtZ || row.createdAt)
5300
5339
  };
5301
5340
  }
@@ -5536,6 +5575,8 @@ var ExperimentsPG = class _ExperimentsPG extends ExperimentsStorage {
5536
5575
  completedAt: input.completedAt.toISOString(),
5537
5576
  retryCount: input.retryCount,
5538
5577
  traceId: input.traceId ?? null,
5578
+ status: input.status ?? null,
5579
+ tags: input.tags ?? null,
5539
5580
  createdAt: nowIso
5540
5581
  }
5541
5582
  });
@@ -5552,6 +5593,8 @@ var ExperimentsPG = class _ExperimentsPG extends ExperimentsStorage {
5552
5593
  completedAt: input.completedAt,
5553
5594
  retryCount: input.retryCount,
5554
5595
  traceId: input.traceId ?? null,
5596
+ status: input.status ?? null,
5597
+ tags: input.tags ?? null,
5555
5598
  createdAt: now
5556
5599
  };
5557
5600
  } catch (error) {
@@ -5565,6 +5608,64 @@ var ExperimentsPG = class _ExperimentsPG extends ExperimentsStorage {
5565
5608
  );
5566
5609
  }
5567
5610
  }
5611
+ async updateExperimentResult(input) {
5612
+ try {
5613
+ const tableName = getTableName2({ indexName: TABLE_EXPERIMENT_RESULTS, schemaName: getSchemaName2(this.#schema) });
5614
+ const setClauses = [];
5615
+ const values = [];
5616
+ let paramIndex = 1;
5617
+ if (input.status !== void 0) {
5618
+ setClauses.push(`"status" = $${paramIndex++}`);
5619
+ values.push(input.status);
5620
+ }
5621
+ if (input.tags !== void 0) {
5622
+ setClauses.push(`"tags" = $${paramIndex++}`);
5623
+ values.push(JSON.stringify(input.tags));
5624
+ }
5625
+ if (setClauses.length === 0) {
5626
+ const existing = await this.getExperimentResultById({ id: input.id });
5627
+ if (!existing) {
5628
+ throw new MastraError({
5629
+ id: createStorageErrorId("PG", "UPDATE_EXPERIMENT_RESULT", "NOT_FOUND"),
5630
+ domain: ErrorDomain.STORAGE,
5631
+ category: ErrorCategory.USER,
5632
+ details: { resultId: input.id }
5633
+ });
5634
+ }
5635
+ return existing;
5636
+ }
5637
+ values.push(input.id);
5638
+ let whereClause = `"id" = $${paramIndex}`;
5639
+ if (input.experimentId) {
5640
+ paramIndex++;
5641
+ values.push(input.experimentId);
5642
+ whereClause += ` AND "experimentId" = $${paramIndex}`;
5643
+ }
5644
+ const row = await this.#db.client.oneOrNone(
5645
+ `UPDATE ${tableName} SET ${setClauses.join(", ")} WHERE ${whereClause} RETURNING *`,
5646
+ values
5647
+ );
5648
+ if (!row) {
5649
+ throw new MastraError({
5650
+ id: createStorageErrorId("PG", "UPDATE_EXPERIMENT_RESULT", "NOT_FOUND"),
5651
+ domain: ErrorDomain.STORAGE,
5652
+ category: ErrorCategory.USER,
5653
+ details: { resultId: input.id }
5654
+ });
5655
+ }
5656
+ return this.transformExperimentResultRow(row);
5657
+ } catch (error) {
5658
+ if (error instanceof MastraError) throw error;
5659
+ throw new MastraError(
5660
+ {
5661
+ id: createStorageErrorId("PG", "UPDATE_EXPERIMENT_RESULT", "FAILED"),
5662
+ domain: ErrorDomain.STORAGE,
5663
+ category: ErrorCategory.THIRD_PARTY
5664
+ },
5665
+ error
5666
+ );
5667
+ }
5668
+ }
5568
5669
  async getExperimentResultById({ id }) {
5569
5670
  try {
5570
5671
  const tableName = getTableName2({ indexName: TABLE_EXPERIMENT_RESULTS, schemaName: getSchemaName2(this.#schema) });