@mastra/pg 1.8.5 → 1.8.6
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +18 -0
- package/dist/docs/SKILL.md +1 -1
- package/dist/docs/assets/SOURCE_MAP.json +1 -1
- package/dist/index.cjs +24 -11
- package/dist/index.cjs.map +1 -1
- package/dist/index.js +24 -11
- package/dist/index.js.map +1 -1
- package/dist/storage/domains/datasets/index.d.ts.map +1 -1
- package/dist/storage/domains/memory/index.d.ts.map +1 -1
- package/package.json +5 -5
package/dist/index.js
CHANGED
|
@@ -4333,6 +4333,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
|
|
|
4333
4333
|
await this.#addColumnIfNotExists(TABLE_DATASETS, "scorerIds", "JSONB");
|
|
4334
4334
|
await this.#addColumnIfNotExists(TABLE_DATASET_ITEMS, "requestContext", "JSONB");
|
|
4335
4335
|
await this.#addColumnIfNotExists(TABLE_DATASET_ITEMS, "source", "JSONB");
|
|
4336
|
+
await this.#addColumnIfNotExists(TABLE_DATASET_ITEMS, "expectedTrajectory", "JSONB");
|
|
4336
4337
|
await this.createDefaultIndexes();
|
|
4337
4338
|
await this.createCustomIndexes();
|
|
4338
4339
|
}
|
|
@@ -4415,6 +4416,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
|
|
|
4415
4416
|
datasetVersion: row.datasetVersion,
|
|
4416
4417
|
input: safelyParseJSON(row.input),
|
|
4417
4418
|
groundTruth: row.groundTruth ? safelyParseJSON(row.groundTruth) : void 0,
|
|
4419
|
+
expectedTrajectory: row.expectedTrajectory ? safelyParseJSON(row.expectedTrajectory) : void 0,
|
|
4418
4420
|
requestContext: row.requestContext ? safelyParseJSON(row.requestContext) : void 0,
|
|
4419
4421
|
metadata: row.metadata ? safelyParseJSON(row.metadata) : void 0,
|
|
4420
4422
|
source: row.source ? safelyParseJSON(row.source) : void 0,
|
|
@@ -4431,6 +4433,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
|
|
|
4431
4433
|
isDeleted: Boolean(row.isDeleted),
|
|
4432
4434
|
input: safelyParseJSON(row.input),
|
|
4433
4435
|
groundTruth: row.groundTruth ? safelyParseJSON(row.groundTruth) : void 0,
|
|
4436
|
+
expectedTrajectory: row.expectedTrajectory ? safelyParseJSON(row.expectedTrajectory) : void 0,
|
|
4434
4437
|
requestContext: row.requestContext ? safelyParseJSON(row.requestContext) : void 0,
|
|
4435
4438
|
metadata: row.metadata ? safelyParseJSON(row.metadata) : void 0,
|
|
4436
4439
|
source: row.source ? safelyParseJSON(row.source) : void 0,
|
|
@@ -4700,13 +4703,14 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
|
|
|
4700
4703
|
);
|
|
4701
4704
|
newVersion = row.version;
|
|
4702
4705
|
await t.none(
|
|
4703
|
-
`INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,false,$4,$5,$6,$7,$8,$9,$10,$11,$12)`,
|
|
4706
|
+
`INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","expectedTrajectory","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,false,$4,$5,$6,$7,$8,$9,$10,$11,$12,$13)`,
|
|
4704
4707
|
[
|
|
4705
4708
|
id,
|
|
4706
4709
|
args.datasetId,
|
|
4707
4710
|
newVersion,
|
|
4708
4711
|
JSON.stringify(args.input),
|
|
4709
4712
|
jsonbArg(args.groundTruth),
|
|
4713
|
+
jsonbArg(args.expectedTrajectory),
|
|
4710
4714
|
jsonbArg(args.requestContext),
|
|
4711
4715
|
jsonbArg(args.metadata),
|
|
4712
4716
|
jsonbArg(args.source),
|
|
@@ -4727,6 +4731,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
|
|
|
4727
4731
|
datasetVersion: newVersion,
|
|
4728
4732
|
input: args.input,
|
|
4729
4733
|
groundTruth: args.groundTruth,
|
|
4734
|
+
expectedTrajectory: args.expectedTrajectory,
|
|
4730
4735
|
requestContext: args.requestContext,
|
|
4731
4736
|
metadata: args.metadata,
|
|
4732
4737
|
source: args.source,
|
|
@@ -4773,11 +4778,12 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
|
|
|
4773
4778
|
const versionId = crypto.randomUUID();
|
|
4774
4779
|
const now = /* @__PURE__ */ new Date();
|
|
4775
4780
|
const nowIso = now.toISOString();
|
|
4776
|
-
const mergedInput = args.input
|
|
4777
|
-
const mergedGroundTruth = args.groundTruth
|
|
4778
|
-
const
|
|
4779
|
-
const
|
|
4780
|
-
const
|
|
4781
|
+
const mergedInput = args.input !== void 0 ? args.input : existing.input;
|
|
4782
|
+
const mergedGroundTruth = args.groundTruth !== void 0 ? args.groundTruth : existing.groundTruth;
|
|
4783
|
+
const mergedExpectedTrajectory = args.expectedTrajectory !== void 0 ? args.expectedTrajectory : existing.expectedTrajectory;
|
|
4784
|
+
const mergedRequestContext = args.requestContext !== void 0 ? args.requestContext : existing.requestContext;
|
|
4785
|
+
const mergedMetadata = args.metadata !== void 0 ? args.metadata : existing.metadata;
|
|
4786
|
+
const mergedSource = args.source !== void 0 ? args.source : existing.source;
|
|
4781
4787
|
let newVersion;
|
|
4782
4788
|
await this.#db.client.tx(async (t) => {
|
|
4783
4789
|
const row = await t.one(
|
|
@@ -4790,13 +4796,14 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
|
|
|
4790
4796
|
[newVersion, args.id]
|
|
4791
4797
|
);
|
|
4792
4798
|
await t.none(
|
|
4793
|
-
`INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,false,$4,$5,$6,$7,$8,$9,$10,$11,$12)`,
|
|
4799
|
+
`INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","expectedTrajectory","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,false,$4,$5,$6,$7,$8,$9,$10,$11,$12,$13)`,
|
|
4794
4800
|
[
|
|
4795
4801
|
args.id,
|
|
4796
4802
|
args.datasetId,
|
|
4797
4803
|
newVersion,
|
|
4798
4804
|
JSON.stringify(mergedInput),
|
|
4799
4805
|
jsonbArg(mergedGroundTruth),
|
|
4806
|
+
jsonbArg(mergedExpectedTrajectory),
|
|
4800
4807
|
jsonbArg(mergedRequestContext),
|
|
4801
4808
|
jsonbArg(mergedMetadata),
|
|
4802
4809
|
jsonbArg(mergedSource),
|
|
@@ -4816,6 +4823,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
|
|
|
4816
4823
|
datasetVersion: newVersion,
|
|
4817
4824
|
input: mergedInput,
|
|
4818
4825
|
groundTruth: mergedGroundTruth,
|
|
4826
|
+
expectedTrajectory: mergedExpectedTrajectory,
|
|
4819
4827
|
requestContext: mergedRequestContext,
|
|
4820
4828
|
metadata: mergedMetadata,
|
|
4821
4829
|
source: mergedSource,
|
|
@@ -4864,13 +4872,14 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
|
|
|
4864
4872
|
[newVersion, id]
|
|
4865
4873
|
);
|
|
4866
4874
|
await t.none(
|
|
4867
|
-
`INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,true,$4,$5,$6,$7,$8,$9,$10,$11,$12)`,
|
|
4875
|
+
`INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","expectedTrajectory","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,true,$4,$5,$6,$7,$8,$9,$10,$11,$12,$13)`,
|
|
4868
4876
|
[
|
|
4869
4877
|
id,
|
|
4870
4878
|
datasetId,
|
|
4871
4879
|
newVersion,
|
|
4872
4880
|
JSON.stringify(existing.input),
|
|
4873
4881
|
jsonbArg(existing.groundTruth),
|
|
4882
|
+
jsonbArg(existing.expectedTrajectory),
|
|
4874
4883
|
jsonbArg(existing.requestContext),
|
|
4875
4884
|
jsonbArg(existing.metadata),
|
|
4876
4885
|
jsonbArg(existing.source),
|
|
@@ -4927,13 +4936,14 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
|
|
|
4927
4936
|
newVersion = row.version;
|
|
4928
4937
|
for (const { id, input: itemInput } of itemsWithIds) {
|
|
4929
4938
|
await t.none(
|
|
4930
|
-
`INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,false,$4,$5,$6,$7,$8,$9,$10,$11,$12)`,
|
|
4939
|
+
`INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","expectedTrajectory","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,false,$4,$5,$6,$7,$8,$9,$10,$11,$12,$13)`,
|
|
4931
4940
|
[
|
|
4932
4941
|
id,
|
|
4933
4942
|
input.datasetId,
|
|
4934
4943
|
newVersion,
|
|
4935
4944
|
JSON.stringify(itemInput.input),
|
|
4936
4945
|
jsonbArg(itemInput.groundTruth),
|
|
4946
|
+
jsonbArg(itemInput.expectedTrajectory),
|
|
4937
4947
|
jsonbArg(itemInput.requestContext),
|
|
4938
4948
|
jsonbArg(itemInput.metadata),
|
|
4939
4949
|
jsonbArg(itemInput.source),
|
|
@@ -4955,6 +4965,7 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
|
|
|
4955
4965
|
datasetVersion: newVersion,
|
|
4956
4966
|
input: itemInput.input,
|
|
4957
4967
|
groundTruth: itemInput.groundTruth,
|
|
4968
|
+
expectedTrajectory: itemInput.expectedTrajectory,
|
|
4958
4969
|
requestContext: itemInput.requestContext,
|
|
4959
4970
|
metadata: itemInput.metadata,
|
|
4960
4971
|
source: itemInput.source,
|
|
@@ -5012,13 +5023,14 @@ var DatasetsPG = class _DatasetsPG extends DatasetsStorage {
|
|
|
5012
5023
|
[newVersion, item.id]
|
|
5013
5024
|
);
|
|
5014
5025
|
await t.none(
|
|
5015
|
-
`INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,true,$4,$5,$6,$7,$8,$9,$10,$11,$12)`,
|
|
5026
|
+
`INSERT INTO ${itemsTable} ("id","datasetId","datasetVersion","validTo","isDeleted","input","groundTruth","expectedTrajectory","requestContext","metadata","source","createdAt","createdAtZ","updatedAt","updatedAtZ") VALUES ($1,$2,$3,NULL,true,$4,$5,$6,$7,$8,$9,$10,$11,$12,$13)`,
|
|
5016
5027
|
[
|
|
5017
5028
|
item.id,
|
|
5018
5029
|
input.datasetId,
|
|
5019
5030
|
newVersion,
|
|
5020
5031
|
JSON.stringify(item.input),
|
|
5021
5032
|
jsonbArg(item.groundTruth),
|
|
5033
|
+
jsonbArg(item.expectedTrajectory),
|
|
5022
5034
|
jsonbArg(item.requestContext),
|
|
5023
5035
|
jsonbArg(item.metadata),
|
|
5024
5036
|
jsonbArg(item.source),
|
|
@@ -9034,7 +9046,8 @@ var MemoryPG = class _MemoryPG extends MemoryStorage {
|
|
|
9034
9046
|
lastObservedAt: input.chunk.lastObservedAt,
|
|
9035
9047
|
createdAt: /* @__PURE__ */ new Date(),
|
|
9036
9048
|
suggestedContinuation: input.chunk.suggestedContinuation,
|
|
9037
|
-
currentTask: input.chunk.currentTask
|
|
9049
|
+
currentTask: input.chunk.currentTask,
|
|
9050
|
+
threadTitle: input.chunk.threadTitle
|
|
9038
9051
|
};
|
|
9039
9052
|
const lastBufferedAtTime = input.lastBufferedAtTime ? input.lastBufferedAtTime.toISOString() : null;
|
|
9040
9053
|
const result = await this.#db.client.query(
|