macrocosmos 2.1.0 → 2.1.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2,17 +2,115 @@
2
2
  // Code generated by protoc-gen-ts_proto. DO NOT EDIT.
3
3
  // versions:
4
4
  // protoc-gen-ts_proto v2.7.0
5
- // protoc v6.33.0
5
+ // protoc v6.33.4
6
6
  // source: gravity/v1/gravity.proto
7
7
  Object.defineProperty(exports, "__esModule", { value: true });
8
- exports.AddPersistentGravityTaskRequest = exports.BuildAllDatasetsResponse = exports.BuildAllDatasetsRequest = exports.BuildDatasetResponse = exports.BuildDatasetRequest = exports.CreateGravityTaskResponse = exports.CreateGravityTaskRequest = exports.GetCrawlerResponse = exports.CompleteCrawlerRequest = exports.GetMarketplaceCrawlersResponse = exports.GetCrawlerRequest = exports.NotificationRequest = exports.GravityTask = exports.GetGravityTasksResponse = exports.GetGravityTasksRequest = exports.GravityMarketplaceTaskState = exports.GravityTaskState = exports.CrawlerState = exports.HfRepo = exports.CrawlerNotification = exports.PersistentTopicResponse = exports.PersistentTopic = exports.CrawlerCriteria = exports.InsertCrawlerCriteriaRequest = exports.GravityTaskRequest = exports.UpsertGravityTaskResponse = exports.UpsertGravityTaskRequest = exports.UpsertResponse = exports.UpsertCrawlerRequest = exports.GetPersistentGravityTasksResponse = exports.PersistentGravityTask = exports.Crawler = exports.GetPersistentDatasetWorkflowsResponse = exports.GetMarketplaceDatasetsRequest = exports.UpsertMarketplaceTaskMetadataRequest = exports.AddPersistentDatasetWorkflowsRequest = exports.PersistentDatasetWorkflow = exports.PublishDatasetRequest = exports.GetPopularTagsResponse = exports.PopularTag = exports.GetMarketplaceTaskSuggestionsResponse = exports.GetMarketplaceTaskSuggestionsRequest = exports.UpsertMarketplaceTaskSuggestionsRequest = exports.UpsertHotkeysRequest = exports.GetUserMarketplaceDatasetsResponse = exports.UserMarketplaceDataset = exports.BuyMarketplaceDatasetResponse = exports.BuyMarketplaceDatasetRequest = exports.GetHotkeysResponse = exports.protobufPackage = void 0;
9
- exports.GravityServiceClient = exports.GravityServiceService = exports.CrawlerDataForDD = exports.GetCrawlerDataForDDSubmissionResponse = exports.GetCrawlerDataForDDSubmissionRequest = exports.GetCrawlerHistoryResponse = exports.CrawlerCriteriaAndHistory = exports.CrawlerHistoryEntry = exports.GetCrawlerHistoryRequest = exports.GetGravityTaskDatasetFilesResponse = exports.DatasetFileWithId = exports.CrawlerRawMinerFilesResponse = exports.CrawlerDatasetFiles = exports.GetGravityTaskDatasetFilesRequest = exports.GetMarketplaceDatasetsResponse = exports.DatasetBillingCorrectionResponse = exports.DatasetBillingCorrectionRequest = exports.CancelDatasetResponse = exports.CancelDatasetRequest = exports.CancelGravityTaskResponse = exports.CancelGravityTaskRequest = exports.GetDatasetResponse = exports.GetDatasetRequest = exports.DatasetStep = exports.DatasetFile = exports.InsertDatasetFileRequest = exports.UpsertNebulaRequest = exports.UpsertDatasetRequest = exports.Dataset = exports.Nebula = void 0;
8
+ exports.DatasetFile = exports.InsertDatasetFileRequest = exports.UpsertNebulaRequest = exports.UpsertDatasetRequest = exports.Dataset = exports.Nebula = exports.ChargeForDatasetRowsRequest = exports.BuildAllDatasetsResponse = exports.BuildAllDatasetsRequest = exports.BuildDatasetResponse = exports.BuildDatasetRequest = exports.CreateGravityTaskResponse = exports.CreateGravityTaskRequest = exports.GetCrawlerResponse = exports.CompleteCrawlerRequest = exports.GetMarketplaceCrawlersResponse = exports.GetCrawlerRequest = exports.NotificationRequest = exports.GravityTask = exports.GetGravityTasksResponse = exports.GetGravityTasksRequest = exports.GravityMarketplaceTaskState = exports.GravityTaskState = exports.CrawlerState = exports.HfRepo = exports.CrawlerNotification = exports.CrawlerCriteria = exports.InsertCrawlerCriteriaRequest = exports.GravityTaskRequest = exports.UpsertGravityTaskResponse = exports.UpsertGravityTaskRequest = exports.UpsertResponse = exports.UpsertCrawlerRequest = exports.Crawler = exports.GetMarketplaceDatasetsRequest = exports.UpsertMarketplaceTaskMetadataRequest = exports.PublishDatasetRequest = exports.GetPopularTagsResponse = exports.PopularTag = exports.GetMarketplaceTaskSuggestionsResponse = exports.GetMarketplaceTaskSuggestionsRequest = exports.UpsertMarketplaceTaskSuggestionsRequest = exports.UpsertHotkeysRequest = exports.GetUserMarketplaceDatasetsResponse = exports.UserMarketplaceDataset = exports.BuyMarketplaceDatasetResponse = exports.BuyMarketplaceDatasetRequest = exports.GetHotkeysResponse = exports.UpsertRawMinerFilesRequest = exports.protobufPackage = void 0;
9
+ exports.GravityServiceClient = exports.GravityServiceService = exports.GetPreBuiltUserDatasetsResponse = exports.PreBuiltUserDataset = exports.GetPreBuiltUserDatasetsRequest = exports.UpsertPreBuiltUserDatasetsRequest = exports.ActiveUserTask = exports.ActiveUserCrawler = exports.GetActiveUserTasksResponse = exports.MarketplaceCrawlerDataForDDSubmission = exports.GetMarketplaceCrawlerDataForDDSubmissionResponse = exports.GetMarketplaceCrawlerDataForDDSubmissionRequest = exports.GetCrawlerHistoryResponse = exports.CrawlerCriteriaAndHistory = exports.CrawlerHistoryEntry = exports.GetCrawlerHistoryRequest = exports.GetGravityTaskDatasetFilesResponse = exports.DatasetFileWithId = exports.CrawlerRawMinerFilesResponse = exports.CrawlerDatasetFiles = exports.GetGravityTaskDatasetFilesRequest = exports.GetMarketplaceDatasetsResponse = exports.DatasetBillingCorrectionResponse = exports.DatasetBillingCorrectionRequest = exports.CancelDatasetResponse = exports.CancelDatasetRequest = exports.CancelGravityTaskResponse = exports.CancelGravityTaskRequest = exports.GetDatasetResponse = exports.GetDatasetRequest = exports.DatasetStep = void 0;
10
10
  /* eslint-disable */
11
11
  const wire_1 = require("@bufbuild/protobuf/wire");
12
12
  const grpc_js_1 = require("@grpc/grpc-js");
13
13
  const empty_1 = require("../../google/protobuf/empty");
14
14
  const timestamp_1 = require("../../google/protobuf/timestamp");
15
15
  exports.protobufPackage = "gravity.v1";
16
+ function createBaseUpsertRawMinerFilesRequest() {
17
+ return { crawlerId: "", parquetPaths: [], pathSizes: [] };
18
+ }
19
+ exports.UpsertRawMinerFilesRequest = {
20
+ encode(message, writer = new wire_1.BinaryWriter()) {
21
+ if (message.crawlerId !== "") {
22
+ writer.uint32(10).string(message.crawlerId);
23
+ }
24
+ for (const v of message.parquetPaths) {
25
+ writer.uint32(18).string(v);
26
+ }
27
+ writer.uint32(26).fork();
28
+ for (const v of message.pathSizes) {
29
+ writer.int64(v);
30
+ }
31
+ writer.join();
32
+ return writer;
33
+ },
34
+ decode(input, length) {
35
+ const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
36
+ let end = length === undefined ? reader.len : reader.pos + length;
37
+ const message = createBaseUpsertRawMinerFilesRequest();
38
+ while (reader.pos < end) {
39
+ const tag = reader.uint32();
40
+ switch (tag >>> 3) {
41
+ case 1: {
42
+ if (tag !== 10) {
43
+ break;
44
+ }
45
+ message.crawlerId = reader.string();
46
+ continue;
47
+ }
48
+ case 2: {
49
+ if (tag !== 18) {
50
+ break;
51
+ }
52
+ message.parquetPaths.push(reader.string());
53
+ continue;
54
+ }
55
+ case 3: {
56
+ if (tag === 24) {
57
+ message.pathSizes.push(longToNumber(reader.int64()));
58
+ continue;
59
+ }
60
+ if (tag === 26) {
61
+ const end2 = reader.uint32() + reader.pos;
62
+ while (reader.pos < end2) {
63
+ message.pathSizes.push(longToNumber(reader.int64()));
64
+ }
65
+ continue;
66
+ }
67
+ break;
68
+ }
69
+ }
70
+ if ((tag & 7) === 4 || tag === 0) {
71
+ break;
72
+ }
73
+ reader.skip(tag & 7);
74
+ }
75
+ return message;
76
+ },
77
+ fromJSON(object) {
78
+ return {
79
+ crawlerId: isSet(object.crawlerId)
80
+ ? globalThis.String(object.crawlerId)
81
+ : "",
82
+ parquetPaths: globalThis.Array.isArray(object?.parquetPaths)
83
+ ? object.parquetPaths.map((e) => globalThis.String(e))
84
+ : [],
85
+ pathSizes: globalThis.Array.isArray(object?.pathSizes)
86
+ ? object.pathSizes.map((e) => globalThis.Number(e))
87
+ : [],
88
+ };
89
+ },
90
+ toJSON(message) {
91
+ const obj = {};
92
+ if (message.crawlerId !== "") {
93
+ obj.crawlerId = message.crawlerId;
94
+ }
95
+ if (message.parquetPaths?.length) {
96
+ obj.parquetPaths = message.parquetPaths;
97
+ }
98
+ if (message.pathSizes?.length) {
99
+ obj.pathSizes = message.pathSizes.map(e => Math.round(e));
100
+ }
101
+ return obj;
102
+ },
103
+ create(base) {
104
+ return exports.UpsertRawMinerFilesRequest.fromPartial(base ?? {});
105
+ },
106
+ fromPartial(object) {
107
+ const message = createBaseUpsertRawMinerFilesRequest();
108
+ message.crawlerId = object.crawlerId ?? "";
109
+ message.parquetPaths = object.parquetPaths?.map(e => e) || [];
110
+ message.pathSizes = object.pathSizes?.map(e => e) || [];
111
+ return message;
112
+ },
113
+ };
16
114
  function createBaseGetHotkeysResponse() {
17
115
  return { hotkeys: [] };
18
116
  }
@@ -796,132 +894,6 @@ exports.PublishDatasetRequest = {
796
894
  return message;
797
895
  },
798
896
  };
799
- function createBasePersistentDatasetWorkflow() {
800
- return { datasetId: "", status: "" };
801
- }
802
- exports.PersistentDatasetWorkflow = {
803
- encode(message, writer = new wire_1.BinaryWriter()) {
804
- if (message.datasetId !== "") {
805
- writer.uint32(10).string(message.datasetId);
806
- }
807
- if (message.status !== "") {
808
- writer.uint32(18).string(message.status);
809
- }
810
- return writer;
811
- },
812
- decode(input, length) {
813
- const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
814
- let end = length === undefined ? reader.len : reader.pos + length;
815
- const message = createBasePersistentDatasetWorkflow();
816
- while (reader.pos < end) {
817
- const tag = reader.uint32();
818
- switch (tag >>> 3) {
819
- case 1: {
820
- if (tag !== 10) {
821
- break;
822
- }
823
- message.datasetId = reader.string();
824
- continue;
825
- }
826
- case 2: {
827
- if (tag !== 18) {
828
- break;
829
- }
830
- message.status = reader.string();
831
- continue;
832
- }
833
- }
834
- if ((tag & 7) === 4 || tag === 0) {
835
- break;
836
- }
837
- reader.skip(tag & 7);
838
- }
839
- return message;
840
- },
841
- fromJSON(object) {
842
- return {
843
- datasetId: isSet(object.datasetId)
844
- ? globalThis.String(object.datasetId)
845
- : "",
846
- status: isSet(object.status) ? globalThis.String(object.status) : "",
847
- };
848
- },
849
- toJSON(message) {
850
- const obj = {};
851
- if (message.datasetId !== "") {
852
- obj.datasetId = message.datasetId;
853
- }
854
- if (message.status !== "") {
855
- obj.status = message.status;
856
- }
857
- return obj;
858
- },
859
- create(base) {
860
- return exports.PersistentDatasetWorkflow.fromPartial(base ?? {});
861
- },
862
- fromPartial(object) {
863
- const message = createBasePersistentDatasetWorkflow();
864
- message.datasetId = object.datasetId ?? "";
865
- message.status = object.status ?? "";
866
- return message;
867
- },
868
- };
869
- function createBaseAddPersistentDatasetWorkflowsRequest() {
870
- return { datasetWorkflows: [] };
871
- }
872
- exports.AddPersistentDatasetWorkflowsRequest = {
873
- encode(message, writer = new wire_1.BinaryWriter()) {
874
- for (const v of message.datasetWorkflows) {
875
- exports.PersistentDatasetWorkflow.encode(v, writer.uint32(10).fork()).join();
876
- }
877
- return writer;
878
- },
879
- decode(input, length) {
880
- const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
881
- let end = length === undefined ? reader.len : reader.pos + length;
882
- const message = createBaseAddPersistentDatasetWorkflowsRequest();
883
- while (reader.pos < end) {
884
- const tag = reader.uint32();
885
- switch (tag >>> 3) {
886
- case 1: {
887
- if (tag !== 10) {
888
- break;
889
- }
890
- message.datasetWorkflows.push(exports.PersistentDatasetWorkflow.decode(reader, reader.uint32()));
891
- continue;
892
- }
893
- }
894
- if ((tag & 7) === 4 || tag === 0) {
895
- break;
896
- }
897
- reader.skip(tag & 7);
898
- }
899
- return message;
900
- },
901
- fromJSON(object) {
902
- return {
903
- datasetWorkflows: globalThis.Array.isArray(object?.datasetWorkflows)
904
- ? object.datasetWorkflows.map((e) => exports.PersistentDatasetWorkflow.fromJSON(e))
905
- : [],
906
- };
907
- },
908
- toJSON(message) {
909
- const obj = {};
910
- if (message.datasetWorkflows?.length) {
911
- obj.datasetWorkflows = message.datasetWorkflows.map(e => exports.PersistentDatasetWorkflow.toJSON(e));
912
- }
913
- return obj;
914
- },
915
- create(base) {
916
- return exports.AddPersistentDatasetWorkflowsRequest.fromPartial(base ?? {});
917
- },
918
- fromPartial(object) {
919
- const message = createBaseAddPersistentDatasetWorkflowsRequest();
920
- message.datasetWorkflows =
921
- object.datasetWorkflows?.map(e => exports.PersistentDatasetWorkflow.fromPartial(e)) || [];
922
- return message;
923
- },
924
- };
925
897
  function createBaseUpsertMarketplaceTaskMetadataRequest() {
926
898
  return {
927
899
  gravityTaskId: "",
@@ -1104,62 +1076,6 @@ exports.GetMarketplaceDatasetsRequest = {
1104
1076
  return message;
1105
1077
  },
1106
1078
  };
1107
- function createBaseGetPersistentDatasetWorkflowsResponse() {
1108
- return { datasetWorkflows: [] };
1109
- }
1110
- exports.GetPersistentDatasetWorkflowsResponse = {
1111
- encode(message, writer = new wire_1.BinaryWriter()) {
1112
- for (const v of message.datasetWorkflows) {
1113
- exports.PersistentDatasetWorkflow.encode(v, writer.uint32(10).fork()).join();
1114
- }
1115
- return writer;
1116
- },
1117
- decode(input, length) {
1118
- const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
1119
- let end = length === undefined ? reader.len : reader.pos + length;
1120
- const message = createBaseGetPersistentDatasetWorkflowsResponse();
1121
- while (reader.pos < end) {
1122
- const tag = reader.uint32();
1123
- switch (tag >>> 3) {
1124
- case 1: {
1125
- if (tag !== 10) {
1126
- break;
1127
- }
1128
- message.datasetWorkflows.push(exports.PersistentDatasetWorkflow.decode(reader, reader.uint32()));
1129
- continue;
1130
- }
1131
- }
1132
- if ((tag & 7) === 4 || tag === 0) {
1133
- break;
1134
- }
1135
- reader.skip(tag & 7);
1136
- }
1137
- return message;
1138
- },
1139
- fromJSON(object) {
1140
- return {
1141
- datasetWorkflows: globalThis.Array.isArray(object?.datasetWorkflows)
1142
- ? object.datasetWorkflows.map((e) => exports.PersistentDatasetWorkflow.fromJSON(e))
1143
- : [],
1144
- };
1145
- },
1146
- toJSON(message) {
1147
- const obj = {};
1148
- if (message.datasetWorkflows?.length) {
1149
- obj.datasetWorkflows = message.datasetWorkflows.map(e => exports.PersistentDatasetWorkflow.toJSON(e));
1150
- }
1151
- return obj;
1152
- },
1153
- create(base) {
1154
- return exports.GetPersistentDatasetWorkflowsResponse.fromPartial(base ?? {});
1155
- },
1156
- fromPartial(object) {
1157
- const message = createBaseGetPersistentDatasetWorkflowsResponse();
1158
- message.datasetWorkflows =
1159
- object.datasetWorkflows?.map(e => exports.PersistentDatasetWorkflow.fromPartial(e)) || [];
1160
- return message;
1161
- },
1162
- };
1163
1079
  function createBaseCrawler() {
1164
1080
  return {
1165
1081
  crawlerId: "",
@@ -1349,23 +1265,23 @@ exports.Crawler = {
1349
1265
  return message;
1350
1266
  },
1351
1267
  };
1352
- function createBasePersistentGravityTask() {
1353
- return { gravityTaskId: "", ingestDt: "" };
1268
+ function createBaseUpsertCrawlerRequest() {
1269
+ return { gravityTaskId: "", crawler: undefined };
1354
1270
  }
1355
- exports.PersistentGravityTask = {
1271
+ exports.UpsertCrawlerRequest = {
1356
1272
  encode(message, writer = new wire_1.BinaryWriter()) {
1357
1273
  if (message.gravityTaskId !== "") {
1358
1274
  writer.uint32(10).string(message.gravityTaskId);
1359
1275
  }
1360
- if (message.ingestDt !== "") {
1361
- writer.uint32(18).string(message.ingestDt);
1276
+ if (message.crawler !== undefined) {
1277
+ exports.Crawler.encode(message.crawler, writer.uint32(18).fork()).join();
1362
1278
  }
1363
1279
  return writer;
1364
1280
  },
1365
1281
  decode(input, length) {
1366
1282
  const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
1367
1283
  let end = length === undefined ? reader.len : reader.pos + length;
1368
- const message = createBasePersistentGravityTask();
1284
+ const message = createBaseUpsertCrawlerRequest();
1369
1285
  while (reader.pos < end) {
1370
1286
  const tag = reader.uint32();
1371
1287
  switch (tag >>> 3) {
@@ -1380,7 +1296,7 @@ exports.PersistentGravityTask = {
1380
1296
  if (tag !== 18) {
1381
1297
  break;
1382
1298
  }
1383
- message.ingestDt = reader.string();
1299
+ message.crawler = exports.Crawler.decode(reader, reader.uint32());
1384
1300
  continue;
1385
1301
  }
1386
1302
  }
@@ -1396,9 +1312,9 @@ exports.PersistentGravityTask = {
1396
1312
  gravityTaskId: isSet(object.gravityTaskId)
1397
1313
  ? globalThis.String(object.gravityTaskId)
1398
1314
  : "",
1399
- ingestDt: isSet(object.ingestDt)
1400
- ? globalThis.String(object.ingestDt)
1401
- : "",
1315
+ crawler: isSet(object.crawler)
1316
+ ? exports.Crawler.fromJSON(object.crawler)
1317
+ : undefined,
1402
1318
  };
1403
1319
  },
1404
1320
  toJSON(message) {
@@ -1406,35 +1322,38 @@ exports.PersistentGravityTask = {
1406
1322
  if (message.gravityTaskId !== "") {
1407
1323
  obj.gravityTaskId = message.gravityTaskId;
1408
1324
  }
1409
- if (message.ingestDt !== "") {
1410
- obj.ingestDt = message.ingestDt;
1325
+ if (message.crawler !== undefined) {
1326
+ obj.crawler = exports.Crawler.toJSON(message.crawler);
1411
1327
  }
1412
1328
  return obj;
1413
1329
  },
1414
1330
  create(base) {
1415
- return exports.PersistentGravityTask.fromPartial(base ?? {});
1331
+ return exports.UpsertCrawlerRequest.fromPartial(base ?? {});
1416
1332
  },
1417
1333
  fromPartial(object) {
1418
- const message = createBasePersistentGravityTask();
1334
+ const message = createBaseUpsertCrawlerRequest();
1419
1335
  message.gravityTaskId = object.gravityTaskId ?? "";
1420
- message.ingestDt = object.ingestDt ?? "";
1336
+ message.crawler =
1337
+ object.crawler !== undefined && object.crawler !== null
1338
+ ? exports.Crawler.fromPartial(object.crawler)
1339
+ : undefined;
1421
1340
  return message;
1422
1341
  },
1423
1342
  };
1424
- function createBaseGetPersistentGravityTasksResponse() {
1425
- return { persistentGravityTasks: [] };
1343
+ function createBaseUpsertResponse() {
1344
+ return { message: "" };
1426
1345
  }
1427
- exports.GetPersistentGravityTasksResponse = {
1346
+ exports.UpsertResponse = {
1428
1347
  encode(message, writer = new wire_1.BinaryWriter()) {
1429
- for (const v of message.persistentGravityTasks) {
1430
- exports.PersistentGravityTask.encode(v, writer.uint32(10).fork()).join();
1348
+ if (message.message !== "") {
1349
+ writer.uint32(10).string(message.message);
1431
1350
  }
1432
1351
  return writer;
1433
1352
  },
1434
1353
  decode(input, length) {
1435
1354
  const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
1436
1355
  let end = length === undefined ? reader.len : reader.pos + length;
1437
- const message = createBaseGetPersistentGravityTasksResponse();
1356
+ const message = createBaseUpsertResponse();
1438
1357
  while (reader.pos < end) {
1439
1358
  const tag = reader.uint32();
1440
1359
  switch (tag >>> 3) {
@@ -1442,7 +1361,7 @@ exports.GetPersistentGravityTasksResponse = {
1442
1361
  if (tag !== 10) {
1443
1362
  break;
1444
1363
  }
1445
- message.persistentGravityTasks.push(exports.PersistentGravityTask.decode(reader, reader.uint32()));
1364
+ message.message = reader.string();
1446
1365
  continue;
1447
1366
  }
1448
1367
  }
@@ -1455,149 +1374,18 @@ exports.GetPersistentGravityTasksResponse = {
1455
1374
  },
1456
1375
  fromJSON(object) {
1457
1376
  return {
1458
- persistentGravityTasks: globalThis.Array.isArray(object?.persistentGravityTasks)
1459
- ? object.persistentGravityTasks.map((e) => exports.PersistentGravityTask.fromJSON(e))
1460
- : [],
1377
+ message: isSet(object.message) ? globalThis.String(object.message) : "",
1461
1378
  };
1462
1379
  },
1463
1380
  toJSON(message) {
1464
1381
  const obj = {};
1465
- if (message.persistentGravityTasks?.length) {
1466
- obj.persistentGravityTasks = message.persistentGravityTasks.map(e => exports.PersistentGravityTask.toJSON(e));
1382
+ if (message.message !== "") {
1383
+ obj.message = message.message;
1467
1384
  }
1468
1385
  return obj;
1469
1386
  },
1470
1387
  create(base) {
1471
- return exports.GetPersistentGravityTasksResponse.fromPartial(base ?? {});
1472
- },
1473
- fromPartial(object) {
1474
- const message = createBaseGetPersistentGravityTasksResponse();
1475
- message.persistentGravityTasks =
1476
- object.persistentGravityTasks?.map(e => exports.PersistentGravityTask.fromPartial(e)) || [];
1477
- return message;
1478
- },
1479
- };
1480
- function createBaseUpsertCrawlerRequest() {
1481
- return { gravityTaskId: "", crawler: undefined };
1482
- }
1483
- exports.UpsertCrawlerRequest = {
1484
- encode(message, writer = new wire_1.BinaryWriter()) {
1485
- if (message.gravityTaskId !== "") {
1486
- writer.uint32(10).string(message.gravityTaskId);
1487
- }
1488
- if (message.crawler !== undefined) {
1489
- exports.Crawler.encode(message.crawler, writer.uint32(18).fork()).join();
1490
- }
1491
- return writer;
1492
- },
1493
- decode(input, length) {
1494
- const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
1495
- let end = length === undefined ? reader.len : reader.pos + length;
1496
- const message = createBaseUpsertCrawlerRequest();
1497
- while (reader.pos < end) {
1498
- const tag = reader.uint32();
1499
- switch (tag >>> 3) {
1500
- case 1: {
1501
- if (tag !== 10) {
1502
- break;
1503
- }
1504
- message.gravityTaskId = reader.string();
1505
- continue;
1506
- }
1507
- case 2: {
1508
- if (tag !== 18) {
1509
- break;
1510
- }
1511
- message.crawler = exports.Crawler.decode(reader, reader.uint32());
1512
- continue;
1513
- }
1514
- }
1515
- if ((tag & 7) === 4 || tag === 0) {
1516
- break;
1517
- }
1518
- reader.skip(tag & 7);
1519
- }
1520
- return message;
1521
- },
1522
- fromJSON(object) {
1523
- return {
1524
- gravityTaskId: isSet(object.gravityTaskId)
1525
- ? globalThis.String(object.gravityTaskId)
1526
- : "",
1527
- crawler: isSet(object.crawler)
1528
- ? exports.Crawler.fromJSON(object.crawler)
1529
- : undefined,
1530
- };
1531
- },
1532
- toJSON(message) {
1533
- const obj = {};
1534
- if (message.gravityTaskId !== "") {
1535
- obj.gravityTaskId = message.gravityTaskId;
1536
- }
1537
- if (message.crawler !== undefined) {
1538
- obj.crawler = exports.Crawler.toJSON(message.crawler);
1539
- }
1540
- return obj;
1541
- },
1542
- create(base) {
1543
- return exports.UpsertCrawlerRequest.fromPartial(base ?? {});
1544
- },
1545
- fromPartial(object) {
1546
- const message = createBaseUpsertCrawlerRequest();
1547
- message.gravityTaskId = object.gravityTaskId ?? "";
1548
- message.crawler =
1549
- object.crawler !== undefined && object.crawler !== null
1550
- ? exports.Crawler.fromPartial(object.crawler)
1551
- : undefined;
1552
- return message;
1553
- },
1554
- };
1555
- function createBaseUpsertResponse() {
1556
- return { message: "" };
1557
- }
1558
- exports.UpsertResponse = {
1559
- encode(message, writer = new wire_1.BinaryWriter()) {
1560
- if (message.message !== "") {
1561
- writer.uint32(10).string(message.message);
1562
- }
1563
- return writer;
1564
- },
1565
- decode(input, length) {
1566
- const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
1567
- let end = length === undefined ? reader.len : reader.pos + length;
1568
- const message = createBaseUpsertResponse();
1569
- while (reader.pos < end) {
1570
- const tag = reader.uint32();
1571
- switch (tag >>> 3) {
1572
- case 1: {
1573
- if (tag !== 10) {
1574
- break;
1575
- }
1576
- message.message = reader.string();
1577
- continue;
1578
- }
1579
- }
1580
- if ((tag & 7) === 4 || tag === 0) {
1581
- break;
1582
- }
1583
- reader.skip(tag & 7);
1584
- }
1585
- return message;
1586
- },
1587
- fromJSON(object) {
1588
- return {
1589
- message: isSet(object.message) ? globalThis.String(object.message) : "",
1590
- };
1591
- },
1592
- toJSON(message) {
1593
- const obj = {};
1594
- if (message.message !== "") {
1595
- obj.message = message.message;
1596
- }
1597
- return obj;
1598
- },
1599
- create(base) {
1600
- return exports.UpsertResponse.fromPartial(base ?? {});
1388
+ return exports.UpsertResponse.fromPartial(base ?? {});
1601
1389
  },
1602
1390
  fromPartial(object) {
1603
1391
  const message = createBaseUpsertResponse();
@@ -2112,132 +1900,6 @@ exports.CrawlerCriteria = {
2112
1900
  return message;
2113
1901
  },
2114
1902
  };
2115
- function createBasePersistentTopic() {
2116
- return { platform: "", topic: undefined };
2117
- }
2118
- exports.PersistentTopic = {
2119
- encode(message, writer = new wire_1.BinaryWriter()) {
2120
- if (message.platform !== "") {
2121
- writer.uint32(10).string(message.platform);
2122
- }
2123
- if (message.topic !== undefined) {
2124
- writer.uint32(18).string(message.topic);
2125
- }
2126
- return writer;
2127
- },
2128
- decode(input, length) {
2129
- const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
2130
- let end = length === undefined ? reader.len : reader.pos + length;
2131
- const message = createBasePersistentTopic();
2132
- while (reader.pos < end) {
2133
- const tag = reader.uint32();
2134
- switch (tag >>> 3) {
2135
- case 1: {
2136
- if (tag !== 10) {
2137
- break;
2138
- }
2139
- message.platform = reader.string();
2140
- continue;
2141
- }
2142
- case 2: {
2143
- if (tag !== 18) {
2144
- break;
2145
- }
2146
- message.topic = reader.string();
2147
- continue;
2148
- }
2149
- }
2150
- if ((tag & 7) === 4 || tag === 0) {
2151
- break;
2152
- }
2153
- reader.skip(tag & 7);
2154
- }
2155
- return message;
2156
- },
2157
- fromJSON(object) {
2158
- return {
2159
- platform: isSet(object.platform)
2160
- ? globalThis.String(object.platform)
2161
- : "",
2162
- topic: isSet(object.topic) ? globalThis.String(object.topic) : undefined,
2163
- };
2164
- },
2165
- toJSON(message) {
2166
- const obj = {};
2167
- if (message.platform !== "") {
2168
- obj.platform = message.platform;
2169
- }
2170
- if (message.topic !== undefined) {
2171
- obj.topic = message.topic;
2172
- }
2173
- return obj;
2174
- },
2175
- create(base) {
2176
- return exports.PersistentTopic.fromPartial(base ?? {});
2177
- },
2178
- fromPartial(object) {
2179
- const message = createBasePersistentTopic();
2180
- message.platform = object.platform ?? "";
2181
- message.topic = object.topic ?? undefined;
2182
- return message;
2183
- },
2184
- };
2185
- function createBasePersistentTopicResponse() {
2186
- return { persistentTopics: [] };
2187
- }
2188
- exports.PersistentTopicResponse = {
2189
- encode(message, writer = new wire_1.BinaryWriter()) {
2190
- for (const v of message.persistentTopics) {
2191
- exports.PersistentTopic.encode(v, writer.uint32(10).fork()).join();
2192
- }
2193
- return writer;
2194
- },
2195
- decode(input, length) {
2196
- const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
2197
- let end = length === undefined ? reader.len : reader.pos + length;
2198
- const message = createBasePersistentTopicResponse();
2199
- while (reader.pos < end) {
2200
- const tag = reader.uint32();
2201
- switch (tag >>> 3) {
2202
- case 1: {
2203
- if (tag !== 10) {
2204
- break;
2205
- }
2206
- message.persistentTopics.push(exports.PersistentTopic.decode(reader, reader.uint32()));
2207
- continue;
2208
- }
2209
- }
2210
- if ((tag & 7) === 4 || tag === 0) {
2211
- break;
2212
- }
2213
- reader.skip(tag & 7);
2214
- }
2215
- return message;
2216
- },
2217
- fromJSON(object) {
2218
- return {
2219
- persistentTopics: globalThis.Array.isArray(object?.persistentTopics)
2220
- ? object.persistentTopics.map((e) => exports.PersistentTopic.fromJSON(e))
2221
- : [],
2222
- };
2223
- },
2224
- toJSON(message) {
2225
- const obj = {};
2226
- if (message.persistentTopics?.length) {
2227
- obj.persistentTopics = message.persistentTopics.map(e => exports.PersistentTopic.toJSON(e));
2228
- }
2229
- return obj;
2230
- },
2231
- create(base) {
2232
- return exports.PersistentTopicResponse.fromPartial(base ?? {});
2233
- },
2234
- fromPartial(object) {
2235
- const message = createBasePersistentTopicResponse();
2236
- message.persistentTopics =
2237
- object.persistentTopics?.map(e => exports.PersistentTopic.fromPartial(e)) || [];
2238
- return message;
2239
- },
2240
- };
2241
1903
  function createBaseCrawlerNotification() {
2242
1904
  return { to: "", link: "" };
2243
1905
  }
@@ -3660,7 +3322,7 @@ function createBaseBuildDatasetRequest() {
3660
3322
  crawlerId: "",
3661
3323
  notificationRequests: [],
3662
3324
  maxRows: 0,
3663
- isMarketplace: undefined,
3325
+ isPeriodic: undefined,
3664
3326
  };
3665
3327
  }
3666
3328
  exports.BuildDatasetRequest = {
@@ -3674,8 +3336,8 @@ exports.BuildDatasetRequest = {
3674
3336
  if (message.maxRows !== 0) {
3675
3337
  writer.uint32(24).int64(message.maxRows);
3676
3338
  }
3677
- if (message.isMarketplace !== undefined) {
3678
- writer.uint32(32).bool(message.isMarketplace);
3339
+ if (message.isPeriodic !== undefined) {
3340
+ writer.uint32(32).bool(message.isPeriodic);
3679
3341
  }
3680
3342
  return writer;
3681
3343
  },
@@ -3711,7 +3373,7 @@ exports.BuildDatasetRequest = {
3711
3373
  if (tag !== 32) {
3712
3374
  break;
3713
3375
  }
3714
- message.isMarketplace = reader.bool();
3376
+ message.isPeriodic = reader.bool();
3715
3377
  continue;
3716
3378
  }
3717
3379
  }
@@ -3731,8 +3393,8 @@ exports.BuildDatasetRequest = {
3731
3393
  ? object.notificationRequests.map((e) => exports.NotificationRequest.fromJSON(e))
3732
3394
  : [],
3733
3395
  maxRows: isSet(object.maxRows) ? globalThis.Number(object.maxRows) : 0,
3734
- isMarketplace: isSet(object.isMarketplace)
3735
- ? globalThis.Boolean(object.isMarketplace)
3396
+ isPeriodic: isSet(object.isPeriodic)
3397
+ ? globalThis.Boolean(object.isPeriodic)
3736
3398
  : undefined,
3737
3399
  };
3738
3400
  },
@@ -3747,8 +3409,8 @@ exports.BuildDatasetRequest = {
3747
3409
  if (message.maxRows !== 0) {
3748
3410
  obj.maxRows = Math.round(message.maxRows);
3749
3411
  }
3750
- if (message.isMarketplace !== undefined) {
3751
- obj.isMarketplace = message.isMarketplace;
3412
+ if (message.isPeriodic !== undefined) {
3413
+ obj.isPeriodic = message.isPeriodic;
3752
3414
  }
3753
3415
  return obj;
3754
3416
  },
@@ -3761,7 +3423,7 @@ exports.BuildDatasetRequest = {
3761
3423
  message.notificationRequests =
3762
3424
  object.notificationRequests?.map(e => exports.NotificationRequest.fromPartial(e)) || [];
3763
3425
  message.maxRows = object.maxRows ?? 0;
3764
- message.isMarketplace = object.isMarketplace ?? undefined;
3426
+ message.isPeriodic = object.isPeriodic ?? undefined;
3765
3427
  return message;
3766
3428
  },
3767
3429
  };
@@ -3841,11 +3503,7 @@ exports.BuildDatasetResponse = {
3841
3503
  },
3842
3504
  };
3843
3505
  function createBaseBuildAllDatasetsRequest() {
3844
- return {
3845
- gravityTaskId: "",
3846
- buildCrawlersConfig: [],
3847
- isMarketplace: undefined,
3848
- };
3506
+ return { gravityTaskId: "", buildCrawlersConfig: [] };
3849
3507
  }
3850
3508
  exports.BuildAllDatasetsRequest = {
3851
3509
  encode(message, writer = new wire_1.BinaryWriter()) {
@@ -3855,9 +3513,6 @@ exports.BuildAllDatasetsRequest = {
3855
3513
  for (const v of message.buildCrawlersConfig) {
3856
3514
  exports.BuildDatasetRequest.encode(v, writer.uint32(18).fork()).join();
3857
3515
  }
3858
- if (message.isMarketplace !== undefined) {
3859
- writer.uint32(24).bool(message.isMarketplace);
3860
- }
3861
3516
  return writer;
3862
3517
  },
3863
3518
  decode(input, length) {
@@ -3881,13 +3536,6 @@ exports.BuildAllDatasetsRequest = {
3881
3536
  message.buildCrawlersConfig.push(exports.BuildDatasetRequest.decode(reader, reader.uint32()));
3882
3537
  continue;
3883
3538
  }
3884
- case 3: {
3885
- if (tag !== 24) {
3886
- break;
3887
- }
3888
- message.isMarketplace = reader.bool();
3889
- continue;
3890
- }
3891
3539
  }
3892
3540
  if ((tag & 7) === 4 || tag === 0) {
3893
3541
  break;
@@ -3904,9 +3552,6 @@ exports.BuildAllDatasetsRequest = {
3904
3552
  buildCrawlersConfig: globalThis.Array.isArray(object?.buildCrawlersConfig)
3905
3553
  ? object.buildCrawlersConfig.map((e) => exports.BuildDatasetRequest.fromJSON(e))
3906
3554
  : [],
3907
- isMarketplace: isSet(object.isMarketplace)
3908
- ? globalThis.Boolean(object.isMarketplace)
3909
- : undefined,
3910
3555
  };
3911
3556
  },
3912
3557
  toJSON(message) {
@@ -3917,9 +3562,6 @@ exports.BuildAllDatasetsRequest = {
3917
3562
  if (message.buildCrawlersConfig?.length) {
3918
3563
  obj.buildCrawlersConfig = message.buildCrawlersConfig.map(e => exports.BuildDatasetRequest.toJSON(e));
3919
3564
  }
3920
- if (message.isMarketplace !== undefined) {
3921
- obj.isMarketplace = message.isMarketplace;
3922
- }
3923
3565
  return obj;
3924
3566
  },
3925
3567
  create(base) {
@@ -3930,7 +3572,6 @@ exports.BuildAllDatasetsRequest = {
3930
3572
  message.gravityTaskId = object.gravityTaskId ?? "";
3931
3573
  message.buildCrawlersConfig =
3932
3574
  object.buildCrawlersConfig?.map(e => exports.BuildDatasetRequest.fromPartial(e)) || [];
3933
- message.isMarketplace = object.isMarketplace ?? undefined;
3934
3575
  return message;
3935
3576
  },
3936
3577
  };
@@ -4006,20 +3647,23 @@ exports.BuildAllDatasetsResponse = {
4006
3647
  return message;
4007
3648
  },
4008
3649
  };
4009
- function createBaseAddPersistentGravityTaskRequest() {
4010
- return { gravityTaskId: "" };
3650
+ function createBaseChargeForDatasetRowsRequest() {
3651
+ return { crawlerId: "", rowCount: 0 };
4011
3652
  }
4012
- exports.AddPersistentGravityTaskRequest = {
3653
+ exports.ChargeForDatasetRowsRequest = {
4013
3654
  encode(message, writer = new wire_1.BinaryWriter()) {
4014
- if (message.gravityTaskId !== "") {
4015
- writer.uint32(10).string(message.gravityTaskId);
3655
+ if (message.crawlerId !== "") {
3656
+ writer.uint32(10).string(message.crawlerId);
3657
+ }
3658
+ if (message.rowCount !== 0) {
3659
+ writer.uint32(16).int64(message.rowCount);
4016
3660
  }
4017
3661
  return writer;
4018
3662
  },
4019
3663
  decode(input, length) {
4020
3664
  const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
4021
3665
  let end = length === undefined ? reader.len : reader.pos + length;
4022
- const message = createBaseAddPersistentGravityTaskRequest();
3666
+ const message = createBaseChargeForDatasetRowsRequest();
4023
3667
  while (reader.pos < end) {
4024
3668
  const tag = reader.uint32();
4025
3669
  switch (tag >>> 3) {
@@ -4027,37 +3671,51 @@ exports.AddPersistentGravityTaskRequest = {
4027
3671
  if (tag !== 10) {
4028
3672
  break;
4029
3673
  }
4030
- message.gravityTaskId = reader.string();
3674
+ message.crawlerId = reader.string();
4031
3675
  continue;
4032
3676
  }
4033
- }
4034
- if ((tag & 7) === 4 || tag === 0) {
4035
- break;
4036
- }
3677
+ case 2: {
3678
+ if (tag !== 16) {
3679
+ break;
3680
+ }
3681
+ message.rowCount = longToNumber(reader.int64());
3682
+ continue;
3683
+ }
3684
+ }
3685
+ if ((tag & 7) === 4 || tag === 0) {
3686
+ break;
3687
+ }
4037
3688
  reader.skip(tag & 7);
4038
3689
  }
4039
3690
  return message;
4040
3691
  },
4041
3692
  fromJSON(object) {
4042
3693
  return {
4043
- gravityTaskId: isSet(object.gravityTaskId)
4044
- ? globalThis.String(object.gravityTaskId)
3694
+ crawlerId: isSet(object.crawlerId)
3695
+ ? globalThis.String(object.crawlerId)
4045
3696
  : "",
3697
+ rowCount: isSet(object.rowCount)
3698
+ ? globalThis.Number(object.rowCount)
3699
+ : 0,
4046
3700
  };
4047
3701
  },
4048
3702
  toJSON(message) {
4049
3703
  const obj = {};
4050
- if (message.gravityTaskId !== "") {
4051
- obj.gravityTaskId = message.gravityTaskId;
3704
+ if (message.crawlerId !== "") {
3705
+ obj.crawlerId = message.crawlerId;
3706
+ }
3707
+ if (message.rowCount !== 0) {
3708
+ obj.rowCount = Math.round(message.rowCount);
4052
3709
  }
4053
3710
  return obj;
4054
3711
  },
4055
3712
  create(base) {
4056
- return exports.AddPersistentGravityTaskRequest.fromPartial(base ?? {});
3713
+ return exports.ChargeForDatasetRowsRequest.fromPartial(base ?? {});
4057
3714
  },
4058
3715
  fromPartial(object) {
4059
- const message = createBaseAddPersistentGravityTaskRequest();
4060
- message.gravityTaskId = object.gravityTaskId ?? "";
3716
+ const message = createBaseChargeForDatasetRowsRequest();
3717
+ message.crawlerId = object.crawlerId ?? "";
3718
+ message.rowCount = object.rowCount ?? 0;
4061
3719
  return message;
4062
3720
  },
4063
3721
  };
@@ -5451,7 +5109,7 @@ exports.CrawlerDatasetFiles = {
5451
5109
  },
5452
5110
  };
5453
5111
  function createBaseCrawlerRawMinerFilesResponse() {
5454
- return { crawlerId: "", s3Paths: [] };
5112
+ return { crawlerId: "", s3Paths: [], fileSizeBytes: [] };
5455
5113
  }
5456
5114
  exports.CrawlerRawMinerFilesResponse = {
5457
5115
  encode(message, writer = new wire_1.BinaryWriter()) {
@@ -5461,6 +5119,11 @@ exports.CrawlerRawMinerFilesResponse = {
5461
5119
  for (const v of message.s3Paths) {
5462
5120
  writer.uint32(18).string(v);
5463
5121
  }
5122
+ writer.uint32(26).fork();
5123
+ for (const v of message.fileSizeBytes) {
5124
+ writer.int64(v);
5125
+ }
5126
+ writer.join();
5464
5127
  return writer;
5465
5128
  },
5466
5129
  decode(input, length) {
@@ -5484,6 +5147,20 @@ exports.CrawlerRawMinerFilesResponse = {
5484
5147
  message.s3Paths.push(reader.string());
5485
5148
  continue;
5486
5149
  }
5150
+ case 3: {
5151
+ if (tag === 24) {
5152
+ message.fileSizeBytes.push(longToNumber(reader.int64()));
5153
+ continue;
5154
+ }
5155
+ if (tag === 26) {
5156
+ const end2 = reader.uint32() + reader.pos;
5157
+ while (reader.pos < end2) {
5158
+ message.fileSizeBytes.push(longToNumber(reader.int64()));
5159
+ }
5160
+ continue;
5161
+ }
5162
+ break;
5163
+ }
5487
5164
  }
5488
5165
  if ((tag & 7) === 4 || tag === 0) {
5489
5166
  break;
@@ -5500,6 +5177,9 @@ exports.CrawlerRawMinerFilesResponse = {
5500
5177
  s3Paths: globalThis.Array.isArray(object?.s3Paths)
5501
5178
  ? object.s3Paths.map((e) => globalThis.String(e))
5502
5179
  : [],
5180
+ fileSizeBytes: globalThis.Array.isArray(object?.fileSizeBytes)
5181
+ ? object.fileSizeBytes.map((e) => globalThis.Number(e))
5182
+ : [],
5503
5183
  };
5504
5184
  },
5505
5185
  toJSON(message) {
@@ -5510,6 +5190,9 @@ exports.CrawlerRawMinerFilesResponse = {
5510
5190
  if (message.s3Paths?.length) {
5511
5191
  obj.s3Paths = message.s3Paths;
5512
5192
  }
5193
+ if (message.fileSizeBytes?.length) {
5194
+ obj.fileSizeBytes = message.fileSizeBytes.map(e => Math.round(e));
5195
+ }
5513
5196
  return obj;
5514
5197
  },
5515
5198
  create(base) {
@@ -5519,6 +5202,7 @@ exports.CrawlerRawMinerFilesResponse = {
5519
5202
  const message = createBaseCrawlerRawMinerFilesResponse();
5520
5203
  message.crawlerId = object.crawlerId ?? "";
5521
5204
  message.s3Paths = object.s3Paths?.map(e => e) || [];
5205
+ message.fileSizeBytes = object.fileSizeBytes?.map(e => e) || [];
5522
5206
  return message;
5523
5207
  },
5524
5208
  };
@@ -6157,20 +5841,20 @@ exports.GetCrawlerHistoryResponse = {
6157
5841
  return message;
6158
5842
  },
6159
5843
  };
6160
- function createBaseGetCrawlerDataForDDSubmissionRequest() {
6161
- return { dsns: [] };
5844
+ function createBaseGetMarketplaceCrawlerDataForDDSubmissionRequest() {
5845
+ return { marketplaceUserId: "" };
6162
5846
  }
6163
- exports.GetCrawlerDataForDDSubmissionRequest = {
5847
+ exports.GetMarketplaceCrawlerDataForDDSubmissionRequest = {
6164
5848
  encode(message, writer = new wire_1.BinaryWriter()) {
6165
- for (const v of message.dsns) {
6166
- writer.uint32(10).string(v);
5849
+ if (message.marketplaceUserId !== "") {
5850
+ writer.uint32(10).string(message.marketplaceUserId);
6167
5851
  }
6168
5852
  return writer;
6169
5853
  },
6170
5854
  decode(input, length) {
6171
5855
  const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
6172
5856
  let end = length === undefined ? reader.len : reader.pos + length;
6173
- const message = createBaseGetCrawlerDataForDDSubmissionRequest();
5857
+ const message = createBaseGetMarketplaceCrawlerDataForDDSubmissionRequest();
6174
5858
  while (reader.pos < end) {
6175
5859
  const tag = reader.uint32();
6176
5860
  switch (tag >>> 3) {
@@ -6178,7 +5862,7 @@ exports.GetCrawlerDataForDDSubmissionRequest = {
6178
5862
  if (tag !== 10) {
6179
5863
  break;
6180
5864
  }
6181
- message.dsns.push(reader.string());
5865
+ message.marketplaceUserId = reader.string();
6182
5866
  continue;
6183
5867
  }
6184
5868
  }
@@ -6191,41 +5875,41 @@ exports.GetCrawlerDataForDDSubmissionRequest = {
6191
5875
  },
6192
5876
  fromJSON(object) {
6193
5877
  return {
6194
- dsns: globalThis.Array.isArray(object?.dsns)
6195
- ? object.dsns.map((e) => globalThis.String(e))
6196
- : [],
5878
+ marketplaceUserId: isSet(object.marketplaceUserId)
5879
+ ? globalThis.String(object.marketplaceUserId)
5880
+ : "",
6197
5881
  };
6198
5882
  },
6199
5883
  toJSON(message) {
6200
5884
  const obj = {};
6201
- if (message.dsns?.length) {
6202
- obj.dsns = message.dsns;
5885
+ if (message.marketplaceUserId !== "") {
5886
+ obj.marketplaceUserId = message.marketplaceUserId;
6203
5887
  }
6204
5888
  return obj;
6205
5889
  },
6206
5890
  create(base) {
6207
- return exports.GetCrawlerDataForDDSubmissionRequest.fromPartial(base ?? {});
5891
+ return exports.GetMarketplaceCrawlerDataForDDSubmissionRequest.fromPartial(base ?? {});
6208
5892
  },
6209
5893
  fromPartial(object) {
6210
- const message = createBaseGetCrawlerDataForDDSubmissionRequest();
6211
- message.dsns = object.dsns?.map(e => e) || [];
5894
+ const message = createBaseGetMarketplaceCrawlerDataForDDSubmissionRequest();
5895
+ message.marketplaceUserId = object.marketplaceUserId ?? "";
6212
5896
  return message;
6213
5897
  },
6214
5898
  };
6215
- function createBaseGetCrawlerDataForDDSubmissionResponse() {
5899
+ function createBaseGetMarketplaceCrawlerDataForDDSubmissionResponse() {
6216
5900
  return { crawlers: [] };
6217
5901
  }
6218
- exports.GetCrawlerDataForDDSubmissionResponse = {
5902
+ exports.GetMarketplaceCrawlerDataForDDSubmissionResponse = {
6219
5903
  encode(message, writer = new wire_1.BinaryWriter()) {
6220
5904
  for (const v of message.crawlers) {
6221
- exports.CrawlerDataForDD.encode(v, writer.uint32(10).fork()).join();
5905
+ exports.MarketplaceCrawlerDataForDDSubmission.encode(v, writer.uint32(10).fork()).join();
6222
5906
  }
6223
5907
  return writer;
6224
5908
  },
6225
5909
  decode(input, length) {
6226
5910
  const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
6227
5911
  let end = length === undefined ? reader.len : reader.pos + length;
6228
- const message = createBaseGetCrawlerDataForDDSubmissionResponse();
5912
+ const message = createBaseGetMarketplaceCrawlerDataForDDSubmissionResponse();
6229
5913
  while (reader.pos < end) {
6230
5914
  const tag = reader.uint32();
6231
5915
  switch (tag >>> 3) {
@@ -6233,7 +5917,7 @@ exports.GetCrawlerDataForDDSubmissionResponse = {
6233
5917
  if (tag !== 10) {
6234
5918
  break;
6235
5919
  }
6236
- message.crawlers.push(exports.CrawlerDataForDD.decode(reader, reader.uint32()));
5920
+ message.crawlers.push(exports.MarketplaceCrawlerDataForDDSubmission.decode(reader, reader.uint32()));
6237
5921
  continue;
6238
5922
  }
6239
5923
  }
@@ -6247,28 +5931,28 @@ exports.GetCrawlerDataForDDSubmissionResponse = {
6247
5931
  fromJSON(object) {
6248
5932
  return {
6249
5933
  crawlers: globalThis.Array.isArray(object?.crawlers)
6250
- ? object.crawlers.map((e) => exports.CrawlerDataForDD.fromJSON(e))
5934
+ ? object.crawlers.map((e) => exports.MarketplaceCrawlerDataForDDSubmission.fromJSON(e))
6251
5935
  : [],
6252
5936
  };
6253
5937
  },
6254
5938
  toJSON(message) {
6255
5939
  const obj = {};
6256
5940
  if (message.crawlers?.length) {
6257
- obj.crawlers = message.crawlers.map(e => exports.CrawlerDataForDD.toJSON(e));
5941
+ obj.crawlers = message.crawlers.map(e => exports.MarketplaceCrawlerDataForDDSubmission.toJSON(e));
6258
5942
  }
6259
5943
  return obj;
6260
5944
  },
6261
5945
  create(base) {
6262
- return exports.GetCrawlerDataForDDSubmissionResponse.fromPartial(base ?? {});
5946
+ return exports.GetMarketplaceCrawlerDataForDDSubmissionResponse.fromPartial(base ?? {});
6263
5947
  },
6264
5948
  fromPartial(object) {
6265
- const message = createBaseGetCrawlerDataForDDSubmissionResponse();
5949
+ const message = createBaseGetMarketplaceCrawlerDataForDDSubmissionResponse();
6266
5950
  message.crawlers =
6267
- object.crawlers?.map(e => exports.CrawlerDataForDD.fromPartial(e)) || [];
5951
+ object.crawlers?.map(e => exports.MarketplaceCrawlerDataForDDSubmission.fromPartial(e)) || [];
6268
5952
  return message;
6269
5953
  },
6270
5954
  };
6271
- function createBaseCrawlerDataForDD() {
5955
+ function createBaseMarketplaceCrawlerDataForDDSubmission() {
6272
5956
  return {
6273
5957
  crawlerId: "",
6274
5958
  platform: "",
@@ -6276,9 +5960,18 @@ function createBaseCrawlerDataForDD() {
6276
5960
  keyword: undefined,
6277
5961
  postStartDatetime: undefined,
6278
5962
  postEndDatetime: undefined,
5963
+ startTime: undefined,
5964
+ deregistrationTime: undefined,
5965
+ archiveTime: undefined,
5966
+ status: "",
5967
+ bytesCollected: 0,
5968
+ recordsCollected: 0,
5969
+ notificationTo: "",
5970
+ notificationLink: "",
5971
+ userId: "",
6279
5972
  };
6280
5973
  }
6281
- exports.CrawlerDataForDD = {
5974
+ exports.MarketplaceCrawlerDataForDDSubmission = {
6282
5975
  encode(message, writer = new wire_1.BinaryWriter()) {
6283
5976
  if (message.crawlerId !== "") {
6284
5977
  writer.uint32(10).string(message.crawlerId);
@@ -6298,12 +5991,39 @@ exports.CrawlerDataForDD = {
6298
5991
  if (message.postEndDatetime !== undefined) {
6299
5992
  writer.uint32(50).string(message.postEndDatetime);
6300
5993
  }
5994
+ if (message.startTime !== undefined) {
5995
+ timestamp_1.Timestamp.encode(toTimestamp(message.startTime), writer.uint32(58).fork()).join();
5996
+ }
5997
+ if (message.deregistrationTime !== undefined) {
5998
+ timestamp_1.Timestamp.encode(toTimestamp(message.deregistrationTime), writer.uint32(66).fork()).join();
5999
+ }
6000
+ if (message.archiveTime !== undefined) {
6001
+ timestamp_1.Timestamp.encode(toTimestamp(message.archiveTime), writer.uint32(74).fork()).join();
6002
+ }
6003
+ if (message.status !== "") {
6004
+ writer.uint32(82).string(message.status);
6005
+ }
6006
+ if (message.bytesCollected !== 0) {
6007
+ writer.uint32(88).uint64(message.bytesCollected);
6008
+ }
6009
+ if (message.recordsCollected !== 0) {
6010
+ writer.uint32(96).uint64(message.recordsCollected);
6011
+ }
6012
+ if (message.notificationTo !== "") {
6013
+ writer.uint32(106).string(message.notificationTo);
6014
+ }
6015
+ if (message.notificationLink !== "") {
6016
+ writer.uint32(114).string(message.notificationLink);
6017
+ }
6018
+ if (message.userId !== "") {
6019
+ writer.uint32(122).string(message.userId);
6020
+ }
6301
6021
  return writer;
6302
6022
  },
6303
6023
  decode(input, length) {
6304
6024
  const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
6305
6025
  let end = length === undefined ? reader.len : reader.pos + length;
6306
- const message = createBaseCrawlerDataForDD();
6026
+ const message = createBaseMarketplaceCrawlerDataForDDSubmission();
6307
6027
  while (reader.pos < end) {
6308
6028
  const tag = reader.uint32();
6309
6029
  switch (tag >>> 3) {
@@ -6349,6 +6069,69 @@ exports.CrawlerDataForDD = {
6349
6069
  message.postEndDatetime = reader.string();
6350
6070
  continue;
6351
6071
  }
6072
+ case 7: {
6073
+ if (tag !== 58) {
6074
+ break;
6075
+ }
6076
+ message.startTime = fromTimestamp(timestamp_1.Timestamp.decode(reader, reader.uint32()));
6077
+ continue;
6078
+ }
6079
+ case 8: {
6080
+ if (tag !== 66) {
6081
+ break;
6082
+ }
6083
+ message.deregistrationTime = fromTimestamp(timestamp_1.Timestamp.decode(reader, reader.uint32()));
6084
+ continue;
6085
+ }
6086
+ case 9: {
6087
+ if (tag !== 74) {
6088
+ break;
6089
+ }
6090
+ message.archiveTime = fromTimestamp(timestamp_1.Timestamp.decode(reader, reader.uint32()));
6091
+ continue;
6092
+ }
6093
+ case 10: {
6094
+ if (tag !== 82) {
6095
+ break;
6096
+ }
6097
+ message.status = reader.string();
6098
+ continue;
6099
+ }
6100
+ case 11: {
6101
+ if (tag !== 88) {
6102
+ break;
6103
+ }
6104
+ message.bytesCollected = longToNumber(reader.uint64());
6105
+ continue;
6106
+ }
6107
+ case 12: {
6108
+ if (tag !== 96) {
6109
+ break;
6110
+ }
6111
+ message.recordsCollected = longToNumber(reader.uint64());
6112
+ continue;
6113
+ }
6114
+ case 13: {
6115
+ if (tag !== 106) {
6116
+ break;
6117
+ }
6118
+ message.notificationTo = reader.string();
6119
+ continue;
6120
+ }
6121
+ case 14: {
6122
+ if (tag !== 114) {
6123
+ break;
6124
+ }
6125
+ message.notificationLink = reader.string();
6126
+ continue;
6127
+ }
6128
+ case 15: {
6129
+ if (tag !== 122) {
6130
+ break;
6131
+ }
6132
+ message.userId = reader.string();
6133
+ continue;
6134
+ }
6352
6135
  }
6353
6136
  if ((tag & 7) === 4 || tag === 0) {
6354
6137
  break;
@@ -6365,7 +6148,9 @@ exports.CrawlerDataForDD = {
6365
6148
  platform: isSet(object.platform)
6366
6149
  ? globalThis.String(object.platform)
6367
6150
  : "",
6368
- topic: isSet(object.topic) ? globalThis.String(object.topic) : undefined,
6151
+ topic: isSet(object.topic)
6152
+ ? globalThis.String(object.topic)
6153
+ : undefined,
6369
6154
  keyword: isSet(object.keyword)
6370
6155
  ? globalThis.String(object.keyword)
6371
6156
  : undefined,
@@ -6375,6 +6160,29 @@ exports.CrawlerDataForDD = {
6375
6160
  postEndDatetime: isSet(object.postEndDatetime)
6376
6161
  ? globalThis.String(object.postEndDatetime)
6377
6162
  : undefined,
6163
+ startTime: isSet(object.startTime)
6164
+ ? fromJsonTimestamp(object.startTime)
6165
+ : undefined,
6166
+ deregistrationTime: isSet(object.deregistrationTime)
6167
+ ? fromJsonTimestamp(object.deregistrationTime)
6168
+ : undefined,
6169
+ archiveTime: isSet(object.archiveTime)
6170
+ ? fromJsonTimestamp(object.archiveTime)
6171
+ : undefined,
6172
+ status: isSet(object.status) ? globalThis.String(object.status) : "",
6173
+ bytesCollected: isSet(object.bytesCollected)
6174
+ ? globalThis.Number(object.bytesCollected)
6175
+ : 0,
6176
+ recordsCollected: isSet(object.recordsCollected)
6177
+ ? globalThis.Number(object.recordsCollected)
6178
+ : 0,
6179
+ notificationTo: isSet(object.notificationTo)
6180
+ ? globalThis.String(object.notificationTo)
6181
+ : "",
6182
+ notificationLink: isSet(object.notificationLink)
6183
+ ? globalThis.String(object.notificationLink)
6184
+ : "",
6185
+ userId: isSet(object.userId) ? globalThis.String(object.userId) : "",
6378
6186
  };
6379
6187
  },
6380
6188
  toJSON(message) {
@@ -6397,112 +6205,594 @@ exports.CrawlerDataForDD = {
6397
6205
  if (message.postEndDatetime !== undefined) {
6398
6206
  obj.postEndDatetime = message.postEndDatetime;
6399
6207
  }
6208
+ if (message.startTime !== undefined) {
6209
+ obj.startTime = message.startTime.toISOString();
6210
+ }
6211
+ if (message.deregistrationTime !== undefined) {
6212
+ obj.deregistrationTime = message.deregistrationTime.toISOString();
6213
+ }
6214
+ if (message.archiveTime !== undefined) {
6215
+ obj.archiveTime = message.archiveTime.toISOString();
6216
+ }
6217
+ if (message.status !== "") {
6218
+ obj.status = message.status;
6219
+ }
6220
+ if (message.bytesCollected !== 0) {
6221
+ obj.bytesCollected = Math.round(message.bytesCollected);
6222
+ }
6223
+ if (message.recordsCollected !== 0) {
6224
+ obj.recordsCollected = Math.round(message.recordsCollected);
6225
+ }
6226
+ if (message.notificationTo !== "") {
6227
+ obj.notificationTo = message.notificationTo;
6228
+ }
6229
+ if (message.notificationLink !== "") {
6230
+ obj.notificationLink = message.notificationLink;
6231
+ }
6232
+ if (message.userId !== "") {
6233
+ obj.userId = message.userId;
6234
+ }
6400
6235
  return obj;
6401
6236
  },
6402
6237
  create(base) {
6403
- return exports.CrawlerDataForDD.fromPartial(base ?? {});
6238
+ return exports.MarketplaceCrawlerDataForDDSubmission.fromPartial(base ?? {});
6404
6239
  },
6405
6240
  fromPartial(object) {
6406
- const message = createBaseCrawlerDataForDD();
6241
+ const message = createBaseMarketplaceCrawlerDataForDDSubmission();
6407
6242
  message.crawlerId = object.crawlerId ?? "";
6408
6243
  message.platform = object.platform ?? "";
6409
6244
  message.topic = object.topic ?? undefined;
6410
6245
  message.keyword = object.keyword ?? undefined;
6411
6246
  message.postStartDatetime = object.postStartDatetime ?? undefined;
6412
6247
  message.postEndDatetime = object.postEndDatetime ?? undefined;
6248
+ message.startTime = object.startTime ?? undefined;
6249
+ message.deregistrationTime = object.deregistrationTime ?? undefined;
6250
+ message.archiveTime = object.archiveTime ?? undefined;
6251
+ message.status = object.status ?? "";
6252
+ message.bytesCollected = object.bytesCollected ?? 0;
6253
+ message.recordsCollected = object.recordsCollected ?? 0;
6254
+ message.notificationTo = object.notificationTo ?? "";
6255
+ message.notificationLink = object.notificationLink ?? "";
6256
+ message.userId = object.userId ?? "";
6413
6257
  return message;
6414
6258
  },
6415
6259
  };
6416
- exports.GravityServiceService = {
6417
- /** Lists all data collection tasks for a user */
6418
- getPopularTags: {
6419
- path: "/gravity.v1.GravityService/GetPopularTags",
6420
- requestStream: false,
6421
- responseStream: false,
6422
- requestSerialize: (value) => Buffer.from(empty_1.Empty.encode(value).finish()),
6423
- requestDeserialize: (value) => empty_1.Empty.decode(value),
6424
- responseSerialize: (value) => Buffer.from(exports.GetPopularTagsResponse.encode(value).finish()),
6425
- responseDeserialize: (value) => exports.GetPopularTagsResponse.decode(value),
6260
+ function createBaseGetActiveUserTasksResponse() {
6261
+ return { activeUserTasks: [] };
6262
+ }
6263
+ exports.GetActiveUserTasksResponse = {
6264
+ encode(message, writer = new wire_1.BinaryWriter()) {
6265
+ for (const v of message.activeUserTasks) {
6266
+ exports.ActiveUserTask.encode(v, writer.uint32(10).fork()).join();
6267
+ }
6268
+ return writer;
6426
6269
  },
6427
- /** Lists all data collection tasks for a user */
6428
- getGravityTasks: {
6429
- path: "/gravity.v1.GravityService/GetGravityTasks",
6430
- requestStream: false,
6431
- responseStream: false,
6432
- requestSerialize: (value) => Buffer.from(exports.GetGravityTasksRequest.encode(value).finish()),
6433
- requestDeserialize: (value) => exports.GetGravityTasksRequest.decode(value),
6434
- responseSerialize: (value) => Buffer.from(exports.GetGravityTasksResponse.encode(value).finish()),
6435
- responseDeserialize: (value) => exports.GetGravityTasksResponse.decode(value),
6270
+ decode(input, length) {
6271
+ const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
6272
+ let end = length === undefined ? reader.len : reader.pos + length;
6273
+ const message = createBaseGetActiveUserTasksResponse();
6274
+ while (reader.pos < end) {
6275
+ const tag = reader.uint32();
6276
+ switch (tag >>> 3) {
6277
+ case 1: {
6278
+ if (tag !== 10) {
6279
+ break;
6280
+ }
6281
+ message.activeUserTasks.push(exports.ActiveUserTask.decode(reader, reader.uint32()));
6282
+ continue;
6283
+ }
6284
+ }
6285
+ if ((tag & 7) === 4 || tag === 0) {
6286
+ break;
6287
+ }
6288
+ reader.skip(tag & 7);
6289
+ }
6290
+ return message;
6436
6291
  },
6437
- /** Get all marketplace crawlers */
6438
- getMarketplaceCrawlers: {
6439
- path: "/gravity.v1.GravityService/GetMarketplaceCrawlers",
6440
- requestStream: false,
6441
- responseStream: false,
6442
- requestSerialize: (value) => Buffer.from(empty_1.Empty.encode(value).finish()),
6443
- requestDeserialize: (value) => empty_1.Empty.decode(value),
6444
- responseSerialize: (value) => Buffer.from(exports.GetMarketplaceCrawlersResponse.encode(value).finish()),
6445
- responseDeserialize: (value) => exports.GetMarketplaceCrawlersResponse.decode(value),
6292
+ fromJSON(object) {
6293
+ return {
6294
+ activeUserTasks: globalThis.Array.isArray(object?.activeUserTasks)
6295
+ ? object.activeUserTasks.map((e) => exports.ActiveUserTask.fromJSON(e))
6296
+ : [],
6297
+ };
6446
6298
  },
6447
- /** Gets raw miner files for a specific crawler */
6448
- getCrawlerRawMinerFiles: {
6449
- path: "/gravity.v1.GravityService/GetCrawlerRawMinerFiles",
6450
- requestStream: false,
6451
- responseStream: false,
6452
- requestSerialize: (value) => Buffer.from(exports.GetCrawlerRequest.encode(value).finish()),
6453
- requestDeserialize: (value) => exports.GetCrawlerRequest.decode(value),
6454
- responseSerialize: (value) => Buffer.from(exports.CrawlerRawMinerFilesResponse.encode(value).finish()),
6455
- responseDeserialize: (value) => exports.CrawlerRawMinerFilesResponse.decode(value),
6299
+ toJSON(message) {
6300
+ const obj = {};
6301
+ if (message.activeUserTasks?.length) {
6302
+ obj.activeUserTasks = message.activeUserTasks.map(e => exports.ActiveUserTask.toJSON(e));
6303
+ }
6304
+ return obj;
6456
6305
  },
6457
- /** Get the parent workflow id (the id of the ui workflow) for this crawler */
6458
- getCrawlerParentTaskId: {
6459
- path: "/gravity.v1.GravityService/GetCrawlerParentTaskId",
6460
- requestStream: false,
6461
- responseStream: false,
6462
- requestSerialize: (value) => Buffer.from(exports.GetCrawlerRequest.encode(value).finish()),
6463
- requestDeserialize: (value) => exports.GetCrawlerRequest.decode(value),
6464
- responseSerialize: (value) => Buffer.from(exports.CreateGravityTaskResponse.encode(value).finish()),
6465
- responseDeserialize: (value) => exports.CreateGravityTaskResponse.decode(value),
6306
+ create(base) {
6307
+ return exports.GetActiveUserTasksResponse.fromPartial(base ?? {});
6308
+ },
6309
+ fromPartial(object) {
6310
+ const message = createBaseGetActiveUserTasksResponse();
6311
+ message.activeUserTasks =
6312
+ object.activeUserTasks?.map(e => exports.ActiveUserTask.fromPartial(e)) || [];
6313
+ return message;
6314
+ },
6315
+ };
6316
+ function createBaseActiveUserCrawler() {
6317
+ return { crawlerId: "", rowCount: 0 };
6318
+ }
6319
+ exports.ActiveUserCrawler = {
6320
+ encode(message, writer = new wire_1.BinaryWriter()) {
6321
+ if (message.crawlerId !== "") {
6322
+ writer.uint32(10).string(message.crawlerId);
6323
+ }
6324
+ if (message.rowCount !== 0) {
6325
+ writer.uint32(16).uint64(message.rowCount);
6326
+ }
6327
+ return writer;
6328
+ },
6329
+ decode(input, length) {
6330
+ const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
6331
+ let end = length === undefined ? reader.len : reader.pos + length;
6332
+ const message = createBaseActiveUserCrawler();
6333
+ while (reader.pos < end) {
6334
+ const tag = reader.uint32();
6335
+ switch (tag >>> 3) {
6336
+ case 1: {
6337
+ if (tag !== 10) {
6338
+ break;
6339
+ }
6340
+ message.crawlerId = reader.string();
6341
+ continue;
6342
+ }
6343
+ case 2: {
6344
+ if (tag !== 16) {
6345
+ break;
6346
+ }
6347
+ message.rowCount = longToNumber(reader.uint64());
6348
+ continue;
6349
+ }
6350
+ }
6351
+ if ((tag & 7) === 4 || tag === 0) {
6352
+ break;
6353
+ }
6354
+ reader.skip(tag & 7);
6355
+ }
6356
+ return message;
6357
+ },
6358
+ fromJSON(object) {
6359
+ return {
6360
+ crawlerId: isSet(object.crawlerId)
6361
+ ? globalThis.String(object.crawlerId)
6362
+ : "",
6363
+ rowCount: isSet(object.rowCount) ? globalThis.Number(object.rowCount) : 0,
6364
+ };
6365
+ },
6366
+ toJSON(message) {
6367
+ const obj = {};
6368
+ if (message.crawlerId !== "") {
6369
+ obj.crawlerId = message.crawlerId;
6370
+ }
6371
+ if (message.rowCount !== 0) {
6372
+ obj.rowCount = Math.round(message.rowCount);
6373
+ }
6374
+ return obj;
6375
+ },
6376
+ create(base) {
6377
+ return exports.ActiveUserCrawler.fromPartial(base ?? {});
6378
+ },
6379
+ fromPartial(object) {
6380
+ const message = createBaseActiveUserCrawler();
6381
+ message.crawlerId = object.crawlerId ?? "";
6382
+ message.rowCount = object.rowCount ?? 0;
6383
+ return message;
6384
+ },
6385
+ };
6386
+ function createBaseActiveUserTask() {
6387
+ return { gravityTaskId: "", crawlers: [] };
6388
+ }
6389
+ exports.ActiveUserTask = {
6390
+ encode(message, writer = new wire_1.BinaryWriter()) {
6391
+ if (message.gravityTaskId !== "") {
6392
+ writer.uint32(10).string(message.gravityTaskId);
6393
+ }
6394
+ for (const v of message.crawlers) {
6395
+ exports.ActiveUserCrawler.encode(v, writer.uint32(18).fork()).join();
6396
+ }
6397
+ return writer;
6398
+ },
6399
+ decode(input, length) {
6400
+ const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
6401
+ let end = length === undefined ? reader.len : reader.pos + length;
6402
+ const message = createBaseActiveUserTask();
6403
+ while (reader.pos < end) {
6404
+ const tag = reader.uint32();
6405
+ switch (tag >>> 3) {
6406
+ case 1: {
6407
+ if (tag !== 10) {
6408
+ break;
6409
+ }
6410
+ message.gravityTaskId = reader.string();
6411
+ continue;
6412
+ }
6413
+ case 2: {
6414
+ if (tag !== 18) {
6415
+ break;
6416
+ }
6417
+ message.crawlers.push(exports.ActiveUserCrawler.decode(reader, reader.uint32()));
6418
+ continue;
6419
+ }
6420
+ }
6421
+ if ((tag & 7) === 4 || tag === 0) {
6422
+ break;
6423
+ }
6424
+ reader.skip(tag & 7);
6425
+ }
6426
+ return message;
6466
6427
  },
6467
- /** Add a persistent gravity task to the Gravity state DB */
6468
- addPersistentGravityTask: {
6469
- path: "/gravity.v1.GravityService/AddPersistentGravityTask",
6428
+ fromJSON(object) {
6429
+ return {
6430
+ gravityTaskId: isSet(object.gravityTaskId)
6431
+ ? globalThis.String(object.gravityTaskId)
6432
+ : "",
6433
+ crawlers: globalThis.Array.isArray(object?.crawlers)
6434
+ ? object.crawlers.map((e) => exports.ActiveUserCrawler.fromJSON(e))
6435
+ : [],
6436
+ };
6437
+ },
6438
+ toJSON(message) {
6439
+ const obj = {};
6440
+ if (message.gravityTaskId !== "") {
6441
+ obj.gravityTaskId = message.gravityTaskId;
6442
+ }
6443
+ if (message.crawlers?.length) {
6444
+ obj.crawlers = message.crawlers.map(e => exports.ActiveUserCrawler.toJSON(e));
6445
+ }
6446
+ return obj;
6447
+ },
6448
+ create(base) {
6449
+ return exports.ActiveUserTask.fromPartial(base ?? {});
6450
+ },
6451
+ fromPartial(object) {
6452
+ const message = createBaseActiveUserTask();
6453
+ message.gravityTaskId = object.gravityTaskId ?? "";
6454
+ message.crawlers =
6455
+ object.crawlers?.map(e => exports.ActiveUserCrawler.fromPartial(e)) || [];
6456
+ return message;
6457
+ },
6458
+ };
6459
+ function createBaseUpsertPreBuiltUserDatasetsRequest() {
6460
+ return { gravityTaskId: "", crawlerId: "", rowCount: 0 };
6461
+ }
6462
+ exports.UpsertPreBuiltUserDatasetsRequest = {
6463
+ encode(message, writer = new wire_1.BinaryWriter()) {
6464
+ if (message.gravityTaskId !== "") {
6465
+ writer.uint32(10).string(message.gravityTaskId);
6466
+ }
6467
+ if (message.crawlerId !== "") {
6468
+ writer.uint32(18).string(message.crawlerId);
6469
+ }
6470
+ if (message.rowCount !== 0) {
6471
+ writer.uint32(24).int64(message.rowCount);
6472
+ }
6473
+ return writer;
6474
+ },
6475
+ decode(input, length) {
6476
+ const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
6477
+ let end = length === undefined ? reader.len : reader.pos + length;
6478
+ const message = createBaseUpsertPreBuiltUserDatasetsRequest();
6479
+ while (reader.pos < end) {
6480
+ const tag = reader.uint32();
6481
+ switch (tag >>> 3) {
6482
+ case 1: {
6483
+ if (tag !== 10) {
6484
+ break;
6485
+ }
6486
+ message.gravityTaskId = reader.string();
6487
+ continue;
6488
+ }
6489
+ case 2: {
6490
+ if (tag !== 18) {
6491
+ break;
6492
+ }
6493
+ message.crawlerId = reader.string();
6494
+ continue;
6495
+ }
6496
+ case 3: {
6497
+ if (tag !== 24) {
6498
+ break;
6499
+ }
6500
+ message.rowCount = longToNumber(reader.int64());
6501
+ continue;
6502
+ }
6503
+ }
6504
+ if ((tag & 7) === 4 || tag === 0) {
6505
+ break;
6506
+ }
6507
+ reader.skip(tag & 7);
6508
+ }
6509
+ return message;
6510
+ },
6511
+ fromJSON(object) {
6512
+ return {
6513
+ gravityTaskId: isSet(object.gravityTaskId)
6514
+ ? globalThis.String(object.gravityTaskId)
6515
+ : "",
6516
+ crawlerId: isSet(object.crawlerId)
6517
+ ? globalThis.String(object.crawlerId)
6518
+ : "",
6519
+ rowCount: isSet(object.rowCount)
6520
+ ? globalThis.Number(object.rowCount)
6521
+ : 0,
6522
+ };
6523
+ },
6524
+ toJSON(message) {
6525
+ const obj = {};
6526
+ if (message.gravityTaskId !== "") {
6527
+ obj.gravityTaskId = message.gravityTaskId;
6528
+ }
6529
+ if (message.crawlerId !== "") {
6530
+ obj.crawlerId = message.crawlerId;
6531
+ }
6532
+ if (message.rowCount !== 0) {
6533
+ obj.rowCount = Math.round(message.rowCount);
6534
+ }
6535
+ return obj;
6536
+ },
6537
+ create(base) {
6538
+ return exports.UpsertPreBuiltUserDatasetsRequest.fromPartial(base ?? {});
6539
+ },
6540
+ fromPartial(object) {
6541
+ const message = createBaseUpsertPreBuiltUserDatasetsRequest();
6542
+ message.gravityTaskId = object.gravityTaskId ?? "";
6543
+ message.crawlerId = object.crawlerId ?? "";
6544
+ message.rowCount = object.rowCount ?? 0;
6545
+ return message;
6546
+ },
6547
+ };
6548
+ function createBaseGetPreBuiltUserDatasetsRequest() {
6549
+ return { gravityTaskId: "" };
6550
+ }
6551
+ exports.GetPreBuiltUserDatasetsRequest = {
6552
+ encode(message, writer = new wire_1.BinaryWriter()) {
6553
+ if (message.gravityTaskId !== "") {
6554
+ writer.uint32(10).string(message.gravityTaskId);
6555
+ }
6556
+ return writer;
6557
+ },
6558
+ decode(input, length) {
6559
+ const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
6560
+ let end = length === undefined ? reader.len : reader.pos + length;
6561
+ const message = createBaseGetPreBuiltUserDatasetsRequest();
6562
+ while (reader.pos < end) {
6563
+ const tag = reader.uint32();
6564
+ switch (tag >>> 3) {
6565
+ case 1: {
6566
+ if (tag !== 10) {
6567
+ break;
6568
+ }
6569
+ message.gravityTaskId = reader.string();
6570
+ continue;
6571
+ }
6572
+ }
6573
+ if ((tag & 7) === 4 || tag === 0) {
6574
+ break;
6575
+ }
6576
+ reader.skip(tag & 7);
6577
+ }
6578
+ return message;
6579
+ },
6580
+ fromJSON(object) {
6581
+ return {
6582
+ gravityTaskId: isSet(object.gravityTaskId)
6583
+ ? globalThis.String(object.gravityTaskId)
6584
+ : "",
6585
+ };
6586
+ },
6587
+ toJSON(message) {
6588
+ const obj = {};
6589
+ if (message.gravityTaskId !== "") {
6590
+ obj.gravityTaskId = message.gravityTaskId;
6591
+ }
6592
+ return obj;
6593
+ },
6594
+ create(base) {
6595
+ return exports.GetPreBuiltUserDatasetsRequest.fromPartial(base ?? {});
6596
+ },
6597
+ fromPartial(object) {
6598
+ const message = createBaseGetPreBuiltUserDatasetsRequest();
6599
+ message.gravityTaskId = object.gravityTaskId ?? "";
6600
+ return message;
6601
+ },
6602
+ };
6603
+ function createBasePreBuiltUserDataset() {
6604
+ return { gravityTaskId: "", crawlerId: "", rowCount: 0 };
6605
+ }
6606
+ exports.PreBuiltUserDataset = {
6607
+ encode(message, writer = new wire_1.BinaryWriter()) {
6608
+ if (message.gravityTaskId !== "") {
6609
+ writer.uint32(10).string(message.gravityTaskId);
6610
+ }
6611
+ if (message.crawlerId !== "") {
6612
+ writer.uint32(18).string(message.crawlerId);
6613
+ }
6614
+ if (message.rowCount !== 0) {
6615
+ writer.uint32(24).int64(message.rowCount);
6616
+ }
6617
+ return writer;
6618
+ },
6619
+ decode(input, length) {
6620
+ const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
6621
+ let end = length === undefined ? reader.len : reader.pos + length;
6622
+ const message = createBasePreBuiltUserDataset();
6623
+ while (reader.pos < end) {
6624
+ const tag = reader.uint32();
6625
+ switch (tag >>> 3) {
6626
+ case 1: {
6627
+ if (tag !== 10) {
6628
+ break;
6629
+ }
6630
+ message.gravityTaskId = reader.string();
6631
+ continue;
6632
+ }
6633
+ case 2: {
6634
+ if (tag !== 18) {
6635
+ break;
6636
+ }
6637
+ message.crawlerId = reader.string();
6638
+ continue;
6639
+ }
6640
+ case 3: {
6641
+ if (tag !== 24) {
6642
+ break;
6643
+ }
6644
+ message.rowCount = longToNumber(reader.int64());
6645
+ continue;
6646
+ }
6647
+ }
6648
+ if ((tag & 7) === 4 || tag === 0) {
6649
+ break;
6650
+ }
6651
+ reader.skip(tag & 7);
6652
+ }
6653
+ return message;
6654
+ },
6655
+ fromJSON(object) {
6656
+ return {
6657
+ gravityTaskId: isSet(object.gravityTaskId)
6658
+ ? globalThis.String(object.gravityTaskId)
6659
+ : "",
6660
+ crawlerId: isSet(object.crawlerId)
6661
+ ? globalThis.String(object.crawlerId)
6662
+ : "",
6663
+ rowCount: isSet(object.rowCount) ? globalThis.Number(object.rowCount) : 0,
6664
+ };
6665
+ },
6666
+ toJSON(message) {
6667
+ const obj = {};
6668
+ if (message.gravityTaskId !== "") {
6669
+ obj.gravityTaskId = message.gravityTaskId;
6670
+ }
6671
+ if (message.crawlerId !== "") {
6672
+ obj.crawlerId = message.crawlerId;
6673
+ }
6674
+ if (message.rowCount !== 0) {
6675
+ obj.rowCount = Math.round(message.rowCount);
6676
+ }
6677
+ return obj;
6678
+ },
6679
+ create(base) {
6680
+ return exports.PreBuiltUserDataset.fromPartial(base ?? {});
6681
+ },
6682
+ fromPartial(object) {
6683
+ const message = createBasePreBuiltUserDataset();
6684
+ message.gravityTaskId = object.gravityTaskId ?? "";
6685
+ message.crawlerId = object.crawlerId ?? "";
6686
+ message.rowCount = object.rowCount ?? 0;
6687
+ return message;
6688
+ },
6689
+ };
6690
+ function createBaseGetPreBuiltUserDatasetsResponse() {
6691
+ return { datasets: [] };
6692
+ }
6693
+ exports.GetPreBuiltUserDatasetsResponse = {
6694
+ encode(message, writer = new wire_1.BinaryWriter()) {
6695
+ for (const v of message.datasets) {
6696
+ exports.PreBuiltUserDataset.encode(v, writer.uint32(10).fork()).join();
6697
+ }
6698
+ return writer;
6699
+ },
6700
+ decode(input, length) {
6701
+ const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
6702
+ let end = length === undefined ? reader.len : reader.pos + length;
6703
+ const message = createBaseGetPreBuiltUserDatasetsResponse();
6704
+ while (reader.pos < end) {
6705
+ const tag = reader.uint32();
6706
+ switch (tag >>> 3) {
6707
+ case 1: {
6708
+ if (tag !== 10) {
6709
+ break;
6710
+ }
6711
+ message.datasets.push(exports.PreBuiltUserDataset.decode(reader, reader.uint32()));
6712
+ continue;
6713
+ }
6714
+ }
6715
+ if ((tag & 7) === 4 || tag === 0) {
6716
+ break;
6717
+ }
6718
+ reader.skip(tag & 7);
6719
+ }
6720
+ return message;
6721
+ },
6722
+ fromJSON(object) {
6723
+ return {
6724
+ datasets: globalThis.Array.isArray(object?.datasets)
6725
+ ? object.datasets.map((e) => exports.PreBuiltUserDataset.fromJSON(e))
6726
+ : [],
6727
+ };
6728
+ },
6729
+ toJSON(message) {
6730
+ const obj = {};
6731
+ if (message.datasets?.length) {
6732
+ obj.datasets = message.datasets.map(e => exports.PreBuiltUserDataset.toJSON(e));
6733
+ }
6734
+ return obj;
6735
+ },
6736
+ create(base) {
6737
+ return exports.GetPreBuiltUserDatasetsResponse.fromPartial(base ?? {});
6738
+ },
6739
+ fromPartial(object) {
6740
+ const message = createBaseGetPreBuiltUserDatasetsResponse();
6741
+ message.datasets =
6742
+ object.datasets?.map(e => exports.PreBuiltUserDataset.fromPartial(e)) || [];
6743
+ return message;
6744
+ },
6745
+ };
6746
+ exports.GravityServiceService = {
6747
+ /** Lists all data collection tasks for a user */
6748
+ getPopularTags: {
6749
+ path: "/gravity.v1.GravityService/GetPopularTags",
6470
6750
  requestStream: false,
6471
6751
  responseStream: false,
6472
- requestSerialize: (value) => Buffer.from(exports.AddPersistentGravityTaskRequest.encode(value).finish()),
6473
- requestDeserialize: (value) => exports.AddPersistentGravityTaskRequest.decode(value),
6474
- responseSerialize: (value) => Buffer.from(exports.UpsertResponse.encode(value).finish()),
6475
- responseDeserialize: (value) => exports.UpsertResponse.decode(value),
6752
+ requestSerialize: (value) => Buffer.from(empty_1.Empty.encode(value).finish()),
6753
+ requestDeserialize: (value) => empty_1.Empty.decode(value),
6754
+ responseSerialize: (value) => Buffer.from(exports.GetPopularTagsResponse.encode(value).finish()),
6755
+ responseDeserialize: (value) => exports.GetPopularTagsResponse.decode(value),
6476
6756
  },
6477
- /** Add a persistent dataset workflow to the Gravity state DB */
6478
- addPersistentDatasetWorkflows: {
6479
- path: "/gravity.v1.GravityService/AddPersistentDatasetWorkflows",
6757
+ /** Lists all data collection tasks for a user */
6758
+ getGravityTasks: {
6759
+ path: "/gravity.v1.GravityService/GetGravityTasks",
6480
6760
  requestStream: false,
6481
6761
  responseStream: false,
6482
- requestSerialize: (value) => Buffer.from(exports.AddPersistentDatasetWorkflowsRequest.encode(value).finish()),
6483
- requestDeserialize: (value) => exports.AddPersistentDatasetWorkflowsRequest.decode(value),
6484
- responseSerialize: (value) => Buffer.from(exports.UpsertResponse.encode(value).finish()),
6485
- responseDeserialize: (value) => exports.UpsertResponse.decode(value),
6762
+ requestSerialize: (value) => Buffer.from(exports.GetGravityTasksRequest.encode(value).finish()),
6763
+ requestDeserialize: (value) => exports.GetGravityTasksRequest.decode(value),
6764
+ responseSerialize: (value) => Buffer.from(exports.GetGravityTasksResponse.encode(value).finish()),
6765
+ responseDeserialize: (value) => exports.GetGravityTasksResponse.decode(value),
6486
6766
  },
6487
- /** Retrieve recent persistent dataset workflows (last 7 days, non-Completed) */
6488
- getPersistentDatasetWorkflows: {
6489
- path: "/gravity.v1.GravityService/GetPersistentDatasetWorkflows",
6767
+ /** Get all marketplace crawlers */
6768
+ getMarketplaceCrawlers: {
6769
+ path: "/gravity.v1.GravityService/GetMarketplaceCrawlers",
6490
6770
  requestStream: false,
6491
6771
  responseStream: false,
6492
6772
  requestSerialize: (value) => Buffer.from(empty_1.Empty.encode(value).finish()),
6493
6773
  requestDeserialize: (value) => empty_1.Empty.decode(value),
6494
- responseSerialize: (value) => Buffer.from(exports.GetPersistentDatasetWorkflowsResponse.encode(value).finish()),
6495
- responseDeserialize: (value) => exports.GetPersistentDatasetWorkflowsResponse.decode(value),
6774
+ responseSerialize: (value) => Buffer.from(exports.GetMarketplaceCrawlersResponse.encode(value).finish()),
6775
+ responseDeserialize: (value) => exports.GetMarketplaceCrawlersResponse.decode(value),
6496
6776
  },
6497
- /** Retrieve all persistent gravity tasks from the Gravity state DB */
6498
- getPersistentGravityTasks: {
6499
- path: "/gravity.v1.GravityService/GetPersistentGravityTasks",
6777
+ /** Gets raw miner files for a specific crawler */
6778
+ getCrawlerRawMinerFiles: {
6779
+ path: "/gravity.v1.GravityService/GetCrawlerRawMinerFiles",
6500
6780
  requestStream: false,
6501
6781
  responseStream: false,
6502
- requestSerialize: (value) => Buffer.from(empty_1.Empty.encode(value).finish()),
6503
- requestDeserialize: (value) => empty_1.Empty.decode(value),
6504
- responseSerialize: (value) => Buffer.from(exports.GetPersistentGravityTasksResponse.encode(value).finish()),
6505
- responseDeserialize: (value) => exports.GetPersistentGravityTasksResponse.decode(value),
6782
+ requestSerialize: (value) => Buffer.from(exports.GetCrawlerRequest.encode(value).finish()),
6783
+ requestDeserialize: (value) => exports.GetCrawlerRequest.decode(value),
6784
+ responseSerialize: (value) => Buffer.from(exports.CrawlerRawMinerFilesResponse.encode(value).finish()),
6785
+ responseDeserialize: (value) => exports.CrawlerRawMinerFilesResponse.decode(value),
6786
+ },
6787
+ /** Get the parent workflow id (the id of the ui workflow) for this crawler */
6788
+ getCrawlerParentTaskId: {
6789
+ path: "/gravity.v1.GravityService/GetCrawlerParentTaskId",
6790
+ requestStream: false,
6791
+ responseStream: false,
6792
+ requestSerialize: (value) => Buffer.from(exports.GetCrawlerRequest.encode(value).finish()),
6793
+ requestDeserialize: (value) => exports.GetCrawlerRequest.decode(value),
6794
+ responseSerialize: (value) => Buffer.from(exports.CreateGravityTaskResponse.encode(value).finish()),
6795
+ responseDeserialize: (value) => exports.CreateGravityTaskResponse.decode(value),
6506
6796
  },
6507
6797
  /** Get a single crawler by its ID */
6508
6798
  getCrawler: {
@@ -6554,6 +6844,16 @@ exports.GravityServiceService = {
6554
6844
  responseSerialize: (value) => Buffer.from(exports.CreateGravityTaskResponse.encode(value).finish()),
6555
6845
  responseDeserialize: (value) => exports.CreateGravityTaskResponse.decode(value),
6556
6846
  },
6847
+ /** Gets all dataset files for a given marketplace gravity task (no user_id check, validates against marketplace tasks table) */
6848
+ getGravityMarketplaceTaskDatasetFiles: {
6849
+ path: "/gravity.v1.GravityService/GetGravityMarketplaceTaskDatasetFiles",
6850
+ requestStream: false,
6851
+ responseStream: false,
6852
+ requestSerialize: (value) => Buffer.from(exports.GetGravityTaskDatasetFilesRequest.encode(value).finish()),
6853
+ requestDeserialize: (value) => exports.GetGravityTaskDatasetFilesRequest.decode(value),
6854
+ responseSerialize: (value) => Buffer.from(exports.GetGravityTaskDatasetFilesResponse.encode(value).finish()),
6855
+ responseDeserialize: (value) => exports.GetGravityTaskDatasetFilesResponse.decode(value),
6856
+ },
6557
6857
  /** Build a dataset for a single crawler */
6558
6858
  buildDataset: {
6559
6859
  path: "/gravity.v1.GravityService/BuildDataset",
@@ -6574,16 +6874,6 @@ exports.GravityServiceService = {
6574
6874
  responseSerialize: (value) => Buffer.from(exports.GetDatasetResponse.encode(value).finish()),
6575
6875
  responseDeserialize: (value) => exports.GetDatasetResponse.decode(value),
6576
6876
  },
6577
- /** Add a persistent topic to the Gravity state DB */
6578
- addPersistentTopic: {
6579
- path: "/gravity.v1.GravityService/AddPersistentTopic",
6580
- requestStream: false,
6581
- responseStream: false,
6582
- requestSerialize: (value) => Buffer.from(exports.PersistentTopic.encode(value).finish()),
6583
- requestDeserialize: (value) => exports.PersistentTopic.decode(value),
6584
- responseSerialize: (value) => Buffer.from(exports.UpsertResponse.encode(value).finish()),
6585
- responseDeserialize: (value) => exports.UpsertResponse.decode(value),
6586
- },
6587
6877
  /** Cancel a gravity task and any crawlers associated with it */
6588
6878
  cancelGravityTask: {
6589
6879
  path: "/gravity.v1.GravityService/CancelGravityTask",
@@ -6634,16 +6924,6 @@ exports.GravityServiceService = {
6634
6924
  responseSerialize: (value) => Buffer.from(exports.GetGravityTaskDatasetFilesResponse.encode(value).finish()),
6635
6925
  responseDeserialize: (value) => exports.GetGravityTaskDatasetFilesResponse.decode(value),
6636
6926
  },
6637
- /** Gets all dataset files for a given persistent gravity task (no user_id check, validates against persistent tasks table) */
6638
- getGravityMarketplaceTaskDatasetFiles: {
6639
- path: "/gravity.v1.GravityService/GetGravityMarketplaceTaskDatasetFiles",
6640
- requestStream: false,
6641
- responseStream: false,
6642
- requestSerialize: (value) => Buffer.from(exports.GetGravityTaskDatasetFilesRequest.encode(value).finish()),
6643
- requestDeserialize: (value) => exports.GetGravityTaskDatasetFilesRequest.decode(value),
6644
- responseSerialize: (value) => Buffer.from(exports.GetGravityTaskDatasetFilesResponse.encode(value).finish()),
6645
- responseDeserialize: (value) => exports.GetGravityTaskDatasetFilesResponse.decode(value),
6646
- },
6647
6927
  /** Publishes a dataset into the Marketplace */
6648
6928
  publishDataset: {
6649
6929
  path: "/gravity.v1.GravityService/PublishDataset",
@@ -6655,14 +6935,24 @@ exports.GravityServiceService = {
6655
6935
  responseDeserialize: (value) => exports.UpsertResponse.decode(value),
6656
6936
  },
6657
6937
  /** Get crawler data for DD submission */
6658
- getCrawlerDataForDdSubmission: {
6659
- path: "/gravity.v1.GravityService/GetCrawlerDataForDDSubmission",
6938
+ getActiveUserTasks: {
6939
+ path: "/gravity.v1.GravityService/GetActiveUserTasks",
6660
6940
  requestStream: false,
6661
6941
  responseStream: false,
6662
- requestSerialize: (value) => Buffer.from(exports.GetCrawlerDataForDDSubmissionRequest.encode(value).finish()),
6663
- requestDeserialize: (value) => exports.GetCrawlerDataForDDSubmissionRequest.decode(value),
6664
- responseSerialize: (value) => Buffer.from(exports.GetCrawlerDataForDDSubmissionResponse.encode(value).finish()),
6665
- responseDeserialize: (value) => exports.GetCrawlerDataForDDSubmissionResponse.decode(value),
6942
+ requestSerialize: (value) => Buffer.from(empty_1.Empty.encode(value).finish()),
6943
+ requestDeserialize: (value) => empty_1.Empty.decode(value),
6944
+ responseSerialize: (value) => Buffer.from(exports.GetActiveUserTasksResponse.encode(value).finish()),
6945
+ responseDeserialize: (value) => exports.GetActiveUserTasksResponse.decode(value),
6946
+ },
6947
+ /** Get crawler data for DD submission for the marketplace user */
6948
+ getMarketplaceCrawlerDataForDdSubmission: {
6949
+ path: "/gravity.v1.GravityService/GetMarketplaceCrawlerDataForDDSubmission",
6950
+ requestStream: false,
6951
+ responseStream: false,
6952
+ requestSerialize: (value) => Buffer.from(exports.GetMarketplaceCrawlerDataForDDSubmissionRequest.encode(value).finish()),
6953
+ requestDeserialize: (value) => exports.GetMarketplaceCrawlerDataForDDSubmissionRequest.decode(value),
6954
+ responseSerialize: (value) => Buffer.from(exports.GetMarketplaceCrawlerDataForDDSubmissionResponse.encode(value).finish()),
6955
+ responseDeserialize: (value) => exports.GetMarketplaceCrawlerDataForDDSubmissionResponse.decode(value),
6666
6956
  },
6667
6957
  /** Upserts a crawler into the Gravity state DB */
6668
6958
  upsertCrawler: {
@@ -6734,15 +7024,25 @@ exports.GravityServiceService = {
6734
7024
  responseSerialize: (value) => Buffer.from(exports.BuildAllDatasetsResponse.encode(value).finish()),
6735
7025
  responseDeserialize: (value) => exports.BuildAllDatasetsResponse.decode(value),
6736
7026
  },
6737
- /** Gets all persistent topics from the Gravity state DB */
6738
- getPersistentTopics: {
6739
- path: "/gravity.v1.GravityService/GetPersistentTopics",
7027
+ /** Builds datasets for multiple crawlers within a single gravity task periodically */
7028
+ buildUserDatasetsPeriodically: {
7029
+ path: "/gravity.v1.GravityService/BuildUserDatasetsPeriodically",
6740
7030
  requestStream: false,
6741
7031
  responseStream: false,
6742
- requestSerialize: (value) => Buffer.from(empty_1.Empty.encode(value).finish()),
6743
- requestDeserialize: (value) => empty_1.Empty.decode(value),
6744
- responseSerialize: (value) => Buffer.from(exports.PersistentTopicResponse.encode(value).finish()),
6745
- responseDeserialize: (value) => exports.PersistentTopicResponse.decode(value),
7032
+ requestSerialize: (value) => Buffer.from(exports.BuildAllDatasetsRequest.encode(value).finish()),
7033
+ requestDeserialize: (value) => exports.BuildAllDatasetsRequest.decode(value),
7034
+ responseSerialize: (value) => Buffer.from(exports.BuildAllDatasetsResponse.encode(value).finish()),
7035
+ responseDeserialize: (value) => exports.BuildAllDatasetsResponse.decode(value),
7036
+ },
7037
+ /** Charges a user for dataset rows */
7038
+ chargeForDatasetRows: {
7039
+ path: "/gravity.v1.GravityService/ChargeForDatasetRows",
7040
+ requestStream: false,
7041
+ responseStream: false,
7042
+ requestSerialize: (value) => Buffer.from(exports.ChargeForDatasetRowsRequest.encode(value).finish()),
7043
+ requestDeserialize: (value) => exports.ChargeForDatasetRowsRequest.decode(value),
7044
+ responseSerialize: (value) => Buffer.from(exports.UpsertResponse.encode(value).finish()),
7045
+ responseDeserialize: (value) => exports.UpsertResponse.decode(value),
6746
7046
  },
6747
7047
  /** Gets crawler history for a gravity task */
6748
7048
  getCrawlerHistory: {
@@ -6769,8 +7069,8 @@ exports.GravityServiceService = {
6769
7069
  path: "/gravity.v1.GravityService/UpsertRawMinerFiles",
6770
7070
  requestStream: false,
6771
7071
  responseStream: false,
6772
- requestSerialize: (value) => Buffer.from(exports.Crawler.encode(value).finish()),
6773
- requestDeserialize: (value) => exports.Crawler.decode(value),
7072
+ requestSerialize: (value) => Buffer.from(exports.UpsertRawMinerFilesRequest.encode(value).finish()),
7073
+ requestDeserialize: (value) => exports.UpsertRawMinerFilesRequest.decode(value),
6774
7074
  responseSerialize: (value) => Buffer.from(exports.UpsertResponse.encode(value).finish()),
6775
7075
  responseDeserialize: (value) => exports.UpsertResponse.decode(value),
6776
7076
  },
@@ -6814,6 +7114,26 @@ exports.GravityServiceService = {
6814
7114
  responseSerialize: (value) => Buffer.from(exports.GetUserMarketplaceDatasetsResponse.encode(value).finish()),
6815
7115
  responseDeserialize: (value) => exports.GetUserMarketplaceDatasetsResponse.decode(value),
6816
7116
  },
7117
+ /** Upserts pre-built user dataset records */
7118
+ upsertPreBuiltUserDatasets: {
7119
+ path: "/gravity.v1.GravityService/UpsertPreBuiltUserDatasets",
7120
+ requestStream: false,
7121
+ responseStream: false,
7122
+ requestSerialize: (value) => Buffer.from(exports.UpsertPreBuiltUserDatasetsRequest.encode(value).finish()),
7123
+ requestDeserialize: (value) => exports.UpsertPreBuiltUserDatasetsRequest.decode(value),
7124
+ responseSerialize: (value) => Buffer.from(exports.UpsertResponse.encode(value).finish()),
7125
+ responseDeserialize: (value) => exports.UpsertResponse.decode(value),
7126
+ },
7127
+ /** Gets pre-built user dataset records for a gravity task */
7128
+ getPreBuiltUserDatasets: {
7129
+ path: "/gravity.v1.GravityService/GetPreBuiltUserDatasets",
7130
+ requestStream: false,
7131
+ responseStream: false,
7132
+ requestSerialize: (value) => Buffer.from(exports.GetPreBuiltUserDatasetsRequest.encode(value).finish()),
7133
+ requestDeserialize: (value) => exports.GetPreBuiltUserDatasetsRequest.decode(value),
7134
+ responseSerialize: (value) => Buffer.from(exports.GetPreBuiltUserDatasetsResponse.encode(value).finish()),
7135
+ responseDeserialize: (value) => exports.GetPreBuiltUserDatasetsResponse.decode(value),
7136
+ },
6817
7137
  };
6818
7138
  exports.GravityServiceClient = (0, grpc_js_1.makeGenericClientConstructor)(exports.GravityServiceService, "gravity.v1.GravityService");
6819
7139
  function toTimestamp(date) {