macrocosmos 1.2.3 → 1.2.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
|
@@ -156,6 +156,14 @@ export interface BuildDatasetResponse {
|
|
|
156
156
|
/** dataset: the dataset that was built */
|
|
157
157
|
dataset?: Dataset | undefined;
|
|
158
158
|
}
|
|
159
|
+
export interface Nebula {
|
|
160
|
+
/** error: nebula build error message */
|
|
161
|
+
error: string;
|
|
162
|
+
/** file_size_bytes: the size of the file in bytes */
|
|
163
|
+
fileSizeBytes: number;
|
|
164
|
+
/** url: the URL of the file */
|
|
165
|
+
url: string;
|
|
166
|
+
}
|
|
159
167
|
/** Dataset contains the progress and results of a dataset build */
|
|
160
168
|
export interface Dataset {
|
|
161
169
|
/** crawler_workflow_id: the ID of the parent crawler for this dataset */
|
|
@@ -174,6 +182,8 @@ export interface Dataset {
|
|
|
174
182
|
steps: DatasetStep[];
|
|
175
183
|
/** total_steps: the total number of steps in the dataset build */
|
|
176
184
|
totalSteps: number;
|
|
185
|
+
/** nebula: the details about the nebula that was built */
|
|
186
|
+
nebula?: Nebula | undefined;
|
|
177
187
|
}
|
|
178
188
|
/** DatasetFile contains the details about a dataset file */
|
|
179
189
|
export interface DatasetFile {
|
|
@@ -248,6 +258,7 @@ export declare const CreateGravityTaskRequest: MessageFns<CreateGravityTaskReque
|
|
|
248
258
|
export declare const CreateGravityTaskResponse: MessageFns<CreateGravityTaskResponse>;
|
|
249
259
|
export declare const BuildDatasetRequest: MessageFns<BuildDatasetRequest>;
|
|
250
260
|
export declare const BuildDatasetResponse: MessageFns<BuildDatasetResponse>;
|
|
261
|
+
export declare const Nebula: MessageFns<Nebula>;
|
|
251
262
|
export declare const Dataset: MessageFns<Dataset>;
|
|
252
263
|
export declare const DatasetFile: MessageFns<DatasetFile>;
|
|
253
264
|
export declare const DatasetStep: MessageFns<DatasetStep>;
|
|
@@ -5,7 +5,7 @@
|
|
|
5
5
|
// protoc v3.20.3
|
|
6
6
|
// source: gravity/v1/gravity.proto
|
|
7
7
|
Object.defineProperty(exports, "__esModule", { value: true });
|
|
8
|
-
exports.GravityServiceClient = exports.GravityServiceService = exports.CancelDatasetResponse = exports.CancelDatasetRequest = exports.CancelGravityTaskResponse = exports.CancelGravityTaskRequest = exports.GetDatasetResponse = exports.GetDatasetRequest = exports.DatasetStep = exports.DatasetFile = exports.Dataset = exports.BuildDatasetResponse = exports.BuildDatasetRequest = exports.CreateGravityTaskResponse = exports.CreateGravityTaskRequest = exports.GetCrawlerResponse = exports.GetCrawlerRequest = exports.NotificationRequest = exports.GravityTask = exports.GetGravityTasksResponse = exports.GetGravityTasksRequest = exports.GravityTaskState = exports.CrawlerState = exports.HfRepo = exports.CrawlerNotification = exports.CrawlerCriteria = exports.Crawler = exports.protobufPackage = void 0;
|
|
8
|
+
exports.GravityServiceClient = exports.GravityServiceService = exports.CancelDatasetResponse = exports.CancelDatasetRequest = exports.CancelGravityTaskResponse = exports.CancelGravityTaskRequest = exports.GetDatasetResponse = exports.GetDatasetRequest = exports.DatasetStep = exports.DatasetFile = exports.Dataset = exports.Nebula = exports.BuildDatasetResponse = exports.BuildDatasetRequest = exports.CreateGravityTaskResponse = exports.CreateGravityTaskRequest = exports.GetCrawlerResponse = exports.GetCrawlerRequest = exports.NotificationRequest = exports.GravityTask = exports.GetGravityTasksResponse = exports.GetGravityTasksRequest = exports.GravityTaskState = exports.CrawlerState = exports.HfRepo = exports.CrawlerNotification = exports.CrawlerCriteria = exports.Crawler = exports.protobufPackage = void 0;
|
|
9
9
|
/* eslint-disable */
|
|
10
10
|
const wire_1 = require("@bufbuild/protobuf/wire");
|
|
11
11
|
const grpc_js_1 = require("@grpc/grpc-js");
|
|
@@ -1415,6 +1415,91 @@ exports.BuildDatasetResponse = {
|
|
|
1415
1415
|
return message;
|
|
1416
1416
|
},
|
|
1417
1417
|
};
|
|
1418
|
+
function createBaseNebula() {
|
|
1419
|
+
return { error: "", fileSizeBytes: 0, url: "" };
|
|
1420
|
+
}
|
|
1421
|
+
exports.Nebula = {
|
|
1422
|
+
encode(message, writer = new wire_1.BinaryWriter()) {
|
|
1423
|
+
if (message.error !== "") {
|
|
1424
|
+
writer.uint32(10).string(message.error);
|
|
1425
|
+
}
|
|
1426
|
+
if (message.fileSizeBytes !== 0) {
|
|
1427
|
+
writer.uint32(16).int64(message.fileSizeBytes);
|
|
1428
|
+
}
|
|
1429
|
+
if (message.url !== "") {
|
|
1430
|
+
writer.uint32(26).string(message.url);
|
|
1431
|
+
}
|
|
1432
|
+
return writer;
|
|
1433
|
+
},
|
|
1434
|
+
decode(input, length) {
|
|
1435
|
+
const reader = input instanceof wire_1.BinaryReader ? input : new wire_1.BinaryReader(input);
|
|
1436
|
+
let end = length === undefined ? reader.len : reader.pos + length;
|
|
1437
|
+
const message = createBaseNebula();
|
|
1438
|
+
while (reader.pos < end) {
|
|
1439
|
+
const tag = reader.uint32();
|
|
1440
|
+
switch (tag >>> 3) {
|
|
1441
|
+
case 1: {
|
|
1442
|
+
if (tag !== 10) {
|
|
1443
|
+
break;
|
|
1444
|
+
}
|
|
1445
|
+
message.error = reader.string();
|
|
1446
|
+
continue;
|
|
1447
|
+
}
|
|
1448
|
+
case 2: {
|
|
1449
|
+
if (tag !== 16) {
|
|
1450
|
+
break;
|
|
1451
|
+
}
|
|
1452
|
+
message.fileSizeBytes = longToNumber(reader.int64());
|
|
1453
|
+
continue;
|
|
1454
|
+
}
|
|
1455
|
+
case 3: {
|
|
1456
|
+
if (tag !== 26) {
|
|
1457
|
+
break;
|
|
1458
|
+
}
|
|
1459
|
+
message.url = reader.string();
|
|
1460
|
+
continue;
|
|
1461
|
+
}
|
|
1462
|
+
}
|
|
1463
|
+
if ((tag & 7) === 4 || tag === 0) {
|
|
1464
|
+
break;
|
|
1465
|
+
}
|
|
1466
|
+
reader.skip(tag & 7);
|
|
1467
|
+
}
|
|
1468
|
+
return message;
|
|
1469
|
+
},
|
|
1470
|
+
fromJSON(object) {
|
|
1471
|
+
return {
|
|
1472
|
+
error: isSet(object.error) ? globalThis.String(object.error) : "",
|
|
1473
|
+
fileSizeBytes: isSet(object.fileSizeBytes)
|
|
1474
|
+
? globalThis.Number(object.fileSizeBytes)
|
|
1475
|
+
: 0,
|
|
1476
|
+
url: isSet(object.url) ? globalThis.String(object.url) : "",
|
|
1477
|
+
};
|
|
1478
|
+
},
|
|
1479
|
+
toJSON(message) {
|
|
1480
|
+
const obj = {};
|
|
1481
|
+
if (message.error !== "") {
|
|
1482
|
+
obj.error = message.error;
|
|
1483
|
+
}
|
|
1484
|
+
if (message.fileSizeBytes !== 0) {
|
|
1485
|
+
obj.fileSizeBytes = Math.round(message.fileSizeBytes);
|
|
1486
|
+
}
|
|
1487
|
+
if (message.url !== "") {
|
|
1488
|
+
obj.url = message.url;
|
|
1489
|
+
}
|
|
1490
|
+
return obj;
|
|
1491
|
+
},
|
|
1492
|
+
create(base) {
|
|
1493
|
+
return exports.Nebula.fromPartial(base ?? {});
|
|
1494
|
+
},
|
|
1495
|
+
fromPartial(object) {
|
|
1496
|
+
const message = createBaseNebula();
|
|
1497
|
+
message.error = object.error ?? "";
|
|
1498
|
+
message.fileSizeBytes = object.fileSizeBytes ?? 0;
|
|
1499
|
+
message.url = object.url ?? "";
|
|
1500
|
+
return message;
|
|
1501
|
+
},
|
|
1502
|
+
};
|
|
1418
1503
|
function createBaseDataset() {
|
|
1419
1504
|
return {
|
|
1420
1505
|
crawlerWorkflowId: "",
|
|
@@ -1425,6 +1510,7 @@ function createBaseDataset() {
|
|
|
1425
1510
|
statusMessage: "",
|
|
1426
1511
|
steps: [],
|
|
1427
1512
|
totalSteps: 0,
|
|
1513
|
+
nebula: undefined,
|
|
1428
1514
|
};
|
|
1429
1515
|
}
|
|
1430
1516
|
exports.Dataset = {
|
|
@@ -1453,6 +1539,9 @@ exports.Dataset = {
|
|
|
1453
1539
|
if (message.totalSteps !== 0) {
|
|
1454
1540
|
writer.uint32(64).int64(message.totalSteps);
|
|
1455
1541
|
}
|
|
1542
|
+
if (message.nebula !== undefined) {
|
|
1543
|
+
exports.Nebula.encode(message.nebula, writer.uint32(74).fork()).join();
|
|
1544
|
+
}
|
|
1456
1545
|
return writer;
|
|
1457
1546
|
},
|
|
1458
1547
|
decode(input, length) {
|
|
@@ -1518,6 +1607,13 @@ exports.Dataset = {
|
|
|
1518
1607
|
message.totalSteps = longToNumber(reader.int64());
|
|
1519
1608
|
continue;
|
|
1520
1609
|
}
|
|
1610
|
+
case 9: {
|
|
1611
|
+
if (tag !== 74) {
|
|
1612
|
+
break;
|
|
1613
|
+
}
|
|
1614
|
+
message.nebula = exports.Nebula.decode(reader, reader.uint32());
|
|
1615
|
+
continue;
|
|
1616
|
+
}
|
|
1521
1617
|
}
|
|
1522
1618
|
if ((tag & 7) === 4 || tag === 0) {
|
|
1523
1619
|
break;
|
|
@@ -1550,6 +1646,7 @@ exports.Dataset = {
|
|
|
1550
1646
|
totalSteps: isSet(object.totalSteps)
|
|
1551
1647
|
? globalThis.Number(object.totalSteps)
|
|
1552
1648
|
: 0,
|
|
1649
|
+
nebula: isSet(object.nebula) ? exports.Nebula.fromJSON(object.nebula) : undefined,
|
|
1553
1650
|
};
|
|
1554
1651
|
},
|
|
1555
1652
|
toJSON(message) {
|
|
@@ -1578,6 +1675,9 @@ exports.Dataset = {
|
|
|
1578
1675
|
if (message.totalSteps !== 0) {
|
|
1579
1676
|
obj.totalSteps = Math.round(message.totalSteps);
|
|
1580
1677
|
}
|
|
1678
|
+
if (message.nebula !== undefined) {
|
|
1679
|
+
obj.nebula = exports.Nebula.toJSON(message.nebula);
|
|
1680
|
+
}
|
|
1581
1681
|
return obj;
|
|
1582
1682
|
},
|
|
1583
1683
|
create(base) {
|
|
@@ -1593,6 +1693,10 @@ exports.Dataset = {
|
|
|
1593
1693
|
message.statusMessage = object.statusMessage ?? "";
|
|
1594
1694
|
message.steps = object.steps?.map(e => exports.DatasetStep.fromPartial(e)) || [];
|
|
1595
1695
|
message.totalSteps = object.totalSteps ?? 0;
|
|
1696
|
+
message.nebula =
|
|
1697
|
+
object.nebula !== undefined && object.nebula !== null
|
|
1698
|
+
? exports.Nebula.fromPartial(object.nebula)
|
|
1699
|
+
: undefined;
|
|
1596
1700
|
return message;
|
|
1597
1701
|
},
|
|
1598
1702
|
};
|
|
@@ -1,9 +1,9 @@
|
|
|
1
|
-
import { GravityServiceClient, GetGravityTasksRequest, GetGravityTasksResponse, GetCrawlerRequest, GetCrawlerResponse, CreateGravityTaskRequest as GeneratedCreateGravityTaskRequest, CreateGravityTaskResponse, BuildDatasetRequest as GeneratedBuildDatasetRequest, BuildDatasetResponse, GetDatasetRequest, GetDatasetResponse, CancelGravityTaskRequest, CancelGravityTaskResponse, CancelDatasetRequest, CancelDatasetResponse, Crawler, CrawlerCriteria, CrawlerNotification, HfRepo, CrawlerState, GravityTaskState, NotificationRequest } from "../../generated/gravity/v1/gravity";
|
|
1
|
+
import { GravityServiceClient, GetGravityTasksRequest, GetGravityTasksResponse, GetCrawlerRequest, GetCrawlerResponse, CreateGravityTaskRequest as GeneratedCreateGravityTaskRequest, CreateGravityTaskResponse, BuildDatasetRequest as GeneratedBuildDatasetRequest, BuildDatasetResponse, GetDatasetRequest, GetDatasetResponse, CancelGravityTaskRequest, CancelGravityTaskResponse, CancelDatasetRequest, CancelDatasetResponse, Crawler, CrawlerCriteria, CrawlerNotification, HfRepo, CrawlerState, GravityTaskState, NotificationRequest, Dataset, DatasetFile, DatasetStep } from "../../generated/gravity/v1/gravity";
|
|
2
2
|
import { BaseClient, BaseClientOptions } from "../BaseClient";
|
|
3
3
|
import { MarkFieldsOptional } from "../util.types";
|
|
4
4
|
type CreateGravityTaskRequest = MarkFieldsOptional<GeneratedCreateGravityTaskRequest, "notificationRequests">;
|
|
5
5
|
type BuildDatasetRequest = MarkFieldsOptional<GeneratedBuildDatasetRequest, "notificationRequests">;
|
|
6
|
-
export type { GetGravityTasksRequest, GetGravityTasksResponse, GetCrawlerRequest, GetCrawlerResponse, CreateGravityTaskRequest, CreateGravityTaskResponse, BuildDatasetRequest, BuildDatasetResponse, GetDatasetRequest, GetDatasetResponse, CancelGravityTaskRequest, CancelGravityTaskResponse, CancelDatasetRequest, CancelDatasetResponse, Crawler, CrawlerCriteria, CrawlerNotification, HfRepo, CrawlerState, GravityTaskState, NotificationRequest, };
|
|
6
|
+
export type { GetGravityTasksRequest, GetGravityTasksResponse, GetCrawlerRequest, GetCrawlerResponse, CreateGravityTaskRequest, CreateGravityTaskResponse, BuildDatasetRequest, BuildDatasetResponse, GetDatasetRequest, GetDatasetResponse, CancelGravityTaskRequest, CancelGravityTaskResponse, CancelDatasetRequest, CancelDatasetResponse, Crawler, CrawlerCriteria, CrawlerNotification, HfRepo, CrawlerState, GravityTaskState, NotificationRequest, Dataset, DatasetFile, DatasetStep, };
|
|
7
7
|
/**
|
|
8
8
|
* Client for interacting with the Gravity API
|
|
9
9
|
* Provides gRPC interface for data collection and dataset management
|
package/new_version.txt
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
1.2.
|
|
1
|
+
1.2.4
|
package/old_version.txt
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
1.2.
|
|
1
|
+
1.2.3
|