@lde/pipeline 0.28.1 → 0.28.3
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +17 -4
- package/dist/sparql/executor.d.ts +44 -0
- package/dist/sparql/executor.d.ts.map +1 -1
- package/dist/sparql/executor.js +57 -1
- package/dist/sparql/index.d.ts +1 -1
- package/dist/sparql/index.d.ts.map +1 -1
- package/dist/sparql/index.js +1 -1
- package/dist/stage.d.ts +7 -1
- package/dist/stage.d.ts.map +1 -1
- package/dist/stage.js +12 -11
- package/package.json +1 -1
package/README.md
CHANGED
|
@@ -50,20 +50,24 @@ new Stage({
|
|
|
50
50
|
|
|
51
51
|
`batchSize` (default: 10) controls how many variable bindings are passed to each executor call as a `VALUES` clause. It also sets the page size for the item selector's SPARQL requests, so that each paginated request fills exactly one executor batch.
|
|
52
52
|
|
|
53
|
-
|
|
53
|
+
Some SPARQL endpoints enforce different result limits for SELECT and CONSTRUCT queries. Since the selector uses SELECT and the executor uses CONSTRUCT, a `LIMIT` clause in the selector query overrides `batchSize` as the page size. Use this when the endpoint caps SELECT results below your desired batch size:
|
|
54
54
|
|
|
55
55
|
```typescript
|
|
56
|
-
// Endpoint caps results at
|
|
56
|
+
// Endpoint caps SELECT results at 500, but each CONSTRUCT can handle 1000 bindings.
|
|
57
57
|
new Stage({
|
|
58
58
|
name: 'per-class',
|
|
59
59
|
itemSelector: new SparqlItemSelector({
|
|
60
|
-
query: 'SELECT DISTINCT ?class WHERE { ?s a ?class } LIMIT
|
|
60
|
+
query: 'SELECT DISTINCT ?class WHERE { ?s a ?class } LIMIT 500',
|
|
61
61
|
}),
|
|
62
62
|
executors: executor,
|
|
63
|
-
batchSize:
|
|
63
|
+
batchSize: 1000, // Two SELECT pages fill one CONSTRUCT batch.
|
|
64
64
|
});
|
|
65
65
|
```
|
|
66
66
|
|
|
67
|
+
#### Concurrency
|
|
68
|
+
|
|
69
|
+
`maxConcurrency` (default: 10) controls how many batches run in parallel. Within each batch, all executors run in parallel too, so the peak number of concurrent SPARQL queries is `maxConcurrency × executorCount`. For example, with `maxConcurrency: 5` and two executors per stage, up to 10 SPARQL queries may be in flight at the same time. Lower `maxConcurrency` if the endpoint can't handle the load.
|
|
70
|
+
|
|
67
71
|
### Item Selector
|
|
68
72
|
|
|
69
73
|
Selects resources from the distribution and fans out executor calls per batch of results. Implements the `ItemSelector` interface:
|
|
@@ -103,6 +107,15 @@ const executor = new SparqlConstructExecutor({
|
|
|
103
107
|
});
|
|
104
108
|
```
|
|
105
109
|
|
|
110
|
+
When querying endpoints that return line-oriented formats like N-Triples (e.g. QLever), enable `lineBuffer` to work around an [N3.js chunk-splitting bug](https://github.com/rdfjs/N3.js/issues/578) that causes intermittent parse errors on large responses:
|
|
111
|
+
|
|
112
|
+
```typescript
|
|
113
|
+
const executor = new SparqlConstructExecutor({
|
|
114
|
+
query: 'CONSTRUCT { ?s ?p ?o } WHERE { ?s ?p ?o }',
|
|
115
|
+
lineBuffer: true,
|
|
116
|
+
});
|
|
117
|
+
```
|
|
118
|
+
|
|
106
119
|
`Executor` is an interface, so you can implement your own for logic that's hard to express in pure SPARQL — for example, cleaning up messy date notations or converting locale-specific dates to ISO 8601. The decorator pattern lets you wrap a SPARQL executor and post-process its quad stream in TypeScript:
|
|
107
120
|
|
|
108
121
|
```typescript
|
|
@@ -1,6 +1,7 @@
|
|
|
1
1
|
import { Dataset, Distribution } from '@lde/dataset';
|
|
2
2
|
import { SparqlEndpointFetcher } from 'fetch-sparql-endpoint';
|
|
3
3
|
import type { NamedNode, Quad } from '@rdfjs/types';
|
|
4
|
+
import { Transform } from 'node:stream';
|
|
4
5
|
/**
|
|
5
6
|
* An executor could not run because the dataset lacks a supported distribution.
|
|
6
7
|
*/
|
|
@@ -42,6 +43,17 @@ export interface SparqlConstructExecutorOptions {
|
|
|
42
43
|
* Optional custom SparqlEndpointFetcher instance.
|
|
43
44
|
*/
|
|
44
45
|
fetcher?: SparqlEndpointFetcher;
|
|
46
|
+
/**
|
|
47
|
+
* Buffer complete lines before passing them to the N3 parser.
|
|
48
|
+
*
|
|
49
|
+
* Works around an [N3.js bug](https://github.com/rdfjs/N3.js/issues/578)
|
|
50
|
+
* where language tags (e.g. `@nl-nl`) split across HTTP chunk boundaries
|
|
51
|
+
* cause parse errors. Enable this when querying endpoints that return
|
|
52
|
+
* line-oriented formats such as N-Triples (e.g. QLever).
|
|
53
|
+
*
|
|
54
|
+
* @default false
|
|
55
|
+
*/
|
|
56
|
+
lineBuffer?: boolean;
|
|
45
57
|
}
|
|
46
58
|
/**
|
|
47
59
|
* A streaming SPARQL CONSTRUCT executor.
|
|
@@ -75,6 +87,7 @@ export declare class SparqlConstructExecutor implements Executor {
|
|
|
75
87
|
private readonly preParsed?;
|
|
76
88
|
private readonly fetcher;
|
|
77
89
|
private readonly retries;
|
|
90
|
+
private readonly lineBuffer;
|
|
78
91
|
private readonly generator;
|
|
79
92
|
constructor(options: SparqlConstructExecutorOptions);
|
|
80
93
|
/**
|
|
@@ -86,6 +99,12 @@ export declare class SparqlConstructExecutor implements Executor {
|
|
|
86
99
|
* @returns AsyncIterable<Quad> stream of results.
|
|
87
100
|
*/
|
|
88
101
|
execute(dataset: Dataset, distribution: Distribution, options?: ExecuteOptions): Promise<AsyncIterable<Quad>>;
|
|
102
|
+
/**
|
|
103
|
+
* Fetch quads from the endpoint, optionally line-buffering the response
|
|
104
|
+
* stream before it reaches the N3 parser to work around
|
|
105
|
+
* {@link https://github.com/rdfjs/N3.js/issues/578 | N3.js#578}.
|
|
106
|
+
*/
|
|
107
|
+
private fetchQuads;
|
|
89
108
|
/**
|
|
90
109
|
* Create an executor from a query file.
|
|
91
110
|
*
|
|
@@ -98,4 +117,29 @@ export declare class SparqlConstructExecutor implements Executor {
|
|
|
98
117
|
* Read a SPARQL query from a file.
|
|
99
118
|
*/
|
|
100
119
|
export declare function readQueryFile(filename: string): Promise<string>;
|
|
120
|
+
/**
|
|
121
|
+
* Buffers incoming data until complete lines (`\n`-terminated) are available,
|
|
122
|
+
* then pushes them downstream as a single chunk.
|
|
123
|
+
*
|
|
124
|
+
* **Why this exists:** `fetch-sparql-endpoint` pipes the raw HTTP response
|
|
125
|
+
* stream directly into N3.js's `StreamParser`. N3.js has a bug
|
|
126
|
+
* ({@link https://github.com/rdfjs/N3.js/issues/578 | N3.js#578}) where
|
|
127
|
+
* tokens that straddle chunk boundaries — most commonly language tags like
|
|
128
|
+
* `@nl-nl` — cause spurious `Unexpected "-nl"` parse errors. The error is
|
|
129
|
+
* non-deterministic and typically surfaces only on responses larger than
|
|
130
|
+
* ~12 MB, because that is when HTTP chunking starts splitting mid-token.
|
|
131
|
+
*
|
|
132
|
+
* By ensuring each chunk passed to the parser ends on a line boundary, we
|
|
133
|
+
* prevent any N-Triples token from being split. Memory overhead is minimal:
|
|
134
|
+
* at most one partial line is buffered at a time.
|
|
135
|
+
*
|
|
136
|
+
* This transform can be removed once N3.js#578 is fixed upstream.
|
|
137
|
+
*
|
|
138
|
+
* @see https://github.com/rdfjs/N3.js/issues/578
|
|
139
|
+
*/
|
|
140
|
+
export declare class LineBufferTransform extends Transform {
|
|
141
|
+
private remainder;
|
|
142
|
+
_transform(chunk: Buffer, _encoding: string, callback: () => void): void;
|
|
143
|
+
_flush(callback: () => void): void;
|
|
144
|
+
}
|
|
101
145
|
//# sourceMappingURL=executor.d.ts.map
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"executor.d.ts","sourceRoot":"","sources":["../../src/sparql/executor.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,OAAO,EAAE,YAAY,EAAiB,MAAM,cAAc,CAAC;AACpE,OAAO,EAAE,qBAAqB,EAAE,MAAM,uBAAuB,CAAC;AAC9D,OAAO,KAAK,EAAE,SAAS,EAAE,IAAI,EAAE,MAAM,cAAc,CAAC;
|
|
1
|
+
{"version":3,"file":"executor.d.ts","sourceRoot":"","sources":["../../src/sparql/executor.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,OAAO,EAAE,YAAY,EAAiB,MAAM,cAAc,CAAC;AACpE,OAAO,EAAE,qBAAqB,EAAE,MAAM,uBAAuB,CAAC;AAC9D,OAAO,KAAK,EAAE,SAAS,EAAE,IAAI,EAAE,MAAM,cAAc,CAAC;AAGpD,OAAO,EAAE,SAAS,EAAE,MAAM,aAAa,CAAC;AAUxC;;GAEG;AACH,qBAAa,YAAY;aACK,OAAO,EAAE,MAAM;gBAAf,OAAO,EAAE,MAAM;CAC5C;AAED,qEAAqE;AACrE,MAAM,MAAM,gBAAgB,GAAG,MAAM,CAAC,MAAM,EAAE,SAAS,CAAC,CAAC;AAEzD,MAAM,WAAW,cAAc;IAC7B;;;OAGG;IACH,QAAQ,CAAC,EAAE,gBAAgB,EAAE,CAAC;CAC/B;AAED,MAAM,WAAW,QAAQ;IACvB,OAAO,CACL,OAAO,EAAE,OAAO,EAChB,YAAY,EAAE,YAAY,EAC1B,OAAO,CAAC,EAAE,cAAc,GACvB,OAAO,CAAC,aAAa,CAAC,IAAI,CAAC,GAAG,YAAY,CAAC,CAAC;CAChD;AAED;;GAEG;AACH,MAAM,WAAW,8BAA8B;IAC7C;;OAEG;IACH,KAAK,EAAE,MAAM,CAAC;IAEd;;;OAGG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;IAEjB;;;OAGG;IACH,OAAO,CAAC,EAAE,MAAM,CAAC;IAEjB;;OAEG;IACH,OAAO,CAAC,EAAE,qBAAqB,CAAC;IAEhC;;;;;;;;;OASG;IACH,UAAU,CAAC,EAAE,OAAO,CAAC;CACtB;AAED;;;;;;;;;;;;;;;;;;;;;;;;;;GA0BG;AACH,qBAAa,uBAAwB,YAAW,QAAQ;IACtD,OAAO,CAAC,QAAQ,CAAC,QAAQ,CAAS;IAClC,OAAO,CAAC,QAAQ,CAAC,SAAS,CAAC,CAAiB;IAC5C,OAAO,CAAC,QAAQ,CAAC,OAAO,CAAwB;IAChD,OAAO,CAAC,QAAQ,CAAC,OAAO,CAAS;IACjC,OAAO,CAAC,QAAQ,CAAC,UAAU,CAAU;IACrC,OAAO,CAAC,QAAQ,CAAC,SAAS,CAAmB;gBAEjC,OAAO,EAAE,8BAA8B;IAoBnD;;;;;;;OAOG;IACG,OAAO,CACX,OAAO,EAAE,OAAO,EAChB,YAAY,EAAE,YAAY,EAC1B,OAAO,CAAC,EAAE,cAAc,GACvB,OAAO,CAAC,aAAa,CAAC,IAAI,CAAC,CAAC;IAwC/B;;;;OAIG;YACW,UAAU;IAmBxB;;;;;OAKG;WACiB,QAAQ,CAC1B,QAAQ,EAAE,MAAM,EAChB,OAAO,CAAC,EAAE,IAAI,CAAC,8BAA8B,EAAE,OAAO,CAAC,GACtD,OAAO,CAAC,uBAAuB,CAAC;CAIpC;AAED;;GAEG;AACH,wBAAsB,aAAa,CAAC,QAAQ,EAAE,MAAM,GAAG,OAAO,CAAC,MAAM,CAAC,CAErE;AAED;;;;;;;;;;;;;;;;;;;GAmBG;AACH,qBAAa,mBAAoB,SAAQ,SAAS;IAChD,OAAO,CAAC,SAAS,CAAM;IAEd,UAAU,CACjB,KAAK,EAAE,MAAM,EACb,SAAS,EAAE,MAAM,EACjB,QAAQ,EAAE,MAAM,IAAI;IAWb,MAAM,CAAC,QAAQ,EAAE,MAAM,IAAI;CAMrC"}
|
package/dist/sparql/executor.js
CHANGED
|
@@ -2,6 +2,8 @@ import { assertSafeIri } from '@lde/dataset';
|
|
|
2
2
|
import { SparqlEndpointFetcher } from 'fetch-sparql-endpoint';
|
|
3
3
|
import { readFile } from 'node:fs/promises';
|
|
4
4
|
import { resolve } from 'node:path';
|
|
5
|
+
import { Transform } from 'node:stream';
|
|
6
|
+
import { StreamParser } from 'n3';
|
|
5
7
|
import { Parser } from '@traqula/parser-sparql-1-1';
|
|
6
8
|
import { Generator } from '@traqula/generator-sparql-1-1';
|
|
7
9
|
import isNetworkError from 'is-network-error';
|
|
@@ -49,10 +51,12 @@ export class SparqlConstructExecutor {
|
|
|
49
51
|
preParsed;
|
|
50
52
|
fetcher;
|
|
51
53
|
retries;
|
|
54
|
+
lineBuffer;
|
|
52
55
|
generator = new Generator();
|
|
53
56
|
constructor(options) {
|
|
54
57
|
this.rawQuery = options.query;
|
|
55
58
|
this.retries = options.retries ?? 3;
|
|
59
|
+
this.lineBuffer = options.lineBuffer ?? false;
|
|
56
60
|
if (!options.query.includes('#subjectFilter#')) {
|
|
57
61
|
const parsed = new Parser().parse(options.query);
|
|
58
62
|
if (parsed.type !== 'query' || parsed.subType !== 'construct') {
|
|
@@ -98,11 +102,25 @@ export class SparqlConstructExecutor {
|
|
|
98
102
|
let query = this.generator.generate(ast);
|
|
99
103
|
assertSafeIri(dataset.iri.toString());
|
|
100
104
|
query = query.replaceAll('?dataset', `<${dataset.iri}>`);
|
|
101
|
-
return await pRetry(() => this.
|
|
105
|
+
return await pRetry(() => this.fetchQuads(endpoint.toString(), query), {
|
|
102
106
|
retries: this.retries,
|
|
103
107
|
shouldRetry: ({ error }) => isTransientError(error),
|
|
104
108
|
});
|
|
105
109
|
}
|
|
110
|
+
/**
|
|
111
|
+
* Fetch quads from the endpoint, optionally line-buffering the response
|
|
112
|
+
* stream before it reaches the N3 parser to work around
|
|
113
|
+
* {@link https://github.com/rdfjs/N3.js/issues/578 | N3.js#578}.
|
|
114
|
+
*/
|
|
115
|
+
async fetchQuads(endpoint, query) {
|
|
116
|
+
if (!this.lineBuffer) {
|
|
117
|
+
return this.fetcher.fetchTriples(endpoint, query);
|
|
118
|
+
}
|
|
119
|
+
const [contentType, , responseStream] = await this.fetcher.fetchRawStream(endpoint, query, SparqlEndpointFetcher.CONTENTTYPE_TURTLE);
|
|
120
|
+
return responseStream
|
|
121
|
+
.pipe(new LineBufferTransform())
|
|
122
|
+
.pipe(new StreamParser({ format: contentType }));
|
|
123
|
+
}
|
|
106
124
|
/**
|
|
107
125
|
* Create an executor from a query file.
|
|
108
126
|
*
|
|
@@ -120,6 +138,44 @@ export class SparqlConstructExecutor {
|
|
|
120
138
|
export async function readQueryFile(filename) {
|
|
121
139
|
return (await readFile(resolve(filename))).toString();
|
|
122
140
|
}
|
|
141
|
+
/**
|
|
142
|
+
* Buffers incoming data until complete lines (`\n`-terminated) are available,
|
|
143
|
+
* then pushes them downstream as a single chunk.
|
|
144
|
+
*
|
|
145
|
+
* **Why this exists:** `fetch-sparql-endpoint` pipes the raw HTTP response
|
|
146
|
+
* stream directly into N3.js's `StreamParser`. N3.js has a bug
|
|
147
|
+
* ({@link https://github.com/rdfjs/N3.js/issues/578 | N3.js#578}) where
|
|
148
|
+
* tokens that straddle chunk boundaries — most commonly language tags like
|
|
149
|
+
* `@nl-nl` — cause spurious `Unexpected "-nl"` parse errors. The error is
|
|
150
|
+
* non-deterministic and typically surfaces only on responses larger than
|
|
151
|
+
* ~12 MB, because that is when HTTP chunking starts splitting mid-token.
|
|
152
|
+
*
|
|
153
|
+
* By ensuring each chunk passed to the parser ends on a line boundary, we
|
|
154
|
+
* prevent any N-Triples token from being split. Memory overhead is minimal:
|
|
155
|
+
* at most one partial line is buffered at a time.
|
|
156
|
+
*
|
|
157
|
+
* This transform can be removed once N3.js#578 is fixed upstream.
|
|
158
|
+
*
|
|
159
|
+
* @see https://github.com/rdfjs/N3.js/issues/578
|
|
160
|
+
*/
|
|
161
|
+
export class LineBufferTransform extends Transform {
|
|
162
|
+
remainder = '';
|
|
163
|
+
_transform(chunk, _encoding, callback) {
|
|
164
|
+
const data = this.remainder + chunk.toString();
|
|
165
|
+
const lines = data.split('\n');
|
|
166
|
+
this.remainder = lines.pop() ?? '';
|
|
167
|
+
if (lines.length > 0) {
|
|
168
|
+
this.push(lines.join('\n') + '\n');
|
|
169
|
+
}
|
|
170
|
+
callback();
|
|
171
|
+
}
|
|
172
|
+
_flush(callback) {
|
|
173
|
+
if (this.remainder.length > 0) {
|
|
174
|
+
this.push(this.remainder);
|
|
175
|
+
}
|
|
176
|
+
callback();
|
|
177
|
+
}
|
|
178
|
+
}
|
|
123
179
|
const transientStatusPattern = /HTTP status (\d+)/;
|
|
124
180
|
function isTransientError(error) {
|
|
125
181
|
if (isNetworkError(error))
|
package/dist/sparql/index.d.ts
CHANGED
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
export { SparqlConstructExecutor, NotSupported, readQueryFile, type ExecuteOptions, type Executor, type SparqlConstructExecutorOptions, type VariableBindings, } from './executor.js';
|
|
1
|
+
export { SparqlConstructExecutor, LineBufferTransform, NotSupported, readQueryFile, type ExecuteOptions, type Executor, type SparqlConstructExecutorOptions, type VariableBindings, } from './executor.js';
|
|
2
2
|
export { SparqlItemSelector, type SparqlItemSelectorOptions, } from './selector.js';
|
|
3
3
|
export { injectValues } from './values.js';
|
|
4
4
|
export { withDefaultGraph } from './graph.js';
|
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/sparql/index.ts"],"names":[],"mappings":"AAAA,OAAO,EACL,uBAAuB,EACvB,YAAY,EACZ,aAAa,EACb,KAAK,cAAc,EACnB,KAAK,QAAQ,EACb,KAAK,8BAA8B,EACnC,KAAK,gBAAgB,GACtB,MAAM,eAAe,CAAC;AACvB,OAAO,EACL,kBAAkB,EAClB,KAAK,yBAAyB,GAC/B,MAAM,eAAe,CAAC;AAEvB,OAAO,EAAE,YAAY,EAAE,MAAM,aAAa,CAAC;AAE3C,OAAO,EAAE,gBAAgB,EAAE,MAAM,YAAY,CAAC"}
|
|
1
|
+
{"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/sparql/index.ts"],"names":[],"mappings":"AAAA,OAAO,EACL,uBAAuB,EACvB,mBAAmB,EACnB,YAAY,EACZ,aAAa,EACb,KAAK,cAAc,EACnB,KAAK,QAAQ,EACb,KAAK,8BAA8B,EACnC,KAAK,gBAAgB,GACtB,MAAM,eAAe,CAAC;AACvB,OAAO,EACL,kBAAkB,EAClB,KAAK,yBAAyB,GAC/B,MAAM,eAAe,CAAC;AAEvB,OAAO,EAAE,YAAY,EAAE,MAAM,aAAa,CAAC;AAE3C,OAAO,EAAE,gBAAgB,EAAE,MAAM,YAAY,CAAC"}
|
package/dist/sparql/index.js
CHANGED
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
export { SparqlConstructExecutor, NotSupported, readQueryFile, } from './executor.js';
|
|
1
|
+
export { SparqlConstructExecutor, LineBufferTransform, NotSupported, readQueryFile, } from './executor.js';
|
|
2
2
|
export { SparqlItemSelector, } from './selector.js';
|
|
3
3
|
export { injectValues } from './values.js';
|
|
4
4
|
export { withDefaultGraph } from './graph.js';
|
package/dist/stage.d.ts
CHANGED
|
@@ -20,7 +20,13 @@ export interface StageOptions {
|
|
|
20
20
|
* @default 10
|
|
21
21
|
*/
|
|
22
22
|
batchSize?: number;
|
|
23
|
-
/**
|
|
23
|
+
/**
|
|
24
|
+
* Maximum concurrent in-flight batches. Within each batch, all executors
|
|
25
|
+
* run in parallel, so the peak number of concurrent SPARQL queries is
|
|
26
|
+
* `maxConcurrency × executorCount`.
|
|
27
|
+
*
|
|
28
|
+
* @default 10
|
|
29
|
+
*/
|
|
24
30
|
maxConcurrency?: number;
|
|
25
31
|
/** Child stages that chain off this stage's output. */
|
|
26
32
|
stages?: Stage[];
|
package/dist/stage.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"stage.d.ts","sourceRoot":"","sources":["../src/stage.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,OAAO,EAAE,YAAY,EAAE,MAAM,cAAc,CAAC;AACrD,OAAO,KAAK,EAAE,IAAI,EAAE,MAAM,cAAc,CAAC;AACzC,OAAO,KAAK,EAAE,QAAQ,EAAE,gBAAgB,EAAE,MAAM,sBAAsB,CAAC;AACvE,OAAO,EAAE,YAAY,EAAE,MAAM,sBAAsB,CAAC;AAEpD,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,gBAAgB,CAAC;AAChD,OAAO,KAAK,EAAE,MAAM,EAAE,MAAM,oBAAoB,CAAC;AAGjD,mEAAmE;AACnE,MAAM,MAAM,aAAa,GAAG,CAC1B,KAAK,EAAE,aAAa,CAAC,IAAI,CAAC,EAC1B,OAAO,EAAE,OAAO,KACb,aAAa,CAAC,IAAI,CAAC,CAAC;AAEzB,MAAM,WAAW,YAAY;IAC3B,IAAI,EAAE,MAAM,CAAC;IACb,SAAS,EAAE,QAAQ,GAAG,QAAQ,EAAE,CAAC;IACjC,YAAY,CAAC,EAAE,YAAY,CAAC;IAC5B;;;;;;;;OAQG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB
|
|
1
|
+
{"version":3,"file":"stage.d.ts","sourceRoot":"","sources":["../src/stage.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,OAAO,EAAE,YAAY,EAAE,MAAM,cAAc,CAAC;AACrD,OAAO,KAAK,EAAE,IAAI,EAAE,MAAM,cAAc,CAAC;AACzC,OAAO,KAAK,EAAE,QAAQ,EAAE,gBAAgB,EAAE,MAAM,sBAAsB,CAAC;AACvE,OAAO,EAAE,YAAY,EAAE,MAAM,sBAAsB,CAAC;AAEpD,OAAO,KAAK,EAAE,SAAS,EAAE,MAAM,gBAAgB,CAAC;AAChD,OAAO,KAAK,EAAE,MAAM,EAAE,MAAM,oBAAoB,CAAC;AAGjD,mEAAmE;AACnE,MAAM,MAAM,aAAa,GAAG,CAC1B,KAAK,EAAE,aAAa,CAAC,IAAI,CAAC,EAC1B,OAAO,EAAE,OAAO,KACb,aAAa,CAAC,IAAI,CAAC,CAAC;AAEzB,MAAM,WAAW,YAAY;IAC3B,IAAI,EAAE,MAAM,CAAC;IACb,SAAS,EAAE,QAAQ,GAAG,QAAQ,EAAE,CAAC;IACjC,YAAY,CAAC,EAAE,YAAY,CAAC;IAC5B;;;;;;;;OAQG;IACH,SAAS,CAAC,EAAE,MAAM,CAAC;IACnB;;;;;;OAMG;IACH,cAAc,CAAC,EAAE,MAAM,CAAC;IACxB,uDAAuD;IACvD,MAAM,CAAC,EAAE,KAAK,EAAE,CAAC;IACjB,qFAAqF;IACrF,UAAU,CAAC,EAAE;QACX,SAAS,EAAE,SAAS,CAAC;QACrB,iEAAiE;QACjE,SAAS,CAAC,EAAE,OAAO,GAAG,MAAM,GAAG,MAAM,CAAC;KACvC,CAAC;CACH;AAED,MAAM,WAAW,UAAU;IACzB,UAAU,CAAC,EAAE,CAAC,cAAc,EAAE,MAAM,EAAE,cAAc,EAAE,MAAM,KAAK,IAAI,CAAC;CACvE;AAED,qBAAa,KAAK;IAChB,QAAQ,CAAC,IAAI,EAAE,MAAM,CAAC;IACtB,QAAQ,CAAC,MAAM,EAAE,SAAS,KAAK,EAAE,CAAC;IAClC,OAAO,CAAC,QAAQ,CAAC,SAAS,CAAa;IACvC,OAAO,CAAC,QAAQ,CAAC,YAAY,CAAC,CAAe;IAC7C,OAAO,CAAC,QAAQ,CAAC,SAAS,CAAS;IACnC,OAAO,CAAC,QAAQ,CAAC,cAAc,CAAS;IACxC,OAAO,CAAC,QAAQ,CAAC,UAAU,CAAC,CAA6B;gBAE7C,OAAO,EAAE,YAAY;IAYjC,mDAAmD;IACnD,IAAI,SAAS,IAAI,SAAS,GAAG,SAAS,CAErC;IAEK,GAAG,CACP,OAAO,EAAE,OAAO,EAChB,YAAY,EAAE,YAAY,EAC1B,MAAM,EAAE,MAAM,EACd,OAAO,CAAC,EAAE,UAAU,GACnB,OAAO,CAAC,YAAY,GAAG,IAAI,CAAC;YAkDjB,eAAe;IAiJ7B;;;OAGG;YACW,cAAc;YAqBd,UAAU;CAqBzB;AAUD,6GAA6G;AAC7G,MAAM,WAAW,YAAY;IAC3B,MAAM,CACJ,YAAY,EAAE,YAAY,EAC1B,SAAS,CAAC,EAAE,MAAM,GACjB,aAAa,CAAC,gBAAgB,CAAC,CAAC;CACpC"}
|
package/dist/stage.js
CHANGED
|
@@ -110,18 +110,19 @@ export class Stage {
|
|
|
110
110
|
break;
|
|
111
111
|
}
|
|
112
112
|
track((async () => {
|
|
113
|
-
|
|
114
|
-
|
|
115
|
-
const result = await executor.execute(dataset, distribution, {
|
|
116
|
-
|
|
117
|
-
|
|
118
|
-
|
|
119
|
-
|
|
120
|
-
|
|
121
|
-
|
|
122
|
-
}
|
|
113
|
+
// Run all executors for this batch in parallel.
|
|
114
|
+
const executorOutputs = await Promise.all(this.executors.map(async (executor) => {
|
|
115
|
+
const result = await executor.execute(dataset, distribution, { bindings });
|
|
116
|
+
if (result instanceof NotSupported)
|
|
117
|
+
return [];
|
|
118
|
+
hasResults = true;
|
|
119
|
+
const quads = [];
|
|
120
|
+
for await (const quad of result) {
|
|
121
|
+
quads.push(quad);
|
|
123
122
|
}
|
|
124
|
-
|
|
123
|
+
return quads;
|
|
124
|
+
}));
|
|
125
|
+
const batchQuads = executorOutputs.flat();
|
|
125
126
|
if (this.validation &&
|
|
126
127
|
batchQuads.length > 0 &&
|
|
127
128
|
onInvalid !== 'write') {
|