@strapi/database 4.6.0-beta.0 → 4.6.0-beta.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/lib/entity-manager/regular-relations.js +55 -67
- package/lib/index.d.ts +5 -2
- package/lib/query/helpers/index.js +1 -0
- package/lib/query/helpers/streams/index.js +5 -0
- package/lib/query/helpers/streams/readable.js +174 -0
- package/lib/query/query-builder.js +11 -0
- package/lib/schema/index.d.ts +1 -1
- package/package.json +2 -2
|
@@ -198,60 +198,42 @@ const cleanOrderColumns = async ({ id, attribute, db, inverseRelIds, transaction
|
|
|
198
198
|
return;
|
|
199
199
|
}
|
|
200
200
|
|
|
201
|
-
// Handle databases that don't support window function ROW_NUMBER
|
|
202
|
-
if (!strapi.db.dialect.supportsWindowFunctions()) {
|
|
203
|
-
await cleanOrderColumnsForOldDatabases({ id, attribute, db, inverseRelIds, transaction: trx });
|
|
204
|
-
return;
|
|
205
|
-
}
|
|
206
|
-
|
|
207
|
-
const { joinTable } = attribute;
|
|
208
|
-
const { joinColumn, inverseJoinColumn, orderColumnName, inverseOrderColumnName } = joinTable;
|
|
209
|
-
const update = [];
|
|
210
|
-
const updateBinding = [];
|
|
211
|
-
const select = ['??'];
|
|
212
|
-
const selectBinding = ['id'];
|
|
213
|
-
const where = [];
|
|
214
|
-
const whereBinding = [];
|
|
215
|
-
|
|
216
|
-
if (hasOrderColumn(attribute) && id) {
|
|
217
|
-
update.push('?? = b.src_order');
|
|
218
|
-
updateBinding.push(orderColumnName);
|
|
219
|
-
select.push('ROW_NUMBER() OVER (PARTITION BY ?? ORDER BY ??) AS src_order');
|
|
220
|
-
selectBinding.push(joinColumn.name, orderColumnName);
|
|
221
|
-
where.push('?? = ?');
|
|
222
|
-
whereBinding.push(joinColumn.name, id);
|
|
223
|
-
}
|
|
224
|
-
|
|
225
|
-
if (hasInverseOrderColumn(attribute) && !isEmpty(inverseRelIds)) {
|
|
226
|
-
update.push('?? = b.inv_order');
|
|
227
|
-
updateBinding.push(inverseOrderColumnName);
|
|
228
|
-
select.push('ROW_NUMBER() OVER (PARTITION BY ?? ORDER BY ??) AS inv_order');
|
|
229
|
-
selectBinding.push(inverseJoinColumn.name, inverseOrderColumnName);
|
|
230
|
-
where.push(`?? IN (${inverseRelIds.map(() => '?').join(', ')})`);
|
|
231
|
-
whereBinding.push(inverseJoinColumn.name, ...inverseRelIds);
|
|
232
|
-
}
|
|
233
|
-
|
|
234
|
-
// raw query as knex doesn't allow updating from a subquery
|
|
235
|
-
// https://github.com/knex/knex/issues/2504
|
|
236
201
|
switch (strapi.db.dialect.client) {
|
|
237
202
|
case 'mysql':
|
|
238
|
-
await db
|
|
239
|
-
.raw(
|
|
240
|
-
`UPDATE
|
|
241
|
-
?? as a,
|
|
242
|
-
(
|
|
243
|
-
SELECT ${select.join(', ')}
|
|
244
|
-
FROM ??
|
|
245
|
-
WHERE ${where.join(' OR ')}
|
|
246
|
-
) AS b
|
|
247
|
-
SET ${update.join(', ')}
|
|
248
|
-
WHERE b.id = a.id`,
|
|
249
|
-
[joinTable.name, ...selectBinding, joinTable.name, ...whereBinding, ...updateBinding]
|
|
250
|
-
)
|
|
251
|
-
.transacting(trx);
|
|
203
|
+
await cleanOrderColumnsForInnoDB({ id, attribute, db, inverseRelIds, transaction: trx });
|
|
252
204
|
break;
|
|
253
205
|
default: {
|
|
206
|
+
const { joinTable } = attribute;
|
|
207
|
+
const { joinColumn, inverseJoinColumn, orderColumnName, inverseOrderColumnName } = joinTable;
|
|
208
|
+
const update = [];
|
|
209
|
+
const updateBinding = [];
|
|
210
|
+
const select = ['??'];
|
|
211
|
+
const selectBinding = ['id'];
|
|
212
|
+
const where = [];
|
|
213
|
+
const whereBinding = [];
|
|
214
|
+
|
|
215
|
+
if (hasOrderColumn(attribute) && id) {
|
|
216
|
+
update.push('?? = b.src_order');
|
|
217
|
+
updateBinding.push(orderColumnName);
|
|
218
|
+
select.push('ROW_NUMBER() OVER (PARTITION BY ?? ORDER BY ??) AS src_order');
|
|
219
|
+
selectBinding.push(joinColumn.name, orderColumnName);
|
|
220
|
+
where.push('?? = ?');
|
|
221
|
+
whereBinding.push(joinColumn.name, id);
|
|
222
|
+
}
|
|
223
|
+
|
|
224
|
+
if (hasInverseOrderColumn(attribute) && !isEmpty(inverseRelIds)) {
|
|
225
|
+
update.push('?? = b.inv_order');
|
|
226
|
+
updateBinding.push(inverseOrderColumnName);
|
|
227
|
+
select.push('ROW_NUMBER() OVER (PARTITION BY ?? ORDER BY ??) AS inv_order');
|
|
228
|
+
selectBinding.push(inverseJoinColumn.name, inverseOrderColumnName);
|
|
229
|
+
where.push(`?? IN (${inverseRelIds.map(() => '?').join(', ')})`);
|
|
230
|
+
whereBinding.push(inverseJoinColumn.name, ...inverseRelIds);
|
|
231
|
+
}
|
|
232
|
+
|
|
254
233
|
const joinTableName = addSchema(joinTable.name);
|
|
234
|
+
|
|
235
|
+
// raw query as knex doesn't allow updating from a subquery
|
|
236
|
+
// https://github.com/knex/knex/issues/2504
|
|
255
237
|
await db.connection
|
|
256
238
|
.raw(
|
|
257
239
|
`UPDATE ?? as a
|
|
@@ -265,24 +247,29 @@ const cleanOrderColumns = async ({ id, attribute, db, inverseRelIds, transaction
|
|
|
265
247
|
[joinTableName, ...updateBinding, ...selectBinding, joinTableName, ...whereBinding]
|
|
266
248
|
)
|
|
267
249
|
.transacting(trx);
|
|
250
|
+
|
|
251
|
+
/*
|
|
252
|
+
`UPDATE :joinTable: as a
|
|
253
|
+
SET :orderColumn: = b.src_order, :inverseOrderColumn: = b.inv_order
|
|
254
|
+
FROM (
|
|
255
|
+
SELECT
|
|
256
|
+
id,
|
|
257
|
+
ROW_NUMBER() OVER ( PARTITION BY :joinColumn: ORDER BY :orderColumn:) AS src_order,
|
|
258
|
+
ROW_NUMBER() OVER ( PARTITION BY :inverseJoinColumn: ORDER BY :inverseOrderColumn:) AS inv_order
|
|
259
|
+
FROM :joinTable:
|
|
260
|
+
WHERE :joinColumn: = :id OR :inverseJoinColumn: IN (:inverseRelIds)
|
|
261
|
+
) AS b
|
|
262
|
+
WHERE b.id = a.id`,
|
|
263
|
+
*/
|
|
268
264
|
}
|
|
269
|
-
/*
|
|
270
|
-
`UPDATE :joinTable: as a
|
|
271
|
-
SET :orderColumn: = b.src_order, :inverseOrderColumn: = b.inv_order
|
|
272
|
-
FROM (
|
|
273
|
-
SELECT
|
|
274
|
-
id,
|
|
275
|
-
ROW_NUMBER() OVER ( PARTITION BY :joinColumn: ORDER BY :orderColumn:) AS src_order,
|
|
276
|
-
ROW_NUMBER() OVER ( PARTITION BY :inverseJoinColumn: ORDER BY :inverseOrderColumn:) AS inv_order
|
|
277
|
-
FROM :joinTable:
|
|
278
|
-
WHERE :joinColumn: = :id OR :inverseJoinColumn: IN (:inverseRelIds)
|
|
279
|
-
) AS b
|
|
280
|
-
WHERE b.id = a.id`,
|
|
281
|
-
*/
|
|
282
265
|
}
|
|
283
266
|
};
|
|
284
267
|
|
|
285
|
-
|
|
268
|
+
/*
|
|
269
|
+
* Ensure that orders are following a 1, 2, 3 sequence, without gap.
|
|
270
|
+
* The use of a temporary table instead of a window function makes the query compatible with MySQL 5 and prevents some deadlocks to happen in innoDB databases
|
|
271
|
+
*/
|
|
272
|
+
const cleanOrderColumnsForInnoDB = async ({
|
|
286
273
|
id,
|
|
287
274
|
attribute,
|
|
288
275
|
db,
|
|
@@ -319,6 +306,9 @@ const cleanOrderColumnsForOldDatabases = async ({
|
|
|
319
306
|
}
|
|
320
307
|
)
|
|
321
308
|
.transacting(trx);
|
|
309
|
+
|
|
310
|
+
// raw query as knex doesn't allow updating from a subquery
|
|
311
|
+
// https://github.com/knex/knex/issues/2504
|
|
322
312
|
await db.connection
|
|
323
313
|
.raw(
|
|
324
314
|
`UPDATE ?? as a, (SELECT * FROM ??) AS b
|
|
@@ -338,7 +328,7 @@ const cleanOrderColumnsForOldDatabases = async ({
|
|
|
338
328
|
await db.connection
|
|
339
329
|
.raw(
|
|
340
330
|
`
|
|
341
|
-
CREATE
|
|
331
|
+
CREATE TABLE ??
|
|
342
332
|
SELECT
|
|
343
333
|
id,
|
|
344
334
|
(
|
|
@@ -371,9 +361,7 @@ const cleanOrderColumnsForOldDatabases = async ({
|
|
|
371
361
|
)
|
|
372
362
|
.transacting(trx);
|
|
373
363
|
} finally {
|
|
374
|
-
await db.connection
|
|
375
|
-
.raw(`DROP TEMPORARY TABLE IF EXISTS ??`, [tempInvOrderTableName])
|
|
376
|
-
.transacting(trx);
|
|
364
|
+
await db.connection.raw(`DROP TABLE IF EXISTS ??`, [tempInvOrderTableName]).transacting(trx);
|
|
377
365
|
}
|
|
378
366
|
}
|
|
379
367
|
};
|
package/lib/index.d.ts
CHANGED
|
@@ -1,3 +1,4 @@
|
|
|
1
|
+
import { Knex } from 'knex';
|
|
1
2
|
import { LifecycleProvider } from './lifecycles';
|
|
2
3
|
import { MigrationProvider } from './migrations';
|
|
3
4
|
import { SchemaProvider } from './schema';
|
|
@@ -31,8 +32,7 @@ type AttributeOperators<T, K extends keyof T> = {
|
|
|
31
32
|
|
|
32
33
|
export type WhereParams<T> = {
|
|
33
34
|
[K in keyof T]?: T[K] | T[K][] | AttributeOperators<T, K>;
|
|
34
|
-
} &
|
|
35
|
-
LogicalOperators<T>;
|
|
35
|
+
} & LogicalOperators<T>;
|
|
36
36
|
|
|
37
37
|
type Sortables<T> = {
|
|
38
38
|
// check sortable
|
|
@@ -158,6 +158,9 @@ export interface Database {
|
|
|
158
158
|
lifecycles: LifecycleProvider;
|
|
159
159
|
migrations: MigrationProvider;
|
|
160
160
|
entityManager: EntityManager;
|
|
161
|
+
queryBuilder: any;
|
|
162
|
+
metadata: any;
|
|
163
|
+
connection: Knex;
|
|
161
164
|
|
|
162
165
|
query<T extends keyof AllTypes>(uid: T): QueryFromContentType<T>;
|
|
163
166
|
}
|
|
@@ -0,0 +1,174 @@
|
|
|
1
|
+
'use strict';
|
|
2
|
+
|
|
3
|
+
const { Readable } = require('stream');
|
|
4
|
+
const { isFinite } = require('lodash/fp');
|
|
5
|
+
|
|
6
|
+
const { applyPopulate } = require('../populate');
|
|
7
|
+
const { fromRow } = require('../transform');
|
|
8
|
+
|
|
9
|
+
const knexQueryDone = Symbol('knexQueryDone');
|
|
10
|
+
const knexPerformingQuery = Symbol('knexPerformingQuery');
|
|
11
|
+
|
|
12
|
+
class ReadableStrapiQuery extends Readable {
|
|
13
|
+
/**
|
|
14
|
+
* @param {object} options
|
|
15
|
+
* @param {ReturnType<typeof import('../../query-builder')>} options.qb The strapi query builder instance
|
|
16
|
+
* @param {string} options.uid The model uid
|
|
17
|
+
* @param {import('../../../index').Database} options.db The Database instance
|
|
18
|
+
* @param {boolean} [options.mapResults] The maximum number of entities to fetch per query
|
|
19
|
+
* @param {number} [options.batchSize] The maximum number of entities to fetch per query
|
|
20
|
+
*/
|
|
21
|
+
constructor({ qb, db, uid, mapResults = true, batchSize = 500 }) {
|
|
22
|
+
super({ objectMode: true, highWaterMark: batchSize });
|
|
23
|
+
|
|
24
|
+
// Extract offset & limit from the query-builder's state
|
|
25
|
+
const { offset, limit } = qb.state;
|
|
26
|
+
|
|
27
|
+
// Original offset value
|
|
28
|
+
this._offset = isFinite(offset) ? offset : 0;
|
|
29
|
+
|
|
30
|
+
// Max amount of entities to fetch, force null as undefined value
|
|
31
|
+
this._limit = isFinite(limit) ? limit : null;
|
|
32
|
+
|
|
33
|
+
// Total amount of entities fetched
|
|
34
|
+
this._fetched = 0;
|
|
35
|
+
|
|
36
|
+
/**
|
|
37
|
+
* Original query
|
|
38
|
+
* @type {import('knex').Knex}
|
|
39
|
+
*/
|
|
40
|
+
this._query = qb.getKnexQuery();
|
|
41
|
+
|
|
42
|
+
// Query Builder instance
|
|
43
|
+
this._qb = qb;
|
|
44
|
+
|
|
45
|
+
// Database related properties
|
|
46
|
+
this._db = db;
|
|
47
|
+
this._uid = uid;
|
|
48
|
+
this._meta = db.metadata.get(uid);
|
|
49
|
+
|
|
50
|
+
// Stream params
|
|
51
|
+
this._batchSize = batchSize;
|
|
52
|
+
this._mapResults = mapResults;
|
|
53
|
+
|
|
54
|
+
// States
|
|
55
|
+
this[knexPerformingQuery] = false;
|
|
56
|
+
}
|
|
57
|
+
|
|
58
|
+
_destroy(err, cb) {
|
|
59
|
+
// If the stream is destroyed while a query is being made, then wait for a
|
|
60
|
+
// kQueryDone event to be emitted before actually destroying the stream
|
|
61
|
+
if (this[knexPerformingQuery]) {
|
|
62
|
+
this.once(knexQueryDone, (er) => cb(err || er));
|
|
63
|
+
} else {
|
|
64
|
+
cb(err);
|
|
65
|
+
}
|
|
66
|
+
}
|
|
67
|
+
|
|
68
|
+
/**
|
|
69
|
+
* Custom ._read() implementation
|
|
70
|
+
*
|
|
71
|
+
* NOTE: Here "size" means the number of entities to be read from the database.
|
|
72
|
+
* Not the actual byte size, as it would means that we need to return partial entities.
|
|
73
|
+
*
|
|
74
|
+
* @param {number} size
|
|
75
|
+
*/
|
|
76
|
+
async _read(size) {
|
|
77
|
+
const query = this._query;
|
|
78
|
+
|
|
79
|
+
// Remove the original offset & limit properties from the query
|
|
80
|
+
// Theoretically, they would be replaced by calling them again, but this is just to be sure
|
|
81
|
+
query.clear('limit').clear('offset');
|
|
82
|
+
|
|
83
|
+
// Define the maximum read size based on the limit and the requested size
|
|
84
|
+
// NOTE: size is equal to _batchSize by default. Since we want to allow customizing it on
|
|
85
|
+
// the fly, we need to use its value instead of batchSize when computing the maxReadSize value
|
|
86
|
+
const maxReadSize =
|
|
87
|
+
// if no limit is defined in the query, use the given size,
|
|
88
|
+
// otherwise, use the smallest value between the two
|
|
89
|
+
this._limit === null ? size : Math.min(size, this._limit);
|
|
90
|
+
|
|
91
|
+
// Compute the limit for the next query
|
|
92
|
+
const limit =
|
|
93
|
+
// If a limit is defined
|
|
94
|
+
this._limit !== null &&
|
|
95
|
+
// And reading `maxReadSize` would fetch too many entities (> _limit)
|
|
96
|
+
this._fetched + maxReadSize > this._limit
|
|
97
|
+
? // Then adjust the limit so that it only get the remaining entities
|
|
98
|
+
this._limit - this._fetched
|
|
99
|
+
: // Else, use the max read size
|
|
100
|
+
maxReadSize;
|
|
101
|
+
|
|
102
|
+
// If we don't have anything left to read (_limit === _fetched),
|
|
103
|
+
// don't bother making the query and end the stream by pushing null
|
|
104
|
+
if (limit <= 0) {
|
|
105
|
+
this.push(null);
|
|
106
|
+
return;
|
|
107
|
+
}
|
|
108
|
+
|
|
109
|
+
// Compute the offset (base offset + number of entities already fetched)
|
|
110
|
+
const offset = this._offset + this._fetched;
|
|
111
|
+
|
|
112
|
+
// Update the query with the new values (offset + limit)
|
|
113
|
+
query.offset(offset).limit(limit);
|
|
114
|
+
|
|
115
|
+
// Lock the ._destroy()
|
|
116
|
+
this[knexPerformingQuery] = true;
|
|
117
|
+
|
|
118
|
+
let results;
|
|
119
|
+
let count;
|
|
120
|
+
let err;
|
|
121
|
+
|
|
122
|
+
try {
|
|
123
|
+
// Execute the query and store the results & count
|
|
124
|
+
results = await query;
|
|
125
|
+
|
|
126
|
+
const { populate } = this._qb.state;
|
|
127
|
+
|
|
128
|
+
// Apply populate if needed
|
|
129
|
+
if (populate) {
|
|
130
|
+
await applyPopulate(results, populate, { qb: this._qb, uid: this._uid, db: this._db });
|
|
131
|
+
}
|
|
132
|
+
|
|
133
|
+
// Map results if asked to
|
|
134
|
+
if (this._mapResults) {
|
|
135
|
+
results = fromRow(this._meta, results);
|
|
136
|
+
}
|
|
137
|
+
|
|
138
|
+
count = results.length;
|
|
139
|
+
} catch (e) {
|
|
140
|
+
err = e;
|
|
141
|
+
}
|
|
142
|
+
|
|
143
|
+
// Unlock the ._destroy()
|
|
144
|
+
this[knexPerformingQuery] = false;
|
|
145
|
+
|
|
146
|
+
// Tell ._destroy() that it's now safe to close the db connection
|
|
147
|
+
if (this.destroyed) {
|
|
148
|
+
this.emit(knexQueryDone);
|
|
149
|
+
return;
|
|
150
|
+
}
|
|
151
|
+
|
|
152
|
+
// If there is an error, destroy with the given error
|
|
153
|
+
if (err) {
|
|
154
|
+
this.destroy(err);
|
|
155
|
+
return;
|
|
156
|
+
}
|
|
157
|
+
|
|
158
|
+
// Update the amount of fetched entities
|
|
159
|
+
this._fetched += count;
|
|
160
|
+
|
|
161
|
+
// While there is at least one value to unpack
|
|
162
|
+
for (const result of results) {
|
|
163
|
+
this.push(result);
|
|
164
|
+
}
|
|
165
|
+
|
|
166
|
+
// If the amount of fetched entities is smaller than the
|
|
167
|
+
// maximum read size, Then push null to close the stream
|
|
168
|
+
if (this._fetched === this._limit || count < this._batchSize) {
|
|
169
|
+
this.push(null);
|
|
170
|
+
}
|
|
171
|
+
}
|
|
172
|
+
}
|
|
173
|
+
|
|
174
|
+
module.exports = ReadableStrapiQuery;
|
|
@@ -2,6 +2,7 @@
|
|
|
2
2
|
|
|
3
3
|
const _ = require('lodash/fp');
|
|
4
4
|
|
|
5
|
+
const { DatabaseError } = require('../errors');
|
|
5
6
|
const helpers = require('./helpers');
|
|
6
7
|
|
|
7
8
|
const createQueryBuilder = (uid, db, initialState = {}) => {
|
|
@@ -488,6 +489,16 @@ const createQueryBuilder = (uid, db, initialState = {}) => {
|
|
|
488
489
|
db.dialect.transformErrors(error);
|
|
489
490
|
}
|
|
490
491
|
},
|
|
492
|
+
|
|
493
|
+
stream({ mapResults = true } = {}) {
|
|
494
|
+
if (state.type === 'select') {
|
|
495
|
+
return new helpers.ReadableQuery({ qb: this, db, uid, mapResults });
|
|
496
|
+
}
|
|
497
|
+
|
|
498
|
+
throw new DatabaseError(
|
|
499
|
+
`query-builder.stream() has been called with an unsupported query type: "${state.type}"`
|
|
500
|
+
);
|
|
501
|
+
},
|
|
491
502
|
};
|
|
492
503
|
};
|
|
493
504
|
|
package/lib/schema/index.d.ts
CHANGED
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@strapi/database",
|
|
3
|
-
"version": "4.6.0-beta.
|
|
3
|
+
"version": "4.6.0-beta.1",
|
|
4
4
|
"description": "Strapi's database layer",
|
|
5
5
|
"homepage": "https://strapi.io",
|
|
6
6
|
"bugs": {
|
|
@@ -43,5 +43,5 @@
|
|
|
43
43
|
"node": ">=14.19.1 <=18.x.x",
|
|
44
44
|
"npm": ">=6.0.0"
|
|
45
45
|
},
|
|
46
|
-
"gitHead": "
|
|
46
|
+
"gitHead": "2c0bcabdf0bf2a269fed50c6f23ba777845968a0"
|
|
47
47
|
}
|