@datapos/datapos-development 0.3.132 β†’ 0.3.134

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md CHANGED
@@ -1,7 +1,7 @@
1
1
  # Data Positioning Development Library
2
2
 
3
3
  [![Quality Gate Status](https://sonarcloud.io/api/project_badges/measure?project=data-positioning_datapos-development&metric=alert_status)](https://sonarcloud.io/summary/new_code?id=data-positioning_datapos-development)
4
- <span><!-- OWASP_BADGE_START -->[![OWASP](https://img.shields.io/badge/OWASP-passed-ForestGreen)](https://data-positioning.github.io/test-lib/dependency-check-reports/dependency-check-report.html)<!-- OWASP_BADGE_END --></span>
4
+ <span><!-- OWASP_BADGE_START -->[![OWASP](https://img.shields.io/badge/OWASP-passed-4CAF50)](https://data-positioning.github.io/test-lib/dependency-check-reports/dependency-check-report.html)<!-- OWASP_BADGE_END --></span>
5
5
  [![npm version](https://img.shields.io/npm/v/@datapos/datapos-development.svg)](https://www.npmjs.com/package/@datapos/datapos-development)
6
6
  [![License: MIT](https://img.shields.io/badge/License-MIT-blue.svg)](./LICENSE)
7
7
 
@@ -62,7 +62,7 @@ The OWASP Dependency Check Report identifies known vulnerabilities in project de
62
62
 
63
63
  ### Dependency Licenses
64
64
 
65
- The following table lists top-level production and peer dependencies only. All dependency licenses (including transitive dependencies) have been recursively verified to conform to Apache-2.0, CC0-1.0, MIT, or n/a. Developers cloning this repository should independently verify dev and optional dependencies; users of the uploaded library are covered by these checks.
65
+ The following table lists top-level production and peer dependencies. All these dependencies (including transitive ones) have been recursively verified to use Apache-2.0, CC0-1.0, or MITβ€”commercially friendly licenses with minimal restrictions. Developers cloning this repository should independently verify dev and optional dependencies; users of the published library are covered by these checks.
66
66
 
67
67
  <!-- DEPENDENCY_LICENSES_START -->
68
68
 
@@ -1,63 +1,63 @@
1
1
  import { exec as b } from "node:child_process";
2
- import { promises as t } from "node:fs";
2
+ import { promises as n } from "node:fs";
3
3
  import { nanoid as $ } from "nanoid";
4
4
  import { promisify as O } from "node:util";
5
5
  const S = ["createObject", "dropObject", "removeRecords", "upsertRecords"], v = ["findObject", "getRecord", "listNodes", "previewObject", "retrieveRecords"], y = O(b);
6
6
  async function j() {
7
7
  try {
8
8
  console.info("πŸš€ Building configuration...");
9
- const e = JSON.parse(await t.readFile("package.json", "utf8")), o = JSON.parse(await t.readFile("config.json", "utf8"));
10
- e.name != null && (o.id = e.name.replace("@datapos/", "").replace("@data-positioning/", "")), e.version != null && (o.version = e.version), await t.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Configuration built.");
9
+ const e = JSON.parse(await n.readFile("package.json", "utf8")), o = JSON.parse(await n.readFile("config.json", "utf8"));
10
+ e.name != null && (o.id = e.name.replace("@datapos/", "").replace("@data-positioning/", "")), e.version != null && (o.version = e.version), await n.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Configuration built.");
11
11
  } catch (e) {
12
12
  console.error("❌ Error building configuration.", e);
13
13
  }
14
14
  }
15
- async function k(e) {
15
+ async function J(e) {
16
16
  try {
17
17
  console.info(`πŸš€ Building public directory index for identifier '${e}'...`);
18
18
  const o = {};
19
19
  async function i(s, r) {
20
20
  console.info(`βš™οΈ Processing directory '${s}'...`);
21
- const d = [], c = s.substring(`public/${e}`.length);
22
- o[c] = d;
21
+ const l = [], c = s.substring(`public/${e}`.length);
22
+ o[c] = l;
23
23
  for (const a of r) {
24
- const l = `${s}/${a}`;
24
+ const d = `${s}/${a}`;
25
25
  try {
26
- const f = await t.stat(l);
26
+ const f = await n.stat(d);
27
27
  if (f.isDirectory()) {
28
- const u = await t.readdir(l), p = { childCount: u.length, name: `${a}`, typeId: "folder" };
29
- d.push(p), await i(l, u);
28
+ const u = await n.readdir(d), p = { childCount: u.length, name: `${a}`, typeId: "folder" };
29
+ l.push(p), await i(d, u);
30
30
  } else {
31
31
  const u = { id: $(), lastModifiedAt: f.mtimeMs, name: a, size: f.size, typeId: "object" };
32
- d.push(u);
32
+ l.push(u);
33
33
  }
34
34
  } catch (f) {
35
35
  throw new Error(`Unable to get information for '${a}' in 'buildPublicDirectoryIndex'. ${String(f)}`);
36
36
  }
37
37
  }
38
- d.sort((a, l) => {
39
- const f = a.typeId.localeCompare(l.typeId);
40
- return f === 0 ? a.name.localeCompare(l.name) : f;
38
+ l.sort((a, d) => {
39
+ const f = a.typeId.localeCompare(d.typeId);
40
+ return f === 0 ? a.name.localeCompare(d.name) : f;
41
41
  });
42
42
  }
43
- const n = await t.readdir(`public/${e}`);
44
- await i(`public/${e}`, n), await t.writeFile(`./public/${e}Index.json`, JSON.stringify(o), "utf8"), console.info("βœ… Public directory index built.");
43
+ const t = await n.readdir(`public/${e}`);
44
+ await i(`public/${e}`, t), await n.writeFile(`./public/${e}Index.json`, JSON.stringify(o), "utf8"), console.info("βœ… Public directory index built.");
45
45
  } catch (o) {
46
46
  console.error("❌ Error building public directory index.", o);
47
47
  }
48
48
  }
49
- async function J() {
49
+ async function k() {
50
50
  try {
51
51
  console.info("πŸš€ Building connector configuration...");
52
- const e = JSON.parse(await t.readFile("package.json", "utf8")), o = JSON.parse(await t.readFile("config.json", "utf8")), i = await t.readFile("src/index.ts", "utf8");
53
- let n = !1, s = !1;
54
- const r = /^\s{4}(?:async\s+)?(private\s+)?(?:public\s+|protected\s+)?([A-Za-z_]\w*)\s*\(/gm, d = [...i.matchAll(r)].filter((a) => a[1] == null && a[2] !== "constructor").map((a) => {
55
- const l = a[2];
56
- return n = n || S.includes(l), s = s || v.includes(l), l;
52
+ const e = JSON.parse(await n.readFile("package.json", "utf8")), o = JSON.parse(await n.readFile("config.json", "utf8")), i = await n.readFile("src/index.ts", "utf8");
53
+ let t = !1, s = !1;
54
+ const r = /^\s{4}(?:async\s+)?(private\s+)?(?:public\s+|protected\s+)?([A-Za-z_]\w*)\s*\(/gm, l = [...i.matchAll(r)].filter((a) => a[1] == null && a[2] !== "constructor").map((a) => {
55
+ const d = a[2];
56
+ return t = t || S.includes(d), s = s || v.includes(d), d;
57
57
  });
58
- d.length > 0 ? console.info(`ℹ️ Implements ${d.length} operations.`) : console.warn("⚠️ Implements no operations.");
59
- const c = s && n ? "bidirectional" : s ? "source" : n ? "destination" : "unknown";
60
- c && console.info(`ℹ️ Supports ${c} usage.`), e.name != null && (o.id = e.name), o.operations = d, o.usageId = c, e.version != null && (o.version = e.version), await t.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Connector configuration built.");
58
+ l.length > 0 ? console.info(`ℹ️ Implements ${l.length} operations.`) : console.warn("⚠️ Implements no operations.");
59
+ const c = s && t ? "bidirectional" : s ? "source" : t ? "destination" : "unknown";
60
+ c && console.info(`ℹ️ Supports ${c} usage.`), e.name != null && (o.id = e.name), o.operations = l, o.usageId = c, e.version != null && (o.version = e.version), await n.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Connector configuration built.");
61
61
  } catch (e) {
62
62
  console.error("❌ Error building connector configuration.", e);
63
63
  }
@@ -65,8 +65,8 @@ async function J() {
65
65
  async function F() {
66
66
  try {
67
67
  console.info("πŸš€ Building context configuration...");
68
- const e = JSON.parse(await t.readFile("package.json", "utf8")), o = JSON.parse(await t.readFile("config.json", "utf8")), i = await t.readFile("src/index.ts", "utf8"), n = /^\s{4}(?:async\s+)?(private\s+)?(?:public\s+|protected\s+)?([A-Za-z_]\w*)\s*\(/gm, s = [...i.matchAll(n)].filter((r) => r[1] == null && r[2] !== "constructor").map((r) => r[2]);
69
- e.name != null && (o.id = e.name), o.operations = s, e.version != null && (o.version = e.version), await t.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Context configuration built.");
68
+ const e = JSON.parse(await n.readFile("package.json", "utf8")), o = JSON.parse(await n.readFile("config.json", "utf8")), i = await n.readFile("src/index.ts", "utf8"), t = /^\s{4}(?:async\s+)?(private\s+)?(?:public\s+|protected\s+)?([A-Za-z_]\w*)\s*\(/gm, s = [...i.matchAll(t)].filter((r) => r[1] == null && r[2] !== "constructor").map((r) => r[2]);
69
+ e.name != null && (o.id = e.name), o.operations = s, e.version != null && (o.version = e.version), await n.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Context configuration built.");
70
70
  } catch (e) {
71
71
  console.error("❌ Error building context configuration.", e);
72
72
  }
@@ -74,8 +74,8 @@ async function F() {
74
74
  async function R() {
75
75
  try {
76
76
  console.info("πŸš€ Building presenter configuration...");
77
- const e = JSON.parse(await t.readFile("package.json", "utf8")), o = JSON.parse(await t.readFile("config.json", "utf8")), i = await t.readFile("src/index.ts", "utf8"), n = /^\s{4}(?:async\s+)?(private\s+)?(?:public\s+|protected\s+)?([A-Za-z_]\w*)\s*\(/gm, s = [...i.matchAll(n)].filter((r) => !r[1] && r[2] !== "constructor").map((r) => r[2]);
78
- e.name != null && (o.id = e.name), o.operations = s, e.version != null && (o.version = e.version), await t.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Presenter configuration built.");
77
+ const e = JSON.parse(await n.readFile("package.json", "utf8")), o = JSON.parse(await n.readFile("config.json", "utf8")), i = await n.readFile("src/index.ts", "utf8"), t = /^\s{4}(?:async\s+)?(private\s+)?(?:public\s+|protected\s+)?([A-Za-z_]\w*)\s*\(/gm, s = [...i.matchAll(t)].filter((r) => !r[1] && r[2] !== "constructor").map((r) => r[2]);
78
+ e.name != null && (o.id = e.name), o.operations = s, e.version != null && (o.version = e.version), await n.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Presenter configuration built.");
79
79
  } catch (e) {
80
80
  console.error("❌ Error building context configuration.", e);
81
81
  }
@@ -83,12 +83,12 @@ async function R() {
83
83
  async function A(e = "./") {
84
84
  try {
85
85
  console.info("πŸš€ Bumping version...");
86
- const o = JSON.parse(await t.readFile(`${e}package.json`, "utf8"));
86
+ const o = JSON.parse(await n.readFile(`${e}package.json`, "utf8"));
87
87
  if (o.version == null)
88
- o.version = "0.0.001", await t.writeFile(`${e}package.json`, JSON.stringify(o, void 0, 4), "utf8"), console.warn(`⚠️ Version initialised to ${o.version}.`);
88
+ o.version = "0.0.001", await n.writeFile(`${e}package.json`, JSON.stringify(o, void 0, 4), "utf8"), console.warn(`⚠️ Version initialised to ${o.version}.`);
89
89
  else {
90
- const i = o.version, n = o.version.split(".");
91
- o.version = `${n[0]}.${n[1]}.${Number(n[2]) + 1}`, await t.writeFile(`${e}package.json`, JSON.stringify(o, void 0, 4), "utf8"), console.info(`βœ… Version bumped from ${i} to ${o.version}.`);
90
+ const i = o.version, t = o.version.split(".");
91
+ o.version = `${t[0]}.${t[1]}.${Number(t[2]) + 1}`, await n.writeFile(`${e}package.json`, JSON.stringify(o, void 0, 4), "utf8"), console.info(`βœ… Version bumped from ${i} to ${o.version}.`);
92
92
  }
93
93
  } catch (o) {
94
94
  console.error("❌ Error bumping package version.", o);
@@ -100,10 +100,12 @@ function D(e) {
100
100
  async function P() {
101
101
  const e = "<!-- DEPENDENCY_LICENSES_START -->", o = "<!-- DEPENDENCY_LICENSES_END -->";
102
102
  try {
103
- const i = (await t.readFile("./licenses.md", "utf8")).trim(), n = await t.readFile("./README.md", "utf8"), s = n.indexOf(e), r = n.indexOf(o);
103
+ const i = (await n.readFile("./licenses.md", "utf8")).trim(), t = await n.readFile("./README.md", "utf8"), s = t.indexOf(e), r = t.indexOf(o);
104
104
  (s === -1 || r === -1) && (console.error("❌ Dependency license markers not found in readme file."), process.exit(1));
105
- const d = n.substring(0, s + e.length) + i + n.substring(r);
106
- await t.writeFile("README.md", d, "utf8"), console.log("βœ… Readme file updated with license information");
105
+ const l = t.substring(0, s + e.length) + `
106
+ ` + i + `
107
+ ` + t.substring(r);
108
+ await n.writeFile("README.md", l, "utf8"), console.log("βœ… Readme file updated with license information");
107
109
  } catch (i) {
108
110
  console.error("❌ Error updating readme file.", i), process.exit(1);
109
111
  }
@@ -111,12 +113,12 @@ async function P() {
111
113
  async function I() {
112
114
  const e = "<!-- OWASP_BADGE_START -->", o = "<!-- OWASP_BADGE_END -->";
113
115
  try {
114
- const i = JSON.parse(await t.readFile("./dependency-check-reports/dependency-check-report.json", "utf-8")), n = { critical: 0, high: 0, moderate: 0, low: 0, unknown: 0 };
116
+ const i = JSON.parse(await n.readFile("./dependency-check-reports/dependency-check-report.json", "utf-8")), t = { critical: 0, high: 0, moderate: 0, low: 0, unknown: 0 };
115
117
  for (const g of i.dependencies)
116
118
  if (g.vulnerabilities != null)
117
119
  for (const m of g.vulnerabilities) {
118
120
  const w = m.severity?.toLowerCase() ?? "unknown";
119
- w in n ? n[w]++ : n.unknown++;
121
+ w in t ? t[w]++ : t.unknown++;
120
122
  }
121
123
  const s = {
122
124
  critical: { color: "D32F2F", label: "critical" },
@@ -124,23 +126,20 @@ async function I() {
124
126
  moderate: { color: "FBC02D", label: "moderate" },
125
127
  low: { color: "6D8C31", label: "low" },
126
128
  unknown: { color: "616161", label: "unknown" }
127
- }, r = Object.values(n).reduce((g, m) => g + m, 0);
128
- console.info(`βœ… Total vulnerabilities found: ${r}`), console.info(
129
- ` Critical: ${n.critical}, High: ${n.high}, Moderate: ${n.moderate}, Low: ${n.low}, Unknown: ${n.unknown}`
130
- );
131
- const d = JSON.parse(await t.readFile("config.json", "utf8")), c = [];
132
- if (r === 0)
133
- c.push(`[![OWASP](https://img.shields.io/badge/OWASP-passed-4CAF50)](https://data-positioning.github.io/${d.id}/dependency-check-reports/dependency-check-report.html)`);
129
+ }, r = JSON.parse(await n.readFile("config.json", "utf8")), l = [];
130
+ if (Object.values(t).reduce((g, m) => g + m, 0) === 0)
131
+ console.info("βœ… No vulnerabilities found."), l.push(`[![OWASP](https://img.shields.io/badge/OWASP-passed-4CAF50)](https://data-positioning.github.io/${r.id}/dependency-check-reports/dependency-check-report.html)`);
134
132
  else
135
- for (const [g, m] of Object.entries(n)) {
136
- if (m === 0) continue;
137
- const w = s[g], h = `https://img.shields.io/badge/OWASP-${m}%20${w.label}-${w.color}`;
138
- c.push(`[![OWASP](${h})](https://data-positioning.github.io/${d.id}/dependency-check-reports/dependency-check-report.html)`);
133
+ for (const [g, m] of Object.entries(t)) {
134
+ const w = s[g];
135
+ if (console.warn(`⚠️ ${m} ${w.label} vulnerability(ies) found.`), m === 0) continue;
136
+ const h = `https://img.shields.io/badge/OWASP-${m}%20${w.label}-${w.color}`;
137
+ l.push(`[![OWASP](${h})](https://data-positioning.github.io/${r.id}/dependency-check-reports/dependency-check-report.html)`);
139
138
  }
140
- const a = await t.readFile("./README.md", "utf8"), l = a.indexOf(e), f = a.indexOf(o);
141
- (l === -1 || f === -1) && (console.error("❌ OWASP badge markers not found in README.md."), process.exit(1));
142
- const u = c.join(" "), p = a.substring(0, l + e.length) + u + a.substring(f);
143
- await t.writeFile("README.md", p, "utf8"), console.info("βœ… OWASP dependency check badges inserted into README.md");
139
+ const a = await n.readFile("./README.md", "utf8"), d = a.indexOf(e), f = a.indexOf(o);
140
+ (d === -1 || f === -1) && (console.error("❌ OWASP badge markers not found in README.md."), process.exit(1));
141
+ const u = l.join(" "), p = a.substring(0, d + e.length) + u + a.substring(f);
142
+ await n.writeFile("README.md", p, "utf8"), console.info("βœ… OWASP dependency check badge(s) inserted into README.md");
144
143
  } catch (i) {
145
144
  console.error("❌ Error updating README with OWASP badges:", i), process.exit(1);
146
145
  }
@@ -148,7 +147,7 @@ async function I() {
148
147
  async function M() {
149
148
  try {
150
149
  console.info("πŸš€ Sending deployment notice...");
151
- const e = JSON.parse(await t.readFile("config.json", "utf8")), o = {
150
+ const e = JSON.parse(await n.readFile("config.json", "utf8")), o = {
152
151
  body: JSON.stringify(e),
153
152
  headers: { "Content-Type": "application/json" },
154
153
  method: "PUT"
@@ -162,7 +161,7 @@ async function M() {
162
161
  async function T() {
163
162
  try {
164
163
  console.info("πŸš€ Synchronising with GitHub....");
165
- const e = JSON.parse(await t.readFile("package.json", "utf8"));
164
+ const e = JSON.parse(await n.readFile("package.json", "utf8"));
166
165
  await y("git add ."), await y(`git commit -m "v${e.version}"`), await y("git push origin main:main"), console.info(`βœ… Synchronised version ${e.version} with GitHub.`);
167
166
  } catch (e) {
168
167
  console.error("❌ Error synchronising with GitHub.", e);
@@ -171,12 +170,12 @@ async function T() {
171
170
  async function _(e, o) {
172
171
  try {
173
172
  console.info("πŸš€ Uploading directory to R2....");
174
- async function i(s, r, d) {
175
- for (const c of d) {
176
- const a = `${s}/${c}`, l = `${r}/${c}`;
177
- if ((await t.stat(a)).isDirectory()) {
178
- const u = await t.readdir(a);
179
- await i(a, l, u);
173
+ async function i(s, r, l) {
174
+ for (const c of l) {
175
+ const a = `${s}/${c}`, d = `${r}/${c}`;
176
+ if ((await n.stat(a)).isDirectory()) {
177
+ const u = await n.readdir(a);
178
+ await i(a, d, u);
180
179
  } else {
181
180
  console.info(`βš™οΈ Uploading '${s}/${c}'...`);
182
181
  const u = `wrangler r2 object put "datapos-sample-data-eu/${r}/${c}" --file="${s}/${c}" --jurisdiction=eu --remote`, p = await y(u);
@@ -184,8 +183,8 @@ async function _(e, o) {
184
183
  }
185
184
  }
186
185
  }
187
- const n = await t.readdir(`${e}/${o}/`);
188
- await i(`${e}/${o}`, o, n), console.info("βœ… Directory uploaded to R2.");
186
+ const t = await n.readdir(`${e}/${o}/`);
187
+ await i(`${e}/${o}`, o, t), console.info("βœ… Directory uploaded to R2.");
189
188
  } catch (i) {
190
189
  console.error("❌ Error uploading directory to R2.", i);
191
190
  }
@@ -193,44 +192,44 @@ async function _(e, o) {
193
192
  async function W() {
194
193
  try {
195
194
  console.info("πŸš€ Uploading module configuration....");
196
- const e = JSON.parse(await t.readFile("config.json", "utf8")), o = e.id, i = {
195
+ const e = JSON.parse(await n.readFile("config.json", "utf8")), o = e.id, i = {
197
196
  body: JSON.stringify(e),
198
197
  headers: { "Content-Type": "application/json" },
199
198
  method: "PUT"
200
- }, n = await fetch(`https://api.datapos.app/states/${o}`, i);
201
- if (!n.ok) throw new Error(await n.text());
199
+ }, t = await fetch(`https://api.datapos.app/states/${o}`, i);
200
+ if (!t.ok) throw new Error(await t.text());
202
201
  console.info("βœ… Module configuration uploaded.");
203
202
  } catch (e) {
204
203
  console.error("❌ Error uploading module configuration.", e);
205
204
  }
206
205
  }
207
- async function U(e) {
206
+ async function B(e) {
208
207
  try {
209
208
  console.info("πŸš€ Uploading module to R2...");
210
- const i = `v${JSON.parse(await t.readFile("package.json", "utf8")).version}`;
211
- async function n(s, r = "") {
212
- const d = await t.readdir(s, { withFileTypes: !0 });
213
- for (const c of d) {
214
- const a = `${s}/${c.name}`, l = r ? `${r}/${c.name}` : c.name;
209
+ const i = `v${JSON.parse(await n.readFile("package.json", "utf8")).version}`;
210
+ async function t(s, r = "") {
211
+ const l = await n.readdir(s, { withFileTypes: !0 });
212
+ for (const c of l) {
213
+ const a = `${s}/${c.name}`, d = r ? `${r}/${c.name}` : c.name;
215
214
  if (!c.isDirectory()) {
216
- const f = `${e}_${i}/${l}`.replace(/\\/g, "/"), u = c.name.endsWith(".js") ? "application/javascript" : c.name.endsWith(".css") ? "text/css" : "application/octet-stream";
217
- console.info(`βš™οΈ Uploading '${l}' β†’ '${f}'...`);
215
+ const f = `${e}_${i}/${d}`.replace(/\\/g, "/"), u = c.name.endsWith(".js") ? "application/javascript" : c.name.endsWith(".css") ? "text/css" : "application/octet-stream";
216
+ console.info(`βš™οΈ Uploading '${d}' β†’ '${f}'...`);
218
217
  const { stderr: p } = await y(`wrangler r2 object put "${f}" --file="${a}" --content-type ${u} --jurisdiction=eu --remote`);
219
218
  if (p) throw new Error(p);
220
219
  }
221
220
  }
222
221
  }
223
- await n("dist"), console.info("βœ… Module uploaded to R2.");
222
+ await t("dist"), console.info("βœ… Module uploaded to R2.");
224
223
  } catch (o) {
225
224
  console.error("❌ Error uploading module to R2.", o);
226
225
  }
227
226
  }
228
227
  export {
229
228
  j as buildConfig,
230
- J as buildConnectorConfig,
229
+ k as buildConnectorConfig,
231
230
  F as buildContextConfig,
232
231
  R as buildPresenterConfig,
233
- k as buildPublicDirectoryIndex,
232
+ J as buildPublicDirectoryIndex,
234
233
  A as bumpVersion,
235
234
  D as echoScriptNotImplemented,
236
235
  P as insertLicensesIntoReadme,
@@ -239,5 +238,5 @@ export {
239
238
  T as syncWithGitHub,
240
239
  _ as uploadDirectoryToR2,
241
240
  W as uploadModuleConfigToDO,
242
- U as uploadModuleToR2
241
+ B as uploadModuleToR2
243
242
  };
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@datapos/datapos-development",
3
- "version": "0.3.132",
3
+ "version": "0.3.134",
4
4
  "description": "A library of utilities for managing the Data Positioning repositories.",
5
5
  "license": "MIT",
6
6
  "author": "Jonathan Terrell <terrell.jm@gmail.com>",
@@ -26,7 +26,8 @@
26
26
  "dist"
27
27
  ],
28
28
  "dependencies": {
29
- "@datapos/datapos-shared": "^0.3.252"
29
+ "@datapos/datapos-shared": "^0.3.252",
30
+ "node-pty": "^1.0.0"
30
31
  },
31
32
  "devDependencies": {
32
33
  "@types/node": "^24.10.1",
@@ -41,8 +42,7 @@
41
42
  "nanoid": "^5.1.6",
42
43
  "npm-check-updates": "^19.1.2",
43
44
  "owasp-dependency-check": "^1.0.0",
44
- "prettier": "^3.6.2",
45
- "retire": "^5.3.0",
45
+ "prettier": "^3.7.1",
46
46
  "rollup-plugin-visualizer": "^6.0.5",
47
47
  "ts-to-zod": "^5.1.0",
48
48
  "type-fest": "^5.2.0",
@@ -55,7 +55,8 @@
55
55
  "scripts": {
56
56
  "audit": "npm audit",
57
57
  "build": "vite build",
58
- "check": "npm outdated; npm-check-updates -i && retire",
58
+ "check:updates": "npm outdated; npm-check-updates -i",
59
+ "check:vulnerabilities": "npm run _check:owaspDependencyCheck; npm run _check:owaspBageInsertiions",
59
60
  "document": "npm run _document:licenceReportJSON && npm run _document:licenceReportMarkdown && npm run _document:licenceReportCheck && npm run _document:insertLicensesIntoReadme && npm run _document:licenceTree && npm run _document:licenceTreeCheck",
60
61
  "format": "prettier --write src/",
61
62
  "lint": "eslint .",