@datapos/datapos-development 0.3.133 β†’ 0.3.135

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md CHANGED
@@ -1,7 +1,7 @@
1
1
  # Data Positioning Development Library
2
2
 
3
3
  [![Quality Gate Status](https://sonarcloud.io/api/project_badges/measure?project=data-positioning_datapos-development&metric=alert_status)](https://sonarcloud.io/summary/new_code?id=data-positioning_datapos-development)
4
- <span><!-- OWASP_BADGE_START -->[![OWASP](https://img.shields.io/badge/OWASP-passed-ForestGreen)](https://data-positioning.github.io/test-lib/dependency-check-reports/dependency-check-report.html)<!-- OWASP_BADGE_END --></span>
4
+ <span><!-- OWASP_BADGE_START -->[![OWASP](https://img.shields.io/badge/OWASP-passed-4CAF50)](https://data-positioning.github.io/test-lib/dependency-check-reports/dependency-check-report.html)<!-- OWASP_BADGE_END --></span>
5
5
  [![npm version](https://img.shields.io/npm/v/@datapos/datapos-development.svg)](https://www.npmjs.com/package/@datapos/datapos-development)
6
6
  [![License: MIT](https://img.shields.io/badge/License-MIT-blue.svg)](./LICENSE)
7
7
 
@@ -62,14 +62,13 @@ The OWASP Dependency Check Report identifies known vulnerabilities in project de
62
62
 
63
63
  ### Dependency Licenses
64
64
 
65
- The following table lists top-level production and peer dependencies only. All dependency licenses (including transitive dependencies) have been recursively verified to conform to Apache-2.0, CC0-1.0, MIT, or n/a. Developers cloning this repository should independently verify dev and optional dependencies; users of the uploaded library are covered by these checks.
65
+ The following table lists top-level production and peer dependencies. All these dependencies (including transitive ones) have been recursively verified to use Apache-2.0, CC0-1.0, or MITβ€”commercially friendly licenses with minimal restrictions. Developers cloning this repository should independently verify dev and optional dependencies; users of the published library are covered by these checks.
66
66
 
67
67
  <!-- DEPENDENCY_LICENSES_START -->
68
-
69
68
  | Name | Type | Installed | Latest | Latest Modified |
70
69
  | :---------------------- | :--: | :-------: | :-----: | :----------------------- |
71
70
  | @datapos/datapos-shared | MIT | 0.3.252 | 0.3.252 | 2025-11-25T16:48:28.532Z |
72
-
71
+ | node-pty | MIT | 1.0.0 | 1.0.0 | 2025-11-03T11:45:17.960Z |
73
72
  <!-- DEPENDENCY_LICENSES_END -->
74
73
 
75
74
  ### Bundle Analysis Report
@@ -1,63 +1,63 @@
1
1
  import { exec as b } from "node:child_process";
2
- import { promises as t } from "node:fs";
2
+ import { promises as n } from "node:fs";
3
3
  import { nanoid as $ } from "nanoid";
4
4
  import { promisify as O } from "node:util";
5
5
  const S = ["createObject", "dropObject", "removeRecords", "upsertRecords"], v = ["findObject", "getRecord", "listNodes", "previewObject", "retrieveRecords"], y = O(b);
6
6
  async function j() {
7
7
  try {
8
8
  console.info("πŸš€ Building configuration...");
9
- const e = JSON.parse(await t.readFile("package.json", "utf8")), o = JSON.parse(await t.readFile("config.json", "utf8"));
10
- e.name != null && (o.id = e.name.replace("@datapos/", "").replace("@data-positioning/", "")), e.version != null && (o.version = e.version), await t.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Configuration built.");
9
+ const e = JSON.parse(await n.readFile("package.json", "utf8")), o = JSON.parse(await n.readFile("config.json", "utf8"));
10
+ e.name != null && (o.id = e.name.replace("@datapos/", "").replace("@data-positioning/", "")), e.version != null && (o.version = e.version), await n.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Configuration built.");
11
11
  } catch (e) {
12
12
  console.error("❌ Error building configuration.", e);
13
13
  }
14
14
  }
15
- async function k(e) {
15
+ async function J(e) {
16
16
  try {
17
17
  console.info(`πŸš€ Building public directory index for identifier '${e}'...`);
18
18
  const o = {};
19
19
  async function i(s, r) {
20
20
  console.info(`βš™οΈ Processing directory '${s}'...`);
21
- const d = [], c = s.substring(`public/${e}`.length);
22
- o[c] = d;
21
+ const l = [], c = s.substring(`public/${e}`.length);
22
+ o[c] = l;
23
23
  for (const a of r) {
24
- const l = `${s}/${a}`;
24
+ const d = `${s}/${a}`;
25
25
  try {
26
- const f = await t.stat(l);
26
+ const f = await n.stat(d);
27
27
  if (f.isDirectory()) {
28
- const u = await t.readdir(l), p = { childCount: u.length, name: `${a}`, typeId: "folder" };
29
- d.push(p), await i(l, u);
28
+ const u = await n.readdir(d), p = { childCount: u.length, name: `${a}`, typeId: "folder" };
29
+ l.push(p), await i(d, u);
30
30
  } else {
31
31
  const u = { id: $(), lastModifiedAt: f.mtimeMs, name: a, size: f.size, typeId: "object" };
32
- d.push(u);
32
+ l.push(u);
33
33
  }
34
34
  } catch (f) {
35
35
  throw new Error(`Unable to get information for '${a}' in 'buildPublicDirectoryIndex'. ${String(f)}`);
36
36
  }
37
37
  }
38
- d.sort((a, l) => {
39
- const f = a.typeId.localeCompare(l.typeId);
40
- return f === 0 ? a.name.localeCompare(l.name) : f;
38
+ l.sort((a, d) => {
39
+ const f = a.typeId.localeCompare(d.typeId);
40
+ return f === 0 ? a.name.localeCompare(d.name) : f;
41
41
  });
42
42
  }
43
- const n = await t.readdir(`public/${e}`);
44
- await i(`public/${e}`, n), await t.writeFile(`./public/${e}Index.json`, JSON.stringify(o), "utf8"), console.info("βœ… Public directory index built.");
43
+ const t = await n.readdir(`public/${e}`);
44
+ await i(`public/${e}`, t), await n.writeFile(`./public/${e}Index.json`, JSON.stringify(o), "utf8"), console.info("βœ… Public directory index built.");
45
45
  } catch (o) {
46
46
  console.error("❌ Error building public directory index.", o);
47
47
  }
48
48
  }
49
- async function J() {
49
+ async function k() {
50
50
  try {
51
51
  console.info("πŸš€ Building connector configuration...");
52
- const e = JSON.parse(await t.readFile("package.json", "utf8")), o = JSON.parse(await t.readFile("config.json", "utf8")), i = await t.readFile("src/index.ts", "utf8");
53
- let n = !1, s = !1;
54
- const r = /^\s{4}(?:async\s+)?(private\s+)?(?:public\s+|protected\s+)?([A-Za-z_]\w*)\s*\(/gm, d = [...i.matchAll(r)].filter((a) => a[1] == null && a[2] !== "constructor").map((a) => {
55
- const l = a[2];
56
- return n = n || S.includes(l), s = s || v.includes(l), l;
52
+ const e = JSON.parse(await n.readFile("package.json", "utf8")), o = JSON.parse(await n.readFile("config.json", "utf8")), i = await n.readFile("src/index.ts", "utf8");
53
+ let t = !1, s = !1;
54
+ const r = /^\s{4}(?:async\s+)?(private\s+)?(?:public\s+|protected\s+)?([A-Za-z_]\w*)\s*\(/gm, l = [...i.matchAll(r)].filter((a) => a[1] == null && a[2] !== "constructor").map((a) => {
55
+ const d = a[2];
56
+ return t = t || S.includes(d), s = s || v.includes(d), d;
57
57
  });
58
- d.length > 0 ? console.info(`ℹ️ Implements ${d.length} operations.`) : console.warn("⚠️ Implements no operations.");
59
- const c = s && n ? "bidirectional" : s ? "source" : n ? "destination" : "unknown";
60
- c && console.info(`ℹ️ Supports ${c} usage.`), e.name != null && (o.id = e.name), o.operations = d, o.usageId = c, e.version != null && (o.version = e.version), await t.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Connector configuration built.");
58
+ l.length > 0 ? console.info(`ℹ️ Implements ${l.length} operations.`) : console.warn("⚠️ Implements no operations.");
59
+ const c = s && t ? "bidirectional" : s ? "source" : t ? "destination" : "unknown";
60
+ c && console.info(`ℹ️ Supports ${c} usage.`), e.name != null && (o.id = e.name), o.operations = l, o.usageId = c, e.version != null && (o.version = e.version), await n.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Connector configuration built.");
61
61
  } catch (e) {
62
62
  console.error("❌ Error building connector configuration.", e);
63
63
  }
@@ -65,8 +65,8 @@ async function J() {
65
65
  async function F() {
66
66
  try {
67
67
  console.info("πŸš€ Building context configuration...");
68
- const e = JSON.parse(await t.readFile("package.json", "utf8")), o = JSON.parse(await t.readFile("config.json", "utf8")), i = await t.readFile("src/index.ts", "utf8"), n = /^\s{4}(?:async\s+)?(private\s+)?(?:public\s+|protected\s+)?([A-Za-z_]\w*)\s*\(/gm, s = [...i.matchAll(n)].filter((r) => r[1] == null && r[2] !== "constructor").map((r) => r[2]);
69
- e.name != null && (o.id = e.name), o.operations = s, e.version != null && (o.version = e.version), await t.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Context configuration built.");
68
+ const e = JSON.parse(await n.readFile("package.json", "utf8")), o = JSON.parse(await n.readFile("config.json", "utf8")), i = await n.readFile("src/index.ts", "utf8"), t = /^\s{4}(?:async\s+)?(private\s+)?(?:public\s+|protected\s+)?([A-Za-z_]\w*)\s*\(/gm, s = [...i.matchAll(t)].filter((r) => r[1] == null && r[2] !== "constructor").map((r) => r[2]);
69
+ e.name != null && (o.id = e.name), o.operations = s, e.version != null && (o.version = e.version), await n.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Context configuration built.");
70
70
  } catch (e) {
71
71
  console.error("❌ Error building context configuration.", e);
72
72
  }
@@ -74,8 +74,8 @@ async function F() {
74
74
  async function R() {
75
75
  try {
76
76
  console.info("πŸš€ Building presenter configuration...");
77
- const e = JSON.parse(await t.readFile("package.json", "utf8")), o = JSON.parse(await t.readFile("config.json", "utf8")), i = await t.readFile("src/index.ts", "utf8"), n = /^\s{4}(?:async\s+)?(private\s+)?(?:public\s+|protected\s+)?([A-Za-z_]\w*)\s*\(/gm, s = [...i.matchAll(n)].filter((r) => !r[1] && r[2] !== "constructor").map((r) => r[2]);
78
- e.name != null && (o.id = e.name), o.operations = s, e.version != null && (o.version = e.version), await t.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Presenter configuration built.");
77
+ const e = JSON.parse(await n.readFile("package.json", "utf8")), o = JSON.parse(await n.readFile("config.json", "utf8")), i = await n.readFile("src/index.ts", "utf8"), t = /^\s{4}(?:async\s+)?(private\s+)?(?:public\s+|protected\s+)?([A-Za-z_]\w*)\s*\(/gm, s = [...i.matchAll(t)].filter((r) => !r[1] && r[2] !== "constructor").map((r) => r[2]);
78
+ e.name != null && (o.id = e.name), o.operations = s, e.version != null && (o.version = e.version), await n.writeFile("config.json", JSON.stringify(o, void 0, 4), "utf8"), console.info("βœ… Presenter configuration built.");
79
79
  } catch (e) {
80
80
  console.error("❌ Error building context configuration.", e);
81
81
  }
@@ -83,12 +83,12 @@ async function R() {
83
83
  async function A(e = "./") {
84
84
  try {
85
85
  console.info("πŸš€ Bumping version...");
86
- const o = JSON.parse(await t.readFile(`${e}package.json`, "utf8"));
86
+ const o = JSON.parse(await n.readFile(`${e}package.json`, "utf8"));
87
87
  if (o.version == null)
88
- o.version = "0.0.001", await t.writeFile(`${e}package.json`, JSON.stringify(o, void 0, 4), "utf8"), console.warn(`⚠️ Version initialised to ${o.version}.`);
88
+ o.version = "0.0.001", await n.writeFile(`${e}package.json`, JSON.stringify(o, void 0, 4), "utf8"), console.warn(`⚠️ Version initialised to ${o.version}.`);
89
89
  else {
90
- const i = o.version, n = o.version.split(".");
91
- o.version = `${n[0]}.${n[1]}.${Number(n[2]) + 1}`, await t.writeFile(`${e}package.json`, JSON.stringify(o, void 0, 4), "utf8"), console.info(`βœ… Version bumped from ${i} to ${o.version}.`);
90
+ const i = o.version, t = o.version.split(".");
91
+ o.version = `${t[0]}.${t[1]}.${Number(t[2]) + 1}`, await n.writeFile(`${e}package.json`, JSON.stringify(o, void 0, 4), "utf8"), console.info(`βœ… Version bumped from ${i} to ${o.version}.`);
92
92
  }
93
93
  } catch (o) {
94
94
  console.error("❌ Error bumping package version.", o);
@@ -100,12 +100,12 @@ function D(e) {
100
100
  async function P() {
101
101
  const e = "<!-- DEPENDENCY_LICENSES_START -->", o = "<!-- DEPENDENCY_LICENSES_END -->";
102
102
  try {
103
- const i = (await t.readFile("./licenses.md", "utf8")).trim(), n = await t.readFile("./README.md", "utf8"), s = n.indexOf(e), r = n.indexOf(o);
103
+ const i = (await n.readFile("./licenses.md", "utf8")).trim(), t = await n.readFile("./README.md", "utf8"), s = t.indexOf(e), r = t.indexOf(o);
104
104
  (s === -1 || r === -1) && (console.error("❌ Dependency license markers not found in readme file."), process.exit(1));
105
- const d = n.substring(0, s + e.length) + `
105
+ const l = t.substring(0, s + e.length) + `
106
106
  ` + i + `
107
- ` + n.substring(r);
108
- await t.writeFile("README.md", d, "utf8"), console.log("βœ… Readme file updated with license information");
107
+ ` + t.substring(r);
108
+ await n.writeFile("README.md", l, "utf8"), console.log("βœ… Readme file updated with license information");
109
109
  } catch (i) {
110
110
  console.error("❌ Error updating readme file.", i), process.exit(1);
111
111
  }
@@ -113,12 +113,12 @@ async function P() {
113
113
  async function I() {
114
114
  const e = "<!-- OWASP_BADGE_START -->", o = "<!-- OWASP_BADGE_END -->";
115
115
  try {
116
- const i = JSON.parse(await t.readFile("./dependency-check-reports/dependency-check-report.json", "utf-8")), n = { critical: 0, high: 0, moderate: 0, low: 0, unknown: 0 };
116
+ const i = JSON.parse(await n.readFile("./dependency-check-reports/dependency-check-report.json", "utf-8")), t = { critical: 0, high: 0, moderate: 0, low: 0, unknown: 0 };
117
117
  for (const g of i.dependencies)
118
118
  if (g.vulnerabilities != null)
119
119
  for (const m of g.vulnerabilities) {
120
120
  const w = m.severity?.toLowerCase() ?? "unknown";
121
- w in n ? n[w]++ : n.unknown++;
121
+ w in t ? t[w]++ : t.unknown++;
122
122
  }
123
123
  const s = {
124
124
  critical: { color: "D32F2F", label: "critical" },
@@ -126,23 +126,20 @@ async function I() {
126
126
  moderate: { color: "FBC02D", label: "moderate" },
127
127
  low: { color: "6D8C31", label: "low" },
128
128
  unknown: { color: "616161", label: "unknown" }
129
- }, r = Object.values(n).reduce((g, m) => g + m, 0);
130
- console.info(`βœ… Total vulnerabilities found: ${r}`), console.info(
131
- ` Critical: ${n.critical}, High: ${n.high}, Moderate: ${n.moderate}, Low: ${n.low}, Unknown: ${n.unknown}`
132
- );
133
- const d = JSON.parse(await t.readFile("config.json", "utf8")), c = [];
134
- if (r === 0)
135
- c.push(`[![OWASP](https://img.shields.io/badge/OWASP-passed-4CAF50)](https://data-positioning.github.io/${d.id}/dependency-check-reports/dependency-check-report.html)`);
129
+ }, r = JSON.parse(await n.readFile("config.json", "utf8")), l = [];
130
+ if (Object.values(t).reduce((g, m) => g + m, 0) === 0)
131
+ console.info("βœ… No vulnerabilities found."), l.push(`[![OWASP](https://img.shields.io/badge/OWASP-passed-4CAF50)](https://data-positioning.github.io/${r.id}/dependency-check-reports/dependency-check-report.html)`);
136
132
  else
137
- for (const [g, m] of Object.entries(n)) {
138
- if (m === 0) continue;
139
- const w = s[g], h = `https://img.shields.io/badge/OWASP-${m}%20${w.label}-${w.color}`;
140
- c.push(`[![OWASP](${h})](https://data-positioning.github.io/${d.id}/dependency-check-reports/dependency-check-report.html)`);
133
+ for (const [g, m] of Object.entries(t)) {
134
+ const w = s[g];
135
+ if (console.warn(`⚠️ ${m} ${w.label} vulnerability(ies) found.`), m === 0) continue;
136
+ const h = `https://img.shields.io/badge/OWASP-${m}%20${w.label}-${w.color}`;
137
+ l.push(`[![OWASP](${h})](https://data-positioning.github.io/${r.id}/dependency-check-reports/dependency-check-report.html)`);
141
138
  }
142
- const a = await t.readFile("./README.md", "utf8"), l = a.indexOf(e), f = a.indexOf(o);
143
- (l === -1 || f === -1) && (console.error("❌ OWASP badge markers not found in README.md."), process.exit(1));
144
- const u = c.join(" "), p = a.substring(0, l + e.length) + u + a.substring(f);
145
- await t.writeFile("README.md", p, "utf8"), console.info("βœ… OWASP dependency check badges inserted into README.md");
139
+ const a = await n.readFile("./README.md", "utf8"), d = a.indexOf(e), f = a.indexOf(o);
140
+ (d === -1 || f === -1) && (console.error("❌ OWASP badge markers not found in README.md."), process.exit(1));
141
+ const u = l.join(" "), p = a.substring(0, d + e.length) + u + a.substring(f);
142
+ await n.writeFile("README.md", p, "utf8"), console.info("βœ… OWASP dependency check badge(s) inserted into README.md");
146
143
  } catch (i) {
147
144
  console.error("❌ Error updating README with OWASP badges:", i), process.exit(1);
148
145
  }
@@ -150,7 +147,7 @@ async function I() {
150
147
  async function M() {
151
148
  try {
152
149
  console.info("πŸš€ Sending deployment notice...");
153
- const e = JSON.parse(await t.readFile("config.json", "utf8")), o = {
150
+ const e = JSON.parse(await n.readFile("config.json", "utf8")), o = {
154
151
  body: JSON.stringify(e),
155
152
  headers: { "Content-Type": "application/json" },
156
153
  method: "PUT"
@@ -164,7 +161,7 @@ async function M() {
164
161
  async function T() {
165
162
  try {
166
163
  console.info("πŸš€ Synchronising with GitHub....");
167
- const e = JSON.parse(await t.readFile("package.json", "utf8"));
164
+ const e = JSON.parse(await n.readFile("package.json", "utf8"));
168
165
  await y("git add ."), await y(`git commit -m "v${e.version}"`), await y("git push origin main:main"), console.info(`βœ… Synchronised version ${e.version} with GitHub.`);
169
166
  } catch (e) {
170
167
  console.error("❌ Error synchronising with GitHub.", e);
@@ -173,12 +170,12 @@ async function T() {
173
170
  async function _(e, o) {
174
171
  try {
175
172
  console.info("πŸš€ Uploading directory to R2....");
176
- async function i(s, r, d) {
177
- for (const c of d) {
178
- const a = `${s}/${c}`, l = `${r}/${c}`;
179
- if ((await t.stat(a)).isDirectory()) {
180
- const u = await t.readdir(a);
181
- await i(a, l, u);
173
+ async function i(s, r, l) {
174
+ for (const c of l) {
175
+ const a = `${s}/${c}`, d = `${r}/${c}`;
176
+ if ((await n.stat(a)).isDirectory()) {
177
+ const u = await n.readdir(a);
178
+ await i(a, d, u);
182
179
  } else {
183
180
  console.info(`βš™οΈ Uploading '${s}/${c}'...`);
184
181
  const u = `wrangler r2 object put "datapos-sample-data-eu/${r}/${c}" --file="${s}/${c}" --jurisdiction=eu --remote`, p = await y(u);
@@ -186,8 +183,8 @@ async function _(e, o) {
186
183
  }
187
184
  }
188
185
  }
189
- const n = await t.readdir(`${e}/${o}/`);
190
- await i(`${e}/${o}`, o, n), console.info("βœ… Directory uploaded to R2.");
186
+ const t = await n.readdir(`${e}/${o}/`);
187
+ await i(`${e}/${o}`, o, t), console.info("βœ… Directory uploaded to R2.");
191
188
  } catch (i) {
192
189
  console.error("❌ Error uploading directory to R2.", i);
193
190
  }
@@ -195,44 +192,44 @@ async function _(e, o) {
195
192
  async function W() {
196
193
  try {
197
194
  console.info("πŸš€ Uploading module configuration....");
198
- const e = JSON.parse(await t.readFile("config.json", "utf8")), o = e.id, i = {
195
+ const e = JSON.parse(await n.readFile("config.json", "utf8")), o = e.id, i = {
199
196
  body: JSON.stringify(e),
200
197
  headers: { "Content-Type": "application/json" },
201
198
  method: "PUT"
202
- }, n = await fetch(`https://api.datapos.app/states/${o}`, i);
203
- if (!n.ok) throw new Error(await n.text());
199
+ }, t = await fetch(`https://api.datapos.app/states/${o}`, i);
200
+ if (!t.ok) throw new Error(await t.text());
204
201
  console.info("βœ… Module configuration uploaded.");
205
202
  } catch (e) {
206
203
  console.error("❌ Error uploading module configuration.", e);
207
204
  }
208
205
  }
209
- async function U(e) {
206
+ async function B(e) {
210
207
  try {
211
208
  console.info("πŸš€ Uploading module to R2...");
212
- const i = `v${JSON.parse(await t.readFile("package.json", "utf8")).version}`;
213
- async function n(s, r = "") {
214
- const d = await t.readdir(s, { withFileTypes: !0 });
215
- for (const c of d) {
216
- const a = `${s}/${c.name}`, l = r ? `${r}/${c.name}` : c.name;
209
+ const i = `v${JSON.parse(await n.readFile("package.json", "utf8")).version}`;
210
+ async function t(s, r = "") {
211
+ const l = await n.readdir(s, { withFileTypes: !0 });
212
+ for (const c of l) {
213
+ const a = `${s}/${c.name}`, d = r ? `${r}/${c.name}` : c.name;
217
214
  if (!c.isDirectory()) {
218
- const f = `${e}_${i}/${l}`.replace(/\\/g, "/"), u = c.name.endsWith(".js") ? "application/javascript" : c.name.endsWith(".css") ? "text/css" : "application/octet-stream";
219
- console.info(`βš™οΈ Uploading '${l}' β†’ '${f}'...`);
215
+ const f = `${e}_${i}/${d}`.replace(/\\/g, "/"), u = c.name.endsWith(".js") ? "application/javascript" : c.name.endsWith(".css") ? "text/css" : "application/octet-stream";
216
+ console.info(`βš™οΈ Uploading '${d}' β†’ '${f}'...`);
220
217
  const { stderr: p } = await y(`wrangler r2 object put "${f}" --file="${a}" --content-type ${u} --jurisdiction=eu --remote`);
221
218
  if (p) throw new Error(p);
222
219
  }
223
220
  }
224
221
  }
225
- await n("dist"), console.info("βœ… Module uploaded to R2.");
222
+ await t("dist"), console.info("βœ… Module uploaded to R2.");
226
223
  } catch (o) {
227
224
  console.error("❌ Error uploading module to R2.", o);
228
225
  }
229
226
  }
230
227
  export {
231
228
  j as buildConfig,
232
- J as buildConnectorConfig,
229
+ k as buildConnectorConfig,
233
230
  F as buildContextConfig,
234
231
  R as buildPresenterConfig,
235
- k as buildPublicDirectoryIndex,
232
+ J as buildPublicDirectoryIndex,
236
233
  A as bumpVersion,
237
234
  D as echoScriptNotImplemented,
238
235
  P as insertLicensesIntoReadme,
@@ -241,5 +238,5 @@ export {
241
238
  T as syncWithGitHub,
242
239
  _ as uploadDirectoryToR2,
243
240
  W as uploadModuleConfigToDO,
244
- U as uploadModuleToR2
241
+ B as uploadModuleToR2
245
242
  };
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@datapos/datapos-development",
3
- "version": "0.3.133",
3
+ "version": "0.3.135",
4
4
  "description": "A library of utilities for managing the Data Positioning repositories.",
5
5
  "license": "MIT",
6
6
  "author": "Jonathan Terrell <terrell.jm@gmail.com>",
@@ -26,7 +26,8 @@
26
26
  "dist"
27
27
  ],
28
28
  "dependencies": {
29
- "@datapos/datapos-shared": "^0.3.252"
29
+ "@datapos/datapos-shared": "^0.3.252",
30
+ "node-pty": "^1.0.0"
30
31
  },
31
32
  "devDependencies": {
32
33
  "@types/node": "^24.10.1",
@@ -41,8 +42,7 @@
41
42
  "nanoid": "^5.1.6",
42
43
  "npm-check-updates": "^19.1.2",
43
44
  "owasp-dependency-check": "^1.0.0",
44
- "prettier": "^3.6.2",
45
- "retire": "^5.3.0",
45
+ "prettier": "^3.7.1",
46
46
  "rollup-plugin-visualizer": "^6.0.5",
47
47
  "ts-to-zod": "^5.1.0",
48
48
  "type-fest": "^5.2.0",
@@ -55,7 +55,8 @@
55
55
  "scripts": {
56
56
  "audit": "npm audit",
57
57
  "build": "vite build",
58
- "check": "npm outdated; npm-check-updates -i && retire",
58
+ "check:updates": "npm outdated; npm-check-updates -i",
59
+ "check:vulnerabilities": "npm run _check:owaspDependencyCheck; npm run _check:owaspBageInsertiions",
59
60
  "document": "npm run _document:licenceReportJSON && npm run _document:licenceReportMarkdown && npm run _document:licenceReportCheck && npm run _document:insertLicensesIntoReadme && npm run _document:licenceTree && npm run _document:licenceTreeCheck",
60
61
  "format": "prettier --write src/",
61
62
  "lint": "eslint .",
@@ -69,9 +70,9 @@
69
70
  "_check:owaspBageInsertiions": "node -e \"import('./dist/datapos-development.es.js').then(m => m.insertOWASPDependencyCheckBadgeIntoReadme())\"",
70
71
  "_document:licenceReportJSON": "license-report --only=prod,peer --department.value=n/a --licensePeriod=n/a --material=n/a --relatedTo.value=n/a > licenses.json",
71
72
  "_document:licenceReportMarkdown": "license-report --config license-report-config.json --only=prod,peer --output=markdown > licenses.md",
72
- "_document:licenceReportCheck": "license-report-check --source ./licenses.json --allowed 'MIT' --allowed 'n/a' --allowed 'Apache-2.0' --allowed 'CC0-1.0' --output=table",
73
+ "_document:licenceReportCheck": "license-report-check --source ./licenses.json --allowed 'MIT' --output=table",
73
74
  "_document:licenceTree": "license-report-recursive --only=prod,peer --department.value=n/a --licensePeriod=n/a --material=n/a --relatedTo.value=n/a --recurse --output=tree > licenseTree.json",
74
- "_document:licenceTreeCheck": "license-report-check --source ./licenseTree.json --allowed 'MIT' --allowed 'n/a' --allowed 'Apache-2.0' --allowed 'CC0-1.0' --output=table",
75
+ "_document:licenceTreeCheck": "license-report-check --source ./licenseTree.json --allowed 'MIT' --output=table",
75
76
  "_document:insertLicensesIntoReadme": "node -e \"import('./dist/datapos-development.es.js').then(m => m.insertLicensesIntoReadme())\"",
76
77
  "_sync:withGitHub": "node -e \"import('./dist/datapos-development.es.js').then(m => m.syncWithGitHub())\"",
77
78
  "_update:sharedDep": "npm install @datapos/datapos-shared@latest"