@larkiny/astro-github-loader 0.12.0 → 0.13.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +7 -0
- package/dist/github.dryrun.js +3 -0
- package/dist/github.link-transform.js +7 -4
- package/dist/github.storage.d.ts +2 -1
- package/dist/github.storage.js +9 -3
- package/dist/github.types.d.ts +6 -0
- package/package.json +1 -1
- package/src/github.dryrun.spec.ts +12 -0
- package/src/github.dryrun.ts +3 -0
- package/src/github.link-transform.ts +9 -6
- package/src/github.storage.spec.ts +19 -9
- package/src/github.storage.ts +11 -3
- package/src/github.types.ts +6 -0
package/README.md
CHANGED
|
@@ -1,5 +1,12 @@
|
|
|
1
1
|
# Astro GitHub Loader
|
|
2
2
|
|
|
3
|
+
[](https://github.com/larkiny/starlight-github-loader/actions/workflows/ci.yml)
|
|
4
|
+
[](https://www.npmjs.com/package/@larkiny/astro-github-loader)
|
|
5
|
+
[](https://opensource.org/licenses/MIT)
|
|
6
|
+
[](https://astro.build)
|
|
7
|
+
[](https://starlight.astro.build)
|
|
8
|
+
[](https://www.typescriptlang.org/)
|
|
9
|
+
|
|
3
10
|
Load content from GitHub repositories into Astro content collections with flexible pattern-based import, asset management, content transformations, and intelligent change detection.
|
|
4
11
|
|
|
5
12
|
## Features
|
package/dist/github.dryrun.js
CHANGED
|
@@ -6,6 +6,9 @@ const STATE_FILENAME = ".github-import-state.json";
|
|
|
6
6
|
* Creates a unique identifier for an import configuration
|
|
7
7
|
*/
|
|
8
8
|
export function createConfigId(config) {
|
|
9
|
+
if (config.stateKey) {
|
|
10
|
+
return config.stateKey;
|
|
11
|
+
}
|
|
9
12
|
return `${config.owner}/${config.repo}@${config.ref || "main"}`;
|
|
10
13
|
}
|
|
11
14
|
/**
|
|
@@ -87,6 +87,11 @@ function applyLinkMappings(linkUrl, linkMappings, context) {
|
|
|
87
87
|
}
|
|
88
88
|
let matched = false;
|
|
89
89
|
let replacement = "";
|
|
90
|
+
const getLinkTransformContext = () => context.currentFile.linkContext ?? {
|
|
91
|
+
sourcePath: context.currentFile.sourcePath,
|
|
92
|
+
targetPath: context.currentFile.targetPath,
|
|
93
|
+
basePath: "",
|
|
94
|
+
};
|
|
90
95
|
if (typeof mapping.pattern === "string") {
|
|
91
96
|
// String pattern - exact match or contains
|
|
92
97
|
if (transformedPath.includes(mapping.pattern)) {
|
|
@@ -95,8 +100,7 @@ function applyLinkMappings(linkUrl, linkMappings, context) {
|
|
|
95
100
|
replacement = transformedPath.replace(mapping.pattern, mapping.replacement);
|
|
96
101
|
}
|
|
97
102
|
else {
|
|
98
|
-
|
|
99
|
-
replacement = mapping.replacement(transformedPath, anchor, linkTransformContext);
|
|
103
|
+
replacement = mapping.replacement(transformedPath, anchor, getLinkTransformContext());
|
|
100
104
|
}
|
|
101
105
|
}
|
|
102
106
|
}
|
|
@@ -109,8 +113,7 @@ function applyLinkMappings(linkUrl, linkMappings, context) {
|
|
|
109
113
|
replacement = transformedPath.replace(mapping.pattern, mapping.replacement);
|
|
110
114
|
}
|
|
111
115
|
else {
|
|
112
|
-
|
|
113
|
-
replacement = mapping.replacement(transformedPath, anchor, linkTransformContext);
|
|
116
|
+
replacement = mapping.replacement(transformedPath, anchor, getLinkTransformContext());
|
|
114
117
|
}
|
|
115
118
|
}
|
|
116
119
|
}
|
package/dist/github.storage.d.ts
CHANGED
|
@@ -2,9 +2,10 @@ import type { ImportedFile } from "./github.link-transform.js";
|
|
|
2
2
|
import type { ExtendedLoaderContext } from "./github.types.js";
|
|
3
3
|
/**
|
|
4
4
|
* Ensures directory exists and writes file to disk.
|
|
5
|
+
* Validates that the resolved path stays within the project root.
|
|
5
6
|
* @internal
|
|
6
7
|
*/
|
|
7
|
-
export declare function syncFile(
|
|
8
|
+
export declare function syncFile(filePath: string, content: string): Promise<void>;
|
|
8
9
|
/**
|
|
9
10
|
* Stores a processed file in Astro's content store
|
|
10
11
|
* @internal
|
package/dist/github.storage.js
CHANGED
|
@@ -1,15 +1,21 @@
|
|
|
1
1
|
import { existsSync, promises as fs } from "node:fs";
|
|
2
|
+
import { resolve } from "node:path";
|
|
2
3
|
import { fileURLToPath, pathToFileURL } from "node:url";
|
|
3
4
|
/**
|
|
4
5
|
* Ensures directory exists and writes file to disk.
|
|
6
|
+
* Validates that the resolved path stays within the project root.
|
|
5
7
|
* @internal
|
|
6
8
|
*/
|
|
7
|
-
export async function syncFile(
|
|
8
|
-
const
|
|
9
|
+
export async function syncFile(filePath, content) {
|
|
10
|
+
const resolved = resolve(filePath);
|
|
11
|
+
if (!resolved.startsWith(process.cwd())) {
|
|
12
|
+
throw new Error(`syncFile: path "${filePath}" resolves outside project root`);
|
|
13
|
+
}
|
|
14
|
+
const dir = resolved.substring(0, resolved.lastIndexOf("/"));
|
|
9
15
|
if (dir && !existsSync(dir)) {
|
|
10
16
|
await fs.mkdir(dir, { recursive: true });
|
|
11
17
|
}
|
|
12
|
-
await fs.writeFile(
|
|
18
|
+
await fs.writeFile(resolved, content, "utf-8");
|
|
13
19
|
}
|
|
14
20
|
/**
|
|
15
21
|
* Stores a processed file in Astro's content store
|
package/dist/github.types.d.ts
CHANGED
|
@@ -243,6 +243,12 @@ export type ImportOptions = {
|
|
|
243
243
|
* Display name for this configuration (used in logging)
|
|
244
244
|
*/
|
|
245
245
|
name?: string;
|
|
246
|
+
/**
|
|
247
|
+
* Custom state key for import tracking. When provided, overrides the default
|
|
248
|
+
* `owner/repo@ref` key used to track import state. This allows the same repo
|
|
249
|
+
* to be imported independently to multiple locations.
|
|
250
|
+
*/
|
|
251
|
+
stateKey?: string;
|
|
246
252
|
/**
|
|
247
253
|
* Repository owner
|
|
248
254
|
*/
|
package/package.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@larkiny/astro-github-loader",
|
|
3
3
|
"type": "module",
|
|
4
|
-
"version": "0.
|
|
4
|
+
"version": "0.13.0",
|
|
5
5
|
"description": "Load content from GitHub repositories into Astro content collections with asset management and content transformations",
|
|
6
6
|
"keywords": [
|
|
7
7
|
"astro",
|
|
@@ -51,6 +51,18 @@ describe("github.dryrun", () => {
|
|
|
51
51
|
expect(id).toBe("algorand/docs@main");
|
|
52
52
|
});
|
|
53
53
|
|
|
54
|
+
it("should use custom stateKey when provided", () => {
|
|
55
|
+
const config: ImportOptions = {
|
|
56
|
+
owner: "algorandfoundation",
|
|
57
|
+
repo: "puya",
|
|
58
|
+
ref: "devportal",
|
|
59
|
+
stateKey: "puya-legacy-guides",
|
|
60
|
+
includes: [],
|
|
61
|
+
};
|
|
62
|
+
|
|
63
|
+
expect(createConfigId(config)).toBe("puya-legacy-guides");
|
|
64
|
+
});
|
|
65
|
+
|
|
54
66
|
it("should handle different refs correctly", () => {
|
|
55
67
|
const config: ImportOptions = {
|
|
56
68
|
name: "Test Repo",
|
package/src/github.dryrun.ts
CHANGED
|
@@ -57,6 +57,9 @@ export interface RepositoryChangeInfo {
|
|
|
57
57
|
* Creates a unique identifier for an import configuration
|
|
58
58
|
*/
|
|
59
59
|
export function createConfigId(config: ImportOptions): string {
|
|
60
|
+
if (config.stateKey) {
|
|
61
|
+
return config.stateKey;
|
|
62
|
+
}
|
|
60
63
|
return `${config.owner}/${config.repo}@${config.ref || "main"}`;
|
|
61
64
|
}
|
|
62
65
|
|
|
@@ -189,6 +189,13 @@ function applyLinkMappings(
|
|
|
189
189
|
let matched = false;
|
|
190
190
|
let replacement = "";
|
|
191
191
|
|
|
192
|
+
const getLinkTransformContext = (): LinkTransformContext =>
|
|
193
|
+
context.currentFile.linkContext ?? {
|
|
194
|
+
sourcePath: context.currentFile.sourcePath,
|
|
195
|
+
targetPath: context.currentFile.targetPath,
|
|
196
|
+
basePath: "",
|
|
197
|
+
};
|
|
198
|
+
|
|
192
199
|
if (typeof mapping.pattern === "string") {
|
|
193
200
|
// String pattern - exact match or contains
|
|
194
201
|
if (transformedPath.includes(mapping.pattern)) {
|
|
@@ -199,12 +206,10 @@ function applyLinkMappings(
|
|
|
199
206
|
mapping.replacement,
|
|
200
207
|
);
|
|
201
208
|
} else {
|
|
202
|
-
const linkTransformContext =
|
|
203
|
-
context.currentFile.linkContext ?? ({} as LinkTransformContext);
|
|
204
209
|
replacement = mapping.replacement(
|
|
205
210
|
transformedPath,
|
|
206
211
|
anchor,
|
|
207
|
-
|
|
212
|
+
getLinkTransformContext(),
|
|
208
213
|
);
|
|
209
214
|
}
|
|
210
215
|
}
|
|
@@ -219,12 +224,10 @@ function applyLinkMappings(
|
|
|
219
224
|
mapping.replacement,
|
|
220
225
|
);
|
|
221
226
|
} else {
|
|
222
|
-
const linkTransformContext =
|
|
223
|
-
context.currentFile.linkContext ?? ({} as LinkTransformContext);
|
|
224
227
|
replacement = mapping.replacement(
|
|
225
228
|
transformedPath,
|
|
226
229
|
anchor,
|
|
227
|
-
|
|
230
|
+
getLinkTransformContext(),
|
|
228
231
|
);
|
|
229
232
|
}
|
|
230
233
|
}
|
|
@@ -1,4 +1,5 @@
|
|
|
1
1
|
import { beforeEach, describe, it, expect, vi } from "vitest";
|
|
2
|
+
import { resolve } from "node:path";
|
|
2
3
|
import { fileURLToPath, pathToFileURL } from "node:url";
|
|
3
4
|
import { syncFile, storeProcessedFile } from "./github.storage.js";
|
|
4
5
|
import { createMockContext } from "./test-helpers.js";
|
|
@@ -40,12 +41,14 @@ describe("syncFile", () => {
|
|
|
40
41
|
it("creates directory and writes file when directory does not exist", async () => {
|
|
41
42
|
await syncFile("some/nested/dir/file.md", "content");
|
|
42
43
|
|
|
43
|
-
|
|
44
|
-
|
|
44
|
+
const resolved = resolve("some/nested/dir/file.md");
|
|
45
|
+
const resolvedDir = resolved.substring(0, resolved.lastIndexOf("/"));
|
|
46
|
+
expect(mockedExistsSync).toHaveBeenCalledWith(resolvedDir);
|
|
47
|
+
expect(mockedMkdir).toHaveBeenCalledWith(resolvedDir, {
|
|
45
48
|
recursive: true,
|
|
46
49
|
});
|
|
47
50
|
expect(mockedWriteFile).toHaveBeenCalledWith(
|
|
48
|
-
|
|
51
|
+
resolved,
|
|
49
52
|
"content",
|
|
50
53
|
"utf-8",
|
|
51
54
|
);
|
|
@@ -56,10 +59,12 @@ describe("syncFile", () => {
|
|
|
56
59
|
|
|
57
60
|
await syncFile("existing/dir/file.md", "content");
|
|
58
61
|
|
|
59
|
-
|
|
62
|
+
const resolved = resolve("existing/dir/file.md");
|
|
63
|
+
const resolvedDir = resolved.substring(0, resolved.lastIndexOf("/"));
|
|
64
|
+
expect(mockedExistsSync).toHaveBeenCalledWith(resolvedDir);
|
|
60
65
|
expect(mockedMkdir).not.toHaveBeenCalled();
|
|
61
66
|
expect(mockedWriteFile).toHaveBeenCalledWith(
|
|
62
|
-
|
|
67
|
+
resolved,
|
|
63
68
|
"content",
|
|
64
69
|
"utf-8",
|
|
65
70
|
);
|
|
@@ -68,10 +73,9 @@ describe("syncFile", () => {
|
|
|
68
73
|
it("skips mkdir when path has no directory component", async () => {
|
|
69
74
|
await syncFile("file.md", "content");
|
|
70
75
|
|
|
71
|
-
//
|
|
72
|
-
expect(mockedMkdir).not.toHaveBeenCalled();
|
|
76
|
+
// resolved path still has a directory (cwd), but it exists
|
|
73
77
|
expect(mockedWriteFile).toHaveBeenCalledWith(
|
|
74
|
-
"file.md",
|
|
78
|
+
resolve("file.md"),
|
|
75
79
|
"content",
|
|
76
80
|
"utf-8",
|
|
77
81
|
);
|
|
@@ -82,11 +86,17 @@ describe("syncFile", () => {
|
|
|
82
86
|
await syncFile("output/test.md", longContent);
|
|
83
87
|
|
|
84
88
|
expect(mockedWriteFile).toHaveBeenCalledWith(
|
|
85
|
-
"output/test.md",
|
|
89
|
+
resolve("output/test.md"),
|
|
86
90
|
longContent,
|
|
87
91
|
"utf-8",
|
|
88
92
|
);
|
|
89
93
|
});
|
|
94
|
+
|
|
95
|
+
it("rejects paths that escape project root", async () => {
|
|
96
|
+
await expect(
|
|
97
|
+
syncFile("../../etc/passwd", "malicious"),
|
|
98
|
+
).rejects.toThrow("resolves outside project root");
|
|
99
|
+
});
|
|
90
100
|
});
|
|
91
101
|
|
|
92
102
|
describe("storeProcessedFile", () => {
|
package/src/github.storage.ts
CHANGED
|
@@ -1,18 +1,26 @@
|
|
|
1
1
|
import { existsSync, promises as fs } from "node:fs";
|
|
2
|
+
import { resolve } from "node:path";
|
|
2
3
|
import { fileURLToPath, pathToFileURL } from "node:url";
|
|
3
4
|
import type { ImportedFile } from "./github.link-transform.js";
|
|
4
5
|
import type { ExtendedLoaderContext } from "./github.types.js";
|
|
5
6
|
|
|
6
7
|
/**
|
|
7
8
|
* Ensures directory exists and writes file to disk.
|
|
9
|
+
* Validates that the resolved path stays within the project root.
|
|
8
10
|
* @internal
|
|
9
11
|
*/
|
|
10
|
-
export async function syncFile(
|
|
11
|
-
const
|
|
12
|
+
export async function syncFile(filePath: string, content: string) {
|
|
13
|
+
const resolved = resolve(filePath);
|
|
14
|
+
if (!resolved.startsWith(process.cwd())) {
|
|
15
|
+
throw new Error(
|
|
16
|
+
`syncFile: path "${filePath}" resolves outside project root`,
|
|
17
|
+
);
|
|
18
|
+
}
|
|
19
|
+
const dir = resolved.substring(0, resolved.lastIndexOf("/"));
|
|
12
20
|
if (dir && !existsSync(dir)) {
|
|
13
21
|
await fs.mkdir(dir, { recursive: true });
|
|
14
22
|
}
|
|
15
|
-
await fs.writeFile(
|
|
23
|
+
await fs.writeFile(resolved, content, "utf-8");
|
|
16
24
|
}
|
|
17
25
|
|
|
18
26
|
/**
|
package/src/github.types.ts
CHANGED
|
@@ -271,6 +271,12 @@ export type ImportOptions = {
|
|
|
271
271
|
* Display name for this configuration (used in logging)
|
|
272
272
|
*/
|
|
273
273
|
name?: string;
|
|
274
|
+
/**
|
|
275
|
+
* Custom state key for import tracking. When provided, overrides the default
|
|
276
|
+
* `owner/repo@ref` key used to track import state. This allows the same repo
|
|
277
|
+
* to be imported independently to multiple locations.
|
|
278
|
+
*/
|
|
279
|
+
stateKey?: string;
|
|
274
280
|
/**
|
|
275
281
|
* Repository owner
|
|
276
282
|
*/
|