@virstack/doc-ingest 1.0.2 → 1.0.4
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +3 -0
- package/dist/cli.js +0 -0
- package/dist/core/config.js +3 -1
- package/dist/core/config.js.map +1 -1
- package/package.json +4 -3
- package/dist/aiAdapters.d.ts +0 -25
- package/dist/aiAdapters.d.ts.map +0 -1
- package/dist/aiAdapters.js +0 -50
- package/dist/aiAdapters.js.map +0 -1
- package/dist/assets/logo.png +0 -0
- package/dist/batchPipeline.d.ts +0 -52
- package/dist/batchPipeline.d.ts.map +0 -1
- package/dist/batchPipeline.js +0 -81
- package/dist/batchPipeline.js.map +0 -1
- package/dist/config.d.ts +0 -26
- package/dist/config.d.ts.map +0 -1
- package/dist/config.js +0 -97
- package/dist/config.js.map +0 -1
- package/dist/logger.d.ts +0 -24
- package/dist/logger.d.ts.map +0 -1
- package/dist/logger.js +0 -36
- package/dist/logger.js.map +0 -1
- package/dist/logo.d.ts +0 -2
- package/dist/logo.d.ts.map +0 -1
- package/dist/logo.js +0 -3
- package/dist/logo.js.map +0 -1
- package/dist/nodes/geminiExtraction.d.ts +0 -19
- package/dist/nodes/geminiExtraction.d.ts.map +0 -1
- package/dist/nodes/geminiExtraction.js +0 -87
- package/dist/nodes/geminiExtraction.js.map +0 -1
- package/dist/nodes/openrouterEmbedder.d.ts +0 -7
- package/dist/nodes/openrouterEmbedder.d.ts.map +0 -1
- package/dist/nodes/openrouterEmbedder.js +0 -31
- package/dist/nodes/openrouterEmbedder.js.map +0 -1
- package/dist/nodes/upstashUpsert.d.ts +0 -7
- package/dist/nodes/upstashUpsert.d.ts.map +0 -1
- package/dist/nodes/upstashUpsert.js +0 -45
- package/dist/nodes/upstashUpsert.js.map +0 -1
- package/dist/pipeline.d.ts +0 -303
- package/dist/pipeline.d.ts.map +0 -1
- package/dist/pipeline.js +0 -93
- package/dist/pipeline.js.map +0 -1
- package/dist/state.d.ts +0 -52
- package/dist/state.d.ts.map +0 -1
- package/dist/state.js +0 -27
- package/dist/state.js.map +0 -1
- package/dist/vectorStore.d.ts +0 -24
- package/dist/vectorStore.d.ts.map +0 -1
- package/dist/vectorStore.js +0 -22
- package/dist/vectorStore.js.map +0 -1
package/dist/pipeline.d.ts
DELETED
|
@@ -1,303 +0,0 @@
|
|
|
1
|
-
export declare function buildPipeline(): import("@langchain/langgraph").CompiledStateGraph<import("@langchain/langgraph").StateType<{
|
|
2
|
-
filePath: {
|
|
3
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
4
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
5
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
6
|
-
};
|
|
7
|
-
mimeType: {
|
|
8
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
9
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
10
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
11
|
-
};
|
|
12
|
-
rawText: {
|
|
13
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
14
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
15
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
16
|
-
};
|
|
17
|
-
pdfChunks: {
|
|
18
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
19
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
20
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
21
|
-
};
|
|
22
|
-
markdownParts: import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
23
|
-
markdown: {
|
|
24
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
25
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
26
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
27
|
-
};
|
|
28
|
-
textChunks: {
|
|
29
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
30
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
31
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
32
|
-
};
|
|
33
|
-
vectors: {
|
|
34
|
-
(): import("@langchain/langgraph").LastValue<number[][]>;
|
|
35
|
-
(annotation: import("@langchain/langgraph").SingleReducer<number[][], number[][]>): import("@langchain/langgraph").BinaryOperatorAggregate<number[][], number[][]>;
|
|
36
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
37
|
-
};
|
|
38
|
-
}>, import("@langchain/langgraph").UpdateType<{
|
|
39
|
-
filePath: {
|
|
40
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
41
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
42
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
43
|
-
};
|
|
44
|
-
mimeType: {
|
|
45
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
46
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
47
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
48
|
-
};
|
|
49
|
-
rawText: {
|
|
50
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
51
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
52
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
53
|
-
};
|
|
54
|
-
pdfChunks: {
|
|
55
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
56
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
57
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
58
|
-
};
|
|
59
|
-
markdownParts: import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
60
|
-
markdown: {
|
|
61
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
62
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
63
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
64
|
-
};
|
|
65
|
-
textChunks: {
|
|
66
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
67
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
68
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
69
|
-
};
|
|
70
|
-
vectors: {
|
|
71
|
-
(): import("@langchain/langgraph").LastValue<number[][]>;
|
|
72
|
-
(annotation: import("@langchain/langgraph").SingleReducer<number[][], number[][]>): import("@langchain/langgraph").BinaryOperatorAggregate<number[][], number[][]>;
|
|
73
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
74
|
-
};
|
|
75
|
-
}>, "markdownMerger" | "markdownNormalizer" | "llmExtractionNode" | "__start__" | "fileTypeRouter" | "libreOfficeToPdf" | "pdfSplitter" | "textExtractorNode" | "saveMarkdown" | "markdownChunker" | "vectorEmbedderNode" | "vectorUpsertNode", {
|
|
76
|
-
filePath: {
|
|
77
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
78
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
79
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
80
|
-
};
|
|
81
|
-
mimeType: {
|
|
82
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
83
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
84
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
85
|
-
};
|
|
86
|
-
rawText: {
|
|
87
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
88
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
89
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
90
|
-
};
|
|
91
|
-
pdfChunks: {
|
|
92
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
93
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
94
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
95
|
-
};
|
|
96
|
-
markdownParts: import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
97
|
-
markdown: {
|
|
98
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
99
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
100
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
101
|
-
};
|
|
102
|
-
textChunks: {
|
|
103
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
104
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
105
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
106
|
-
};
|
|
107
|
-
vectors: {
|
|
108
|
-
(): import("@langchain/langgraph").LastValue<number[][]>;
|
|
109
|
-
(annotation: import("@langchain/langgraph").SingleReducer<number[][], number[][]>): import("@langchain/langgraph").BinaryOperatorAggregate<number[][], number[][]>;
|
|
110
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
111
|
-
};
|
|
112
|
-
}, {
|
|
113
|
-
filePath: {
|
|
114
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
115
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
116
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
117
|
-
};
|
|
118
|
-
mimeType: {
|
|
119
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
120
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
121
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
122
|
-
};
|
|
123
|
-
rawText: {
|
|
124
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
125
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
126
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
127
|
-
};
|
|
128
|
-
pdfChunks: {
|
|
129
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
130
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
131
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
132
|
-
};
|
|
133
|
-
markdownParts: import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
134
|
-
markdown: {
|
|
135
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
136
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
137
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
138
|
-
};
|
|
139
|
-
textChunks: {
|
|
140
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
141
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
142
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
143
|
-
};
|
|
144
|
-
vectors: {
|
|
145
|
-
(): import("@langchain/langgraph").LastValue<number[][]>;
|
|
146
|
-
(annotation: import("@langchain/langgraph").SingleReducer<number[][], number[][]>): import("@langchain/langgraph").BinaryOperatorAggregate<number[][], number[][]>;
|
|
147
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
148
|
-
};
|
|
149
|
-
}, import("@langchain/langgraph").StateDefinition>;
|
|
150
|
-
/**
|
|
151
|
-
* The compiled graph instance.
|
|
152
|
-
* Exported specifically for LangGraph Studio and the LangGraph CLI.
|
|
153
|
-
*/
|
|
154
|
-
export declare const graph: import("@langchain/langgraph").CompiledStateGraph<import("@langchain/langgraph").StateType<{
|
|
155
|
-
filePath: {
|
|
156
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
157
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
158
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
159
|
-
};
|
|
160
|
-
mimeType: {
|
|
161
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
162
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
163
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
164
|
-
};
|
|
165
|
-
rawText: {
|
|
166
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
167
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
168
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
169
|
-
};
|
|
170
|
-
pdfChunks: {
|
|
171
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
172
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
173
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
174
|
-
};
|
|
175
|
-
markdownParts: import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
176
|
-
markdown: {
|
|
177
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
178
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
179
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
180
|
-
};
|
|
181
|
-
textChunks: {
|
|
182
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
183
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
184
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
185
|
-
};
|
|
186
|
-
vectors: {
|
|
187
|
-
(): import("@langchain/langgraph").LastValue<number[][]>;
|
|
188
|
-
(annotation: import("@langchain/langgraph").SingleReducer<number[][], number[][]>): import("@langchain/langgraph").BinaryOperatorAggregate<number[][], number[][]>;
|
|
189
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
190
|
-
};
|
|
191
|
-
}>, import("@langchain/langgraph").UpdateType<{
|
|
192
|
-
filePath: {
|
|
193
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
194
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
195
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
196
|
-
};
|
|
197
|
-
mimeType: {
|
|
198
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
199
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
200
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
201
|
-
};
|
|
202
|
-
rawText: {
|
|
203
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
204
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
205
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
206
|
-
};
|
|
207
|
-
pdfChunks: {
|
|
208
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
209
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
210
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
211
|
-
};
|
|
212
|
-
markdownParts: import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
213
|
-
markdown: {
|
|
214
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
215
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
216
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
217
|
-
};
|
|
218
|
-
textChunks: {
|
|
219
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
220
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
221
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
222
|
-
};
|
|
223
|
-
vectors: {
|
|
224
|
-
(): import("@langchain/langgraph").LastValue<number[][]>;
|
|
225
|
-
(annotation: import("@langchain/langgraph").SingleReducer<number[][], number[][]>): import("@langchain/langgraph").BinaryOperatorAggregate<number[][], number[][]>;
|
|
226
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
227
|
-
};
|
|
228
|
-
}>, "markdownMerger" | "markdownNormalizer" | "llmExtractionNode" | "__start__" | "fileTypeRouter" | "libreOfficeToPdf" | "pdfSplitter" | "textExtractorNode" | "saveMarkdown" | "markdownChunker" | "vectorEmbedderNode" | "vectorUpsertNode", {
|
|
229
|
-
filePath: {
|
|
230
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
231
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
232
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
233
|
-
};
|
|
234
|
-
mimeType: {
|
|
235
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
236
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
237
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
238
|
-
};
|
|
239
|
-
rawText: {
|
|
240
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
241
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
242
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
243
|
-
};
|
|
244
|
-
pdfChunks: {
|
|
245
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
246
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
247
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
248
|
-
};
|
|
249
|
-
markdownParts: import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
250
|
-
markdown: {
|
|
251
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
252
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
253
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
254
|
-
};
|
|
255
|
-
textChunks: {
|
|
256
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
257
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
258
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
259
|
-
};
|
|
260
|
-
vectors: {
|
|
261
|
-
(): import("@langchain/langgraph").LastValue<number[][]>;
|
|
262
|
-
(annotation: import("@langchain/langgraph").SingleReducer<number[][], number[][]>): import("@langchain/langgraph").BinaryOperatorAggregate<number[][], number[][]>;
|
|
263
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
264
|
-
};
|
|
265
|
-
}, {
|
|
266
|
-
filePath: {
|
|
267
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
268
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
269
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
270
|
-
};
|
|
271
|
-
mimeType: {
|
|
272
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
273
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
274
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
275
|
-
};
|
|
276
|
-
rawText: {
|
|
277
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
278
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
279
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
280
|
-
};
|
|
281
|
-
pdfChunks: {
|
|
282
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
283
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
284
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
285
|
-
};
|
|
286
|
-
markdownParts: import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
287
|
-
markdown: {
|
|
288
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
289
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
290
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
291
|
-
};
|
|
292
|
-
textChunks: {
|
|
293
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
294
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
295
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
296
|
-
};
|
|
297
|
-
vectors: {
|
|
298
|
-
(): import("@langchain/langgraph").LastValue<number[][]>;
|
|
299
|
-
(annotation: import("@langchain/langgraph").SingleReducer<number[][], number[][]>): import("@langchain/langgraph").BinaryOperatorAggregate<number[][], number[][]>;
|
|
300
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
301
|
-
};
|
|
302
|
-
}, import("@langchain/langgraph").StateDefinition>;
|
|
303
|
-
//# sourceMappingURL=pipeline.d.ts.map
|
package/dist/pipeline.d.ts.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"pipeline.d.ts","sourceRoot":"","sources":["../src/pipeline.ts"],"names":[],"mappings":"AA0CA,wBAAgB,aAAa;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;mDA4D5B;AAED;;;GAGG;AACH,eAAO,MAAM,KAAK;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;kDAAkB,CAAC"}
|
package/dist/pipeline.js
DELETED
|
@@ -1,93 +0,0 @@
|
|
|
1
|
-
import { StateGraph, END, Send } from "@langchain/langgraph";
|
|
2
|
-
import { PipelineStateAnnotation } from "./state.js";
|
|
3
|
-
import { fileTypeRouter, routeByMimeType } from "./nodes/fileTypeRouter.js";
|
|
4
|
-
import { pdfSplitter } from "./nodes/pdfSplitter.js";
|
|
5
|
-
import { llmExtractionNode, routeAfterLlm } from "./nodes/llmExtractionNode.js";
|
|
6
|
-
import { markdownMerger } from "./nodes/markdownMerger.js";
|
|
7
|
-
import { textExtractorNode } from "./nodes/textExtractorNode.js";
|
|
8
|
-
import { markdownNormalizer } from "./nodes/markdownNormalizer.js";
|
|
9
|
-
import { markdownChunker } from "./nodes/markdownChunker.js";
|
|
10
|
-
import { vectorEmbedderNode } from "./nodes/vectorEmbedderNode.js";
|
|
11
|
-
import { vectorUpsertNode } from "./nodes/vectorUpsertNode.js";
|
|
12
|
-
import { saveMarkdown } from "./nodes/saveMarkdown.js";
|
|
13
|
-
import { libreOfficeToPdf } from "./nodes/libreOfficeToPdf.js";
|
|
14
|
-
/**
|
|
15
|
-
* Builds and compiles the RAG ingestion pipeline as a LangGraph StateGraph.
|
|
16
|
-
*
|
|
17
|
-
* Flow:
|
|
18
|
-
* START → fileTypeRouter
|
|
19
|
-
* ├─ "pdf" → pdfSplitter → [llmExtractionNode (Parallel)] → markdownMerger → markdownNormalizer
|
|
20
|
-
* ├─ "convert" → libreOfficeToPdf → pdfSplitter → (same as pdf branch)
|
|
21
|
-
* └─ "extract" → textExtractorNode → llmExtractionNode → markdownNormalizer
|
|
22
|
-
* markdownNormalizer → saveMarkdown → markdownChunker → vectorEmbedderNode → vectorUpsertNode → END
|
|
23
|
-
*/
|
|
24
|
-
/**
|
|
25
|
-
* Returns an array of 'Send' objects to process each PDF chunk in parallel.
|
|
26
|
-
*/
|
|
27
|
-
function dispatchPdfChunks(state) {
|
|
28
|
-
if (!state.pdfChunks || state.pdfChunks.length === 0) {
|
|
29
|
-
console.warn("[dispatchPdfChunks] No PDF chunks found to process.");
|
|
30
|
-
return [];
|
|
31
|
-
}
|
|
32
|
-
return state.pdfChunks.map((chunk, index) => {
|
|
33
|
-
return new Send("llmExtractionNode", {
|
|
34
|
-
chunk,
|
|
35
|
-
index,
|
|
36
|
-
totalChunks: state.pdfChunks.length,
|
|
37
|
-
});
|
|
38
|
-
});
|
|
39
|
-
}
|
|
40
|
-
export function buildPipeline() {
|
|
41
|
-
const graph = new StateGraph(PipelineStateAnnotation)
|
|
42
|
-
// ── Phase 1: Routing ──
|
|
43
|
-
.addNode("fileTypeRouter", fileTypeRouter)
|
|
44
|
-
// ── Phase 2a: PDF Branch ──
|
|
45
|
-
.addNode("libreOfficeToPdf", libreOfficeToPdf)
|
|
46
|
-
.addNode("pdfSplitter", pdfSplitter)
|
|
47
|
-
.addNode("markdownMerger", markdownMerger)
|
|
48
|
-
// ── Phase 2b: Text / Data Extraction Branch ──
|
|
49
|
-
.addNode("textExtractorNode", textExtractorNode)
|
|
50
|
-
.addNode("llmExtractionNode", llmExtractionNode)
|
|
51
|
-
// ── Phase 3: Normalization & Chunking ──
|
|
52
|
-
.addNode("markdownNormalizer", markdownNormalizer)
|
|
53
|
-
.addNode("saveMarkdown", saveMarkdown)
|
|
54
|
-
.addNode("markdownChunker", markdownChunker)
|
|
55
|
-
// ── Phase 4: Embedding & Indexing ──
|
|
56
|
-
.addNode("vectorEmbedderNode", vectorEmbedderNode)
|
|
57
|
-
.addNode("vectorUpsertNode", vectorUpsertNode)
|
|
58
|
-
// ── Edges ──
|
|
59
|
-
// Start → Router
|
|
60
|
-
.addEdge("__start__", "fileTypeRouter")
|
|
61
|
-
// Router → conditional branch
|
|
62
|
-
.addConditionalEdges("fileTypeRouter", routeByMimeType, {
|
|
63
|
-
pdf: "pdfSplitter",
|
|
64
|
-
convert: "libreOfficeToPdf",
|
|
65
|
-
extract: "textExtractorNode",
|
|
66
|
-
})
|
|
67
|
-
// Convert branch: LibreOffice → pdfSplitter → (joins PDF branch)
|
|
68
|
-
.addEdge("libreOfficeToPdf", "pdfSplitter")
|
|
69
|
-
// PDF branch dispatcher
|
|
70
|
-
.addConditionalEdges("pdfSplitter", dispatchPdfChunks, ["llmExtractionNode"])
|
|
71
|
-
// Unified Document/Text branch flow
|
|
72
|
-
.addEdge("textExtractorNode", "llmExtractionNode")
|
|
73
|
-
// After llmExtractionNode, conditionally merge PDF chunks or normalize Text
|
|
74
|
-
.addConditionalEdges("llmExtractionNode", routeAfterLlm, {
|
|
75
|
-
markdownMerger: "markdownMerger",
|
|
76
|
-
markdownNormalizer: "markdownNormalizer",
|
|
77
|
-
})
|
|
78
|
-
// If PDF branch, finish merger
|
|
79
|
-
.addEdge("markdownMerger", "markdownNormalizer")
|
|
80
|
-
// Shared tail: normalize → save → chunk → embed → upsert → end
|
|
81
|
-
.addEdge("markdownNormalizer", "saveMarkdown")
|
|
82
|
-
.addEdge("saveMarkdown", "markdownChunker")
|
|
83
|
-
.addEdge("markdownChunker", "vectorEmbedderNode")
|
|
84
|
-
.addEdge("vectorEmbedderNode", "vectorUpsertNode")
|
|
85
|
-
.addEdge("vectorUpsertNode", END);
|
|
86
|
-
return graph.compile();
|
|
87
|
-
}
|
|
88
|
-
/**
|
|
89
|
-
* The compiled graph instance.
|
|
90
|
-
* Exported specifically for LangGraph Studio and the LangGraph CLI.
|
|
91
|
-
*/
|
|
92
|
-
export const graph = buildPipeline();
|
|
93
|
-
//# sourceMappingURL=pipeline.js.map
|
package/dist/pipeline.js.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"pipeline.js","sourceRoot":"","sources":["../src/pipeline.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,UAAU,EAAE,GAAG,EAAE,IAAI,EAAE,MAAM,sBAAsB,CAAC;AAC7D,OAAO,EAAE,uBAAuB,EAAsB,MAAM,YAAY,CAAC;AACzE,OAAO,EAAE,cAAc,EAAE,eAAe,EAAE,MAAM,2BAA2B,CAAC;AAC5E,OAAO,EAAE,WAAW,EAAE,MAAM,wBAAwB,CAAC;AACrD,OAAO,EAAE,iBAAiB,EAAE,aAAa,EAAE,MAAM,8BAA8B,CAAC;AAChF,OAAO,EAAE,cAAc,EAAE,MAAM,2BAA2B,CAAC;AAE3D,OAAO,EAAE,iBAAiB,EAAE,MAAM,8BAA8B,CAAC;AACjE,OAAO,EAAE,kBAAkB,EAAE,MAAM,+BAA+B,CAAC;AACnE,OAAO,EAAE,eAAe,EAAE,MAAM,4BAA4B,CAAC;AAC7D,OAAO,EAAE,kBAAkB,EAAE,MAAM,+BAA+B,CAAC;AACnE,OAAO,EAAE,gBAAgB,EAAE,MAAM,6BAA6B,CAAC;AAC/D,OAAO,EAAE,YAAY,EAAE,MAAM,yBAAyB,CAAC;AACvD,OAAO,EAAE,gBAAgB,EAAE,MAAM,6BAA6B,CAAC;AAE/D;;;;;;;;;GASG;AAEH;;GAEG;AACH,SAAS,iBAAiB,CAAC,KAAoB;IAC7C,IAAI,CAAC,KAAK,CAAC,SAAS,IAAI,KAAK,CAAC,SAAS,CAAC,MAAM,KAAK,CAAC,EAAE,CAAC;QACrD,OAAO,CAAC,IAAI,CAAC,qDAAqD,CAAC,CAAC;QACpE,OAAO,EAAE,CAAC;IACZ,CAAC;IACD,OAAO,KAAK,CAAC,SAAS,CAAC,GAAG,CAAC,CAAC,KAAK,EAAE,KAAK,EAAE,EAAE;QAC1C,OAAO,IAAI,IAAI,CAAC,mBAAmB,EAAE;YACnC,KAAK;YACL,KAAK;YACL,WAAW,EAAE,KAAK,CAAC,SAAS,CAAC,MAAM;SACpC,CAAC,CAAC;IACL,CAAC,CAAC,CAAC;AACL,CAAC;AACD,MAAM,UAAU,aAAa;IAC3B,MAAM,KAAK,GAAG,IAAI,UAAU,CAAC,uBAAuB,CAAC;QACnD,yBAAyB;SACxB,OAAO,CAAC,gBAAgB,EAAE,cAAc,CAAC;QAE1C,6BAA6B;SAC5B,OAAO,CAAC,kBAAkB,EAAE,gBAAgB,CAAC;SAC7C,OAAO,CAAC,aAAa,EAAE,WAAW,CAAC;SACnC,OAAO,CAAC,gBAAgB,EAAE,cAAc,CAAC;QAE1C,gDAAgD;SAC/C,OAAO,CAAC,mBAAmB,EAAE,iBAAiB,CAAC;SAC/C,OAAO,CAAC,mBAAmB,EAAE,iBAAiB,CAAC;QAEhD,0CAA0C;SACzC,OAAO,CAAC,oBAAoB,EAAE,kBAAkB,CAAC;SACjD,OAAO,CAAC,cAAc,EAAE,YAAY,CAAC;SACrC,OAAO,CAAC,iBAAiB,EAAE,eAAe,CAAC;QAE5C,sCAAsC;SACrC,OAAO,CAAC,oBAAoB,EAAE,kBAAkB,CAAC;SACjD,OAAO,CAAC,kBAAkB,EAAE,gBAAgB,CAAC;QAE9C,cAAc;QACd,iBAAiB;SAChB,OAAO,CAAC,WAAW,EAAE,gBAAgB,CAAC;QAEvC,8BAA8B;SAC7B,mBAAmB,CAAC,gBAAgB,EAAE,eAAe,EAAE;QACtD,GAAG,EAAE,aAAa;QAClB,OAAO,EAAE,kBAAkB;QAC3B,OAAO,EAAE,mBAAmB;KAC7B,CAAC;QAEF,iEAAiE;SAChE,OAAO,CAAC,kBAAkB,EAAE,aAAa,CAAC;QAE3C,wBAAwB;SACvB,mBAAmB,CAAC,aAAa,EAAE,iBAAiB,EAAE,CAAC,mBAAmB,CAAC,CAAC;QAE7E,oCAAoC;SACnC,OAAO,CAAC,mBAAmB,EAAE,mBAAmB,CAAC;QAElD,4EAA4E;SAC3E,mBAAmB,CAAC,mBAAmB,EAAE,aAAa,EAAE;QACvD,cAAc,EAAE,gBAAgB;QAChC,kBAAkB,EAAE,oBAAoB;KACzC,CAAC;QAEF,+BAA+B;SAC9B,OAAO,CAAC,gBAAgB,EAAE,oBAAoB,CAAC;QAEhD,+DAA+D;SAC9D,OAAO,CAAC,oBAAoB,EAAE,cAAc,CAAC;SAC7C,OAAO,CAAC,cAAc,EAAE,iBAAiB,CAAC;SAC1C,OAAO,CAAC,iBAAiB,EAAE,oBAAoB,CAAC;SAChD,OAAO,CAAC,oBAAoB,EAAE,kBAAkB,CAAC;SACjD,OAAO,CAAC,kBAAkB,EAAE,GAAG,CAAC,CAAC;IAEpC,OAAO,KAAK,CAAC,OAAO,EAAE,CAAC;AACzB,CAAC;AAED;;;GAGG;AACH,MAAM,CAAC,MAAM,KAAK,GAAG,aAAa,EAAE,CAAC"}
|
package/dist/state.d.ts
DELETED
|
@@ -1,52 +0,0 @@
|
|
|
1
|
-
/**
|
|
2
|
-
* LangGraph pipeline state definition.
|
|
3
|
-
* Every node reads from and writes to this shared state.
|
|
4
|
-
*/
|
|
5
|
-
export declare const PipelineStateAnnotation: import("@langchain/langgraph").AnnotationRoot<{
|
|
6
|
-
/** Absolute path to the input file (optional if rawText is provided) */
|
|
7
|
-
filePath: {
|
|
8
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
9
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
10
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
11
|
-
};
|
|
12
|
-
/** Detected MIME type of the input file (optional if rawText is provided) */
|
|
13
|
-
mimeType: {
|
|
14
|
-
(): import("@langchain/langgraph").LastValue<string | undefined>;
|
|
15
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string | undefined, string | undefined>): import("@langchain/langgraph").BinaryOperatorAggregate<string | undefined, string | undefined>;
|
|
16
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
17
|
-
};
|
|
18
|
-
/** Extracted raw text (office / text branch) */
|
|
19
|
-
rawText: {
|
|
20
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
21
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
22
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
23
|
-
};
|
|
24
|
-
/** 10-page PDF chunk buffers (base64 strings, PDF branch) */
|
|
25
|
-
pdfChunks: {
|
|
26
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
27
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
28
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
29
|
-
};
|
|
30
|
-
/** Per-chunk markdown outputs from Gemini (PDF branch) */
|
|
31
|
-
markdownParts: import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
32
|
-
/** Final merged / extracted markdown (both branches converge here) */
|
|
33
|
-
markdown: {
|
|
34
|
-
(): import("@langchain/langgraph").LastValue<string>;
|
|
35
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string, string>): import("@langchain/langgraph").BinaryOperatorAggregate<string, string>;
|
|
36
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
37
|
-
};
|
|
38
|
-
/** Semantic text chunks after splitting */
|
|
39
|
-
textChunks: {
|
|
40
|
-
(): import("@langchain/langgraph").LastValue<string[]>;
|
|
41
|
-
(annotation: import("@langchain/langgraph").SingleReducer<string[], string[]>): import("@langchain/langgraph").BinaryOperatorAggregate<string[], string[]>;
|
|
42
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
43
|
-
};
|
|
44
|
-
/** OpenAI embedding vectors, one per text chunk */
|
|
45
|
-
vectors: {
|
|
46
|
-
(): import("@langchain/langgraph").LastValue<number[][]>;
|
|
47
|
-
(annotation: import("@langchain/langgraph").SingleReducer<number[][], number[][]>): import("@langchain/langgraph").BinaryOperatorAggregate<number[][], number[][]>;
|
|
48
|
-
Root: <S extends import("@langchain/langgraph").StateDefinition>(sd: S) => import("@langchain/langgraph").AnnotationRoot<S>;
|
|
49
|
-
};
|
|
50
|
-
}>;
|
|
51
|
-
export type PipelineState = typeof PipelineStateAnnotation.State;
|
|
52
|
-
//# sourceMappingURL=state.d.ts.map
|
package/dist/state.d.ts.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"state.d.ts","sourceRoot":"","sources":["../src/state.ts"],"names":[],"mappings":"AAEA;;;GAGG;AACH,eAAO,MAAM,uBAAuB;IAClC,wEAAwE;;;;;;IAGxE,6EAA6E;;;;;;IAG7E,gDAAgD;;;;;;IAGhD,6DAA6D;;;;;;IAG7D,0DAA0D;;IAM1D,sEAAsE;;;;;;IAGtE,2CAA2C;;;;;;IAG3C,mDAAmD;;;;;;EAEnD,CAAC;AAEH,MAAM,MAAM,aAAa,GAAG,OAAO,uBAAuB,CAAC,KAAK,CAAC"}
|
package/dist/state.js
DELETED
|
@@ -1,27 +0,0 @@
|
|
|
1
|
-
import { Annotation } from "@langchain/langgraph";
|
|
2
|
-
/**
|
|
3
|
-
* LangGraph pipeline state definition.
|
|
4
|
-
* Every node reads from and writes to this shared state.
|
|
5
|
-
*/
|
|
6
|
-
export const PipelineStateAnnotation = Annotation.Root({
|
|
7
|
-
/** Absolute path to the input file (optional if rawText is provided) */
|
|
8
|
-
filePath: (Annotation),
|
|
9
|
-
/** Detected MIME type of the input file (optional if rawText is provided) */
|
|
10
|
-
mimeType: (Annotation),
|
|
11
|
-
/** Extracted raw text (office / text branch) */
|
|
12
|
-
rawText: (Annotation),
|
|
13
|
-
/** 10-page PDF chunk buffers (base64 strings, PDF branch) */
|
|
14
|
-
pdfChunks: (Annotation),
|
|
15
|
-
/** Per-chunk markdown outputs from Gemini (PDF branch) */
|
|
16
|
-
markdownParts: Annotation({
|
|
17
|
-
reducer: (x, y) => x.concat(y),
|
|
18
|
-
default: () => [],
|
|
19
|
-
}),
|
|
20
|
-
/** Final merged / extracted markdown (both branches converge here) */
|
|
21
|
-
markdown: (Annotation),
|
|
22
|
-
/** Semantic text chunks after splitting */
|
|
23
|
-
textChunks: (Annotation),
|
|
24
|
-
/** OpenAI embedding vectors, one per text chunk */
|
|
25
|
-
vectors: (Annotation),
|
|
26
|
-
});
|
|
27
|
-
//# sourceMappingURL=state.js.map
|
package/dist/state.js.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"state.js","sourceRoot":"","sources":["../src/state.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,UAAU,EAAE,MAAM,sBAAsB,CAAC;AAElD;;;GAGG;AACH,MAAM,CAAC,MAAM,uBAAuB,GAAG,UAAU,CAAC,IAAI,CAAC;IACrD,wEAAwE;IACxE,QAAQ,EAAE,CAAA,UAA8B,CAAA;IAExC,6EAA6E;IAC7E,QAAQ,EAAE,CAAA,UAA8B,CAAA;IAExC,gDAAgD;IAChD,OAAO,EAAE,CAAA,UAAkB,CAAA;IAE3B,6DAA6D;IAC7D,SAAS,EAAE,CAAA,UAAoB,CAAA;IAE/B,0DAA0D;IAC1D,aAAa,EAAE,UAAU,CAAW;QAClC,OAAO,EAAE,CAAC,CAAC,EAAE,CAAC,EAAE,EAAE,CAAC,CAAC,CAAC,MAAM,CAAC,CAAC,CAAC;QAC9B,OAAO,EAAE,GAAG,EAAE,CAAC,EAAE;KAClB,CAAC;IAEF,sEAAsE;IACtE,QAAQ,EAAE,CAAA,UAAkB,CAAA;IAE5B,2CAA2C;IAC3C,UAAU,EAAE,CAAA,UAAoB,CAAA;IAEhC,mDAAmD;IACnD,OAAO,EAAE,CAAA,UAAsB,CAAA;CAChC,CAAC,CAAC"}
|
package/dist/vectorStore.d.ts
DELETED
|
@@ -1,24 +0,0 @@
|
|
|
1
|
-
/**
|
|
2
|
-
* The standard shape of a record that the pipeline will produce.
|
|
3
|
-
*/
|
|
4
|
-
export interface VectorRecord {
|
|
5
|
-
id: string;
|
|
6
|
-
vector: number[];
|
|
7
|
-
metadata: Record<string, any>;
|
|
8
|
-
}
|
|
9
|
-
/**
|
|
10
|
-
* The contract that any vector database adapter must follow.
|
|
11
|
-
*/
|
|
12
|
-
export interface VectorStoreAdapter {
|
|
13
|
-
upsert(records: VectorRecord[]): Promise<void>;
|
|
14
|
-
}
|
|
15
|
-
/**
|
|
16
|
-
* Built-in adapter for Upstash Vector.
|
|
17
|
-
* Used by default when running via the CLI.
|
|
18
|
-
*/
|
|
19
|
-
export declare class UpstashAdapter implements VectorStoreAdapter {
|
|
20
|
-
private index;
|
|
21
|
-
constructor(url: string, token: string);
|
|
22
|
-
upsert(records: VectorRecord[]): Promise<void>;
|
|
23
|
-
}
|
|
24
|
-
//# sourceMappingURL=vectorStore.d.ts.map
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"vectorStore.d.ts","sourceRoot":"","sources":["../src/vectorStore.ts"],"names":[],"mappings":"AAEA;;GAEG;AACH,MAAM,WAAW,YAAY;IAC3B,EAAE,EAAE,MAAM,CAAC;IACX,MAAM,EAAE,MAAM,EAAE,CAAC;IACjB,QAAQ,EAAE,MAAM,CAAC,MAAM,EAAE,GAAG,CAAC,CAAC;CAC/B;AAED;;GAEG;AACH,MAAM,WAAW,kBAAkB;IACjC,MAAM,CAAC,OAAO,EAAE,YAAY,EAAE,GAAG,OAAO,CAAC,IAAI,CAAC,CAAC;CAChD;AAED;;;GAGG;AACH,qBAAa,cAAe,YAAW,kBAAkB;IACvD,OAAO,CAAC,KAAK,CAAQ;gBAET,GAAG,EAAE,MAAM,EAAE,KAAK,EAAE,MAAM;IAIhC,MAAM,CAAC,OAAO,EAAE,YAAY,EAAE,GAAG,OAAO,CAAC,IAAI,CAAC;CAWrD"}
|
package/dist/vectorStore.js
DELETED
|
@@ -1,22 +0,0 @@
|
|
|
1
|
-
import { Index } from "@upstash/vector";
|
|
2
|
-
/**
|
|
3
|
-
* Built-in adapter for Upstash Vector.
|
|
4
|
-
* Used by default when running via the CLI.
|
|
5
|
-
*/
|
|
6
|
-
export class UpstashAdapter {
|
|
7
|
-
index;
|
|
8
|
-
constructor(url, token) {
|
|
9
|
-
this.index = new Index({ url, token });
|
|
10
|
-
}
|
|
11
|
-
async upsert(records) {
|
|
12
|
-
const upstashRecords = records.map((r) => ({
|
|
13
|
-
id: r.id,
|
|
14
|
-
vector: r.vector,
|
|
15
|
-
metadata: r.metadata,
|
|
16
|
-
// For Upstash, the string payload goes in 'data' usually, but metadata is fine.
|
|
17
|
-
data: r.metadata.text || "",
|
|
18
|
-
}));
|
|
19
|
-
await this.index.upsert(upstashRecords);
|
|
20
|
-
}
|
|
21
|
-
}
|
|
22
|
-
//# sourceMappingURL=vectorStore.js.map
|
package/dist/vectorStore.js.map
DELETED
|
@@ -1 +0,0 @@
|
|
|
1
|
-
{"version":3,"file":"vectorStore.js","sourceRoot":"","sources":["../src/vectorStore.ts"],"names":[],"mappings":"AAAA,OAAO,EAAE,KAAK,EAAE,MAAM,iBAAiB,CAAC;AAkBxC;;;GAGG;AACH,MAAM,OAAO,cAAc;IACjB,KAAK,CAAQ;IAErB,YAAY,GAAW,EAAE,KAAa;QACpC,IAAI,CAAC,KAAK,GAAG,IAAI,KAAK,CAAC,EAAE,GAAG,EAAE,KAAK,EAAE,CAAC,CAAC;IACzC,CAAC;IAED,KAAK,CAAC,MAAM,CAAC,OAAuB;QAClC,MAAM,cAAc,GAAG,OAAO,CAAC,GAAG,CAAC,CAAC,CAAC,EAAE,EAAE,CAAC,CAAC;YACzC,EAAE,EAAE,CAAC,CAAC,EAAE;YACR,MAAM,EAAE,CAAC,CAAC,MAAM;YAChB,QAAQ,EAAE,CAAC,CAAC,QAAQ;YACpB,gFAAgF;YAChF,IAAI,EAAE,CAAC,CAAC,QAAQ,CAAC,IAAI,IAAI,EAAE;SAC5B,CAAC,CAAC,CAAC;QAEJ,MAAM,IAAI,CAAC,KAAK,CAAC,MAAM,CAAC,cAAc,CAAC,CAAC;IAC1C,CAAC;CACF"}
|