kreuzberg 4.0.0.pre.rc.6 → 4.0.0.rc1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (175) hide show
  1. checksums.yaml +4 -4
  2. data/.gitignore +0 -6
  3. data/.rubocop.yaml +534 -1
  4. data/Gemfile +2 -1
  5. data/Gemfile.lock +11 -11
  6. data/README.md +5 -10
  7. data/examples/async_patterns.rb +0 -1
  8. data/ext/kreuzberg_rb/extconf.rb +0 -10
  9. data/ext/kreuzberg_rb/native/Cargo.toml +15 -23
  10. data/ext/kreuzberg_rb/native/build.rs +2 -0
  11. data/ext/kreuzberg_rb/native/include/ieeefp.h +1 -1
  12. data/ext/kreuzberg_rb/native/include/msvc_compat/strings.h +1 -1
  13. data/ext/kreuzberg_rb/native/include/strings.h +2 -2
  14. data/ext/kreuzberg_rb/native/include/unistd.h +1 -1
  15. data/ext/kreuzberg_rb/native/src/lib.rs +16 -75
  16. data/kreuzberg.gemspec +14 -57
  17. data/lib/kreuzberg/cache_api.rb +0 -1
  18. data/lib/kreuzberg/cli.rb +2 -2
  19. data/lib/kreuzberg/config.rb +2 -9
  20. data/lib/kreuzberg/errors.rb +7 -75
  21. data/lib/kreuzberg/extraction_api.rb +0 -1
  22. data/lib/kreuzberg/setup_lib_path.rb +0 -1
  23. data/lib/kreuzberg/version.rb +1 -1
  24. data/lib/kreuzberg.rb +0 -21
  25. data/pkg/kreuzberg-4.0.0.rc1.gem +0 -0
  26. data/sig/kreuzberg.rbs +3 -55
  27. data/spec/binding/cli_proxy_spec.rb +4 -2
  28. data/spec/binding/cli_spec.rb +11 -12
  29. data/spec/examples.txt +104 -0
  30. data/spec/fixtures/config.yaml +1 -0
  31. data/spec/spec_helper.rb +1 -1
  32. data/vendor/kreuzberg/Cargo.toml +42 -112
  33. data/vendor/kreuzberg/README.md +2 -2
  34. data/vendor/kreuzberg/build.rs +4 -18
  35. data/vendor/kreuzberg/src/bin/profile_extract.rs +455 -0
  36. data/vendor/kreuzberg/src/cache/mod.rs +3 -27
  37. data/vendor/kreuzberg/src/core/batch_mode.rs +0 -60
  38. data/vendor/kreuzberg/src/core/extractor.rs +81 -202
  39. data/vendor/kreuzberg/src/core/io.rs +2 -4
  40. data/vendor/kreuzberg/src/core/mime.rs +12 -2
  41. data/vendor/kreuzberg/src/core/mod.rs +1 -4
  42. data/vendor/kreuzberg/src/core/pipeline.rs +33 -111
  43. data/vendor/kreuzberg/src/embeddings.rs +16 -125
  44. data/vendor/kreuzberg/src/error.rs +1 -1
  45. data/vendor/kreuzberg/src/extraction/docx.rs +1 -1
  46. data/vendor/kreuzberg/src/extraction/image.rs +13 -13
  47. data/vendor/kreuzberg/src/extraction/libreoffice.rs +1 -0
  48. data/vendor/kreuzberg/src/extraction/mod.rs +5 -9
  49. data/vendor/kreuzberg/src/extraction/office_metadata/mod.rs +0 -2
  50. data/vendor/kreuzberg/src/extraction/pandoc/batch.rs +275 -0
  51. data/vendor/kreuzberg/src/extraction/pandoc/mime_types.rs +178 -0
  52. data/vendor/kreuzberg/src/extraction/pandoc/mod.rs +491 -0
  53. data/vendor/kreuzberg/src/extraction/pandoc/server.rs +496 -0
  54. data/vendor/kreuzberg/src/extraction/pandoc/subprocess.rs +1188 -0
  55. data/vendor/kreuzberg/src/extraction/pandoc/version.rs +162 -0
  56. data/vendor/kreuzberg/src/extractors/archive.rs +0 -21
  57. data/vendor/kreuzberg/src/extractors/docx.rs +128 -16
  58. data/vendor/kreuzberg/src/extractors/email.rs +0 -14
  59. data/vendor/kreuzberg/src/extractors/excel.rs +20 -19
  60. data/vendor/kreuzberg/src/extractors/html.rs +154 -137
  61. data/vendor/kreuzberg/src/extractors/image.rs +4 -7
  62. data/vendor/kreuzberg/src/extractors/mod.rs +9 -106
  63. data/vendor/kreuzberg/src/extractors/pandoc.rs +201 -0
  64. data/vendor/kreuzberg/src/extractors/pdf.rs +15 -12
  65. data/vendor/kreuzberg/src/extractors/pptx.rs +3 -17
  66. data/vendor/kreuzberg/src/extractors/structured.rs +0 -14
  67. data/vendor/kreuzberg/src/extractors/text.rs +5 -23
  68. data/vendor/kreuzberg/src/extractors/xml.rs +0 -7
  69. data/vendor/kreuzberg/src/keywords/rake.rs +1 -0
  70. data/vendor/kreuzberg/src/lib.rs +1 -4
  71. data/vendor/kreuzberg/src/mcp/mod.rs +1 -1
  72. data/vendor/kreuzberg/src/mcp/server.rs +3 -5
  73. data/vendor/kreuzberg/src/ocr/processor.rs +2 -18
  74. data/vendor/kreuzberg/src/pdf/error.rs +1 -1
  75. data/vendor/kreuzberg/src/pdf/table.rs +44 -17
  76. data/vendor/kreuzberg/src/pdf/text.rs +3 -0
  77. data/vendor/kreuzberg/src/plugins/extractor.rs +5 -8
  78. data/vendor/kreuzberg/src/plugins/ocr.rs +11 -2
  79. data/vendor/kreuzberg/src/plugins/processor.rs +1 -2
  80. data/vendor/kreuzberg/src/plugins/registry.rs +0 -13
  81. data/vendor/kreuzberg/src/plugins/validator.rs +8 -9
  82. data/vendor/kreuzberg/src/stopwords/mod.rs +2 -2
  83. data/vendor/kreuzberg/src/types.rs +12 -42
  84. data/vendor/kreuzberg/tests/batch_orchestration.rs +5 -19
  85. data/vendor/kreuzberg/tests/batch_processing.rs +3 -15
  86. data/vendor/kreuzberg/tests/chunking_offset_demo.rs +92 -0
  87. data/vendor/kreuzberg/tests/concurrency_stress.rs +1 -17
  88. data/vendor/kreuzberg/tests/config_features.rs +0 -18
  89. data/vendor/kreuzberg/tests/config_loading_tests.rs +39 -15
  90. data/vendor/kreuzberg/tests/core_integration.rs +7 -24
  91. data/vendor/kreuzberg/tests/csv_integration.rs +81 -71
  92. data/vendor/kreuzberg/tests/docx_metadata_extraction_test.rs +25 -23
  93. data/vendor/kreuzberg/tests/pandoc_integration.rs +503 -0
  94. data/vendor/kreuzberg/tests/pipeline_integration.rs +1 -0
  95. data/vendor/kreuzberg/tests/plugin_postprocessor_test.rs +1 -0
  96. data/vendor/kreuzberg/tests/registry_integration_tests.rs +22 -1
  97. data/vendor/kreuzberg/tests/security_validation.rs +1 -12
  98. metadata +25 -90
  99. data/.rubocop.yml +0 -538
  100. data/ext/kreuzberg_rb/native/Cargo.lock +0 -6535
  101. data/lib/kreuzberg/error_context.rb +0 -32
  102. data/vendor/kreuzberg/benches/otel_overhead.rs +0 -48
  103. data/vendor/kreuzberg/src/extraction/markdown.rs +0 -213
  104. data/vendor/kreuzberg/src/extraction/office_metadata/odt_properties.rs +0 -287
  105. data/vendor/kreuzberg/src/extractors/bibtex.rs +0 -469
  106. data/vendor/kreuzberg/src/extractors/docbook.rs +0 -502
  107. data/vendor/kreuzberg/src/extractors/epub.rs +0 -707
  108. data/vendor/kreuzberg/src/extractors/fictionbook.rs +0 -491
  109. data/vendor/kreuzberg/src/extractors/fictionbook.rs.backup2 +0 -738
  110. data/vendor/kreuzberg/src/extractors/jats.rs +0 -1051
  111. data/vendor/kreuzberg/src/extractors/jupyter.rs +0 -367
  112. data/vendor/kreuzberg/src/extractors/latex.rs +0 -652
  113. data/vendor/kreuzberg/src/extractors/markdown.rs +0 -700
  114. data/vendor/kreuzberg/src/extractors/odt.rs +0 -628
  115. data/vendor/kreuzberg/src/extractors/opml.rs +0 -634
  116. data/vendor/kreuzberg/src/extractors/orgmode.rs +0 -528
  117. data/vendor/kreuzberg/src/extractors/rst.rs +0 -576
  118. data/vendor/kreuzberg/src/extractors/rtf.rs +0 -810
  119. data/vendor/kreuzberg/src/extractors/security.rs +0 -484
  120. data/vendor/kreuzberg/src/extractors/security_tests.rs +0 -367
  121. data/vendor/kreuzberg/src/extractors/typst.rs +0 -650
  122. data/vendor/kreuzberg/src/panic_context.rs +0 -154
  123. data/vendor/kreuzberg/tests/api_extract_multipart.rs +0 -52
  124. data/vendor/kreuzberg/tests/bibtex_parity_test.rs +0 -421
  125. data/vendor/kreuzberg/tests/docbook_extractor_tests.rs +0 -498
  126. data/vendor/kreuzberg/tests/docx_vs_pandoc_comparison.rs +0 -370
  127. data/vendor/kreuzberg/tests/epub_native_extractor_tests.rs +0 -275
  128. data/vendor/kreuzberg/tests/fictionbook_extractor_tests.rs +0 -228
  129. data/vendor/kreuzberg/tests/html_table_test.rs +0 -551
  130. data/vendor/kreuzberg/tests/instrumentation_test.rs +0 -139
  131. data/vendor/kreuzberg/tests/jats_extractor_tests.rs +0 -639
  132. data/vendor/kreuzberg/tests/jupyter_extractor_tests.rs +0 -704
  133. data/vendor/kreuzberg/tests/latex_extractor_tests.rs +0 -496
  134. data/vendor/kreuzberg/tests/markdown_extractor_tests.rs +0 -490
  135. data/vendor/kreuzberg/tests/odt_extractor_tests.rs +0 -695
  136. data/vendor/kreuzberg/tests/opml_extractor_tests.rs +0 -616
  137. data/vendor/kreuzberg/tests/orgmode_extractor_tests.rs +0 -822
  138. data/vendor/kreuzberg/tests/rst_extractor_tests.rs +0 -692
  139. data/vendor/kreuzberg/tests/rtf_extractor_tests.rs +0 -776
  140. data/vendor/kreuzberg/tests/typst_behavioral_tests.rs +0 -1259
  141. data/vendor/kreuzberg/tests/typst_extractor_tests.rs +0 -647
  142. data/vendor/rb-sys/.cargo-ok +0 -1
  143. data/vendor/rb-sys/.cargo_vcs_info.json +0 -6
  144. data/vendor/rb-sys/Cargo.lock +0 -393
  145. data/vendor/rb-sys/Cargo.toml +0 -70
  146. data/vendor/rb-sys/Cargo.toml.orig +0 -57
  147. data/vendor/rb-sys/LICENSE-APACHE +0 -190
  148. data/vendor/rb-sys/LICENSE-MIT +0 -21
  149. data/vendor/rb-sys/bin/release.sh +0 -21
  150. data/vendor/rb-sys/build/features.rs +0 -108
  151. data/vendor/rb-sys/build/main.rs +0 -246
  152. data/vendor/rb-sys/build/stable_api_config.rs +0 -153
  153. data/vendor/rb-sys/build/version.rs +0 -48
  154. data/vendor/rb-sys/readme.md +0 -36
  155. data/vendor/rb-sys/src/bindings.rs +0 -21
  156. data/vendor/rb-sys/src/hidden.rs +0 -11
  157. data/vendor/rb-sys/src/lib.rs +0 -34
  158. data/vendor/rb-sys/src/macros.rs +0 -371
  159. data/vendor/rb-sys/src/memory.rs +0 -53
  160. data/vendor/rb-sys/src/ruby_abi_version.rs +0 -38
  161. data/vendor/rb-sys/src/special_consts.rs +0 -31
  162. data/vendor/rb-sys/src/stable_api/compiled.c +0 -179
  163. data/vendor/rb-sys/src/stable_api/compiled.rs +0 -257
  164. data/vendor/rb-sys/src/stable_api/ruby_2_6.rs +0 -316
  165. data/vendor/rb-sys/src/stable_api/ruby_2_7.rs +0 -316
  166. data/vendor/rb-sys/src/stable_api/ruby_3_0.rs +0 -324
  167. data/vendor/rb-sys/src/stable_api/ruby_3_1.rs +0 -317
  168. data/vendor/rb-sys/src/stable_api/ruby_3_2.rs +0 -315
  169. data/vendor/rb-sys/src/stable_api/ruby_3_3.rs +0 -326
  170. data/vendor/rb-sys/src/stable_api/ruby_3_4.rs +0 -327
  171. data/vendor/rb-sys/src/stable_api.rs +0 -261
  172. data/vendor/rb-sys/src/symbol.rs +0 -31
  173. data/vendor/rb-sys/src/tracking_allocator.rs +0 -332
  174. data/vendor/rb-sys/src/utils.rs +0 -89
  175. data/vendor/rb-sys/src/value_type.rs +0 -7
@@ -1,707 +0,0 @@
1
- //! Native EPUB extractor using permissive-licensed dependencies.
2
- //!
3
- //! This extractor provides native Rust-based EPUB extraction without GPL-licensed
4
- //! dependencies, extracting:
5
- //! - Metadata from OPF (Open Packaging Format) using Dublin Core standards
6
- //! - Content from XHTML files in spine order
7
- //! - Proper handling of EPUB2 and EPUB3 formats
8
- //!
9
- //! Uses only permissive-licensed crates:
10
- //! - `zip` (MIT/Apache) - for reading EPUB container
11
- //! - `roxmltree` (MIT) - for parsing XML
12
- //! - `html-to-markdown-rs` (MIT) - for converting XHTML to plain text
13
-
14
- use crate::Result;
15
- use crate::core::config::ExtractionConfig;
16
- use crate::plugins::{DocumentExtractor, Plugin};
17
- use crate::types::{ExtractionResult, Metadata};
18
- use async_trait::async_trait;
19
- use roxmltree;
20
- use std::collections::BTreeMap;
21
- use std::io::Cursor;
22
- use zip::ZipArchive;
23
-
24
- /// EPUB format extractor using permissive-licensed dependencies.
25
- ///
26
- /// Extracts content and metadata from EPUB files (both EPUB2 and EPUB3)
27
- /// using native Rust parsing without GPL-licensed dependencies.
28
- pub struct EpubExtractor;
29
-
30
- impl EpubExtractor {
31
- /// Create a new EPUB extractor.
32
- pub fn new() -> Self {
33
- Self
34
- }
35
-
36
- /// Extract text content from an EPUB document by reading in spine order
37
- fn extract_content(
38
- archive: &mut ZipArchive<Cursor<Vec<u8>>>,
39
- opf_path: &str,
40
- manifest_dir: &str,
41
- ) -> Result<String> {
42
- let opf_xml = Self::read_file_from_zip(archive, opf_path)?;
43
- let (_, spine_hrefs) = Self::parse_opf(&opf_xml)?;
44
-
45
- let mut content = String::new();
46
-
47
- for (index, href) in spine_hrefs.iter().enumerate() {
48
- let file_path = Self::resolve_path(manifest_dir, href);
49
-
50
- match Self::read_file_from_zip(archive, &file_path) {
51
- Ok(xhtml_content) => {
52
- let text = Self::extract_text_from_xhtml(&xhtml_content);
53
- if !text.is_empty() {
54
- if index > 0 && !content.ends_with('\n') {
55
- content.push('\n');
56
- }
57
- content.push_str(&text);
58
- content.push('\n');
59
- }
60
- }
61
- Err(_) => {
62
- continue;
63
- }
64
- }
65
- }
66
-
67
- Ok(content.trim().to_string())
68
- }
69
-
70
- /// Extract text from XHTML content using html-to-markdown-rs
71
- fn extract_text_from_xhtml(xhtml: &str) -> String {
72
- match crate::extraction::html::convert_html_to_markdown(xhtml, None) {
73
- Ok(markdown) => {
74
- let text = Self::markdown_to_plain_text(&markdown);
75
- Self::remove_html_comments(&text)
76
- }
77
- Err(_) => Self::strip_html_tags(xhtml),
78
- }
79
- }
80
-
81
- /// Remove HTML comments from text
82
- fn remove_html_comments(text: &str) -> String {
83
- let mut result = String::new();
84
- let mut in_comment = false;
85
- let mut chars = text.chars().peekable();
86
-
87
- while let Some(ch) = chars.next() {
88
- if !in_comment && ch == '<' {
89
- if chars.peek() == Some(&'!') {
90
- chars.next();
91
- if chars.peek() == Some(&'-') {
92
- chars.next();
93
- if chars.peek() == Some(&'-') {
94
- chars.next();
95
- in_comment = true;
96
- continue;
97
- } else {
98
- result.push('<');
99
- result.push('!');
100
- result.push('-');
101
- continue;
102
- }
103
- } else {
104
- result.push('<');
105
- result.push('!');
106
- continue;
107
- }
108
- } else {
109
- result.push(ch);
110
- }
111
- } else if in_comment {
112
- if ch == '-' && chars.peek() == Some(&'-') {
113
- chars.next();
114
- if chars.peek() == Some(&'>') {
115
- chars.next();
116
- in_comment = false;
117
- result.push('\n');
118
- }
119
- }
120
- } else {
121
- result.push(ch);
122
- }
123
- }
124
-
125
- result
126
- }
127
-
128
- /// Convert markdown output to plain text by removing markdown syntax
129
- fn markdown_to_plain_text(markdown: &str) -> String {
130
- let mut text = String::new();
131
- let mut in_code_block = false;
132
-
133
- for line in markdown.lines() {
134
- let trimmed = line.trim();
135
-
136
- if trimmed.is_empty() {
137
- if !text.is_empty() && !text.ends_with('\n') {
138
- text.push('\n');
139
- }
140
- continue;
141
- }
142
-
143
- if trimmed.starts_with("```") {
144
- in_code_block = !in_code_block;
145
- continue;
146
- }
147
-
148
- if in_code_block {
149
- text.push_str(trimmed);
150
- text.push('\n');
151
- continue;
152
- }
153
-
154
- let cleaned = if let Some(stripped) = trimmed.strip_prefix("- ").or_else(|| trimmed.strip_prefix("* ")) {
155
- stripped
156
- } else if let Some(stripped) = trimmed.strip_prefix(|c: char| c.is_ascii_digit()) {
157
- if let Some(rest) = stripped.strip_prefix(". ") {
158
- rest
159
- } else {
160
- trimmed
161
- }
162
- } else {
163
- trimmed
164
- };
165
-
166
- let cleaned = cleaned.trim_start_matches('#').trim();
167
-
168
- let cleaned = cleaned
169
- .replace("**", "")
170
- .replace("__", "")
171
- .replace("*", "")
172
- .replace("_", "");
173
-
174
- let cleaned = Self::remove_markdown_links(&cleaned);
175
-
176
- if !cleaned.is_empty() {
177
- text.push_str(&cleaned);
178
- text.push('\n');
179
- }
180
- }
181
-
182
- text.trim().to_string()
183
- }
184
-
185
- /// Remove markdown links [text](url) -> text
186
- fn remove_markdown_links(text: &str) -> String {
187
- let mut result = String::new();
188
- let mut chars = text.chars().peekable();
189
-
190
- while let Some(ch) = chars.next() {
191
- if ch == '[' {
192
- let mut link_text = String::new();
193
- let mut depth = 1;
194
-
195
- while let Some(&next_ch) = chars.peek() {
196
- chars.next();
197
- if next_ch == '[' {
198
- depth += 1;
199
- link_text.push(next_ch);
200
- } else if next_ch == ']' {
201
- depth -= 1;
202
- if depth == 0 {
203
- break;
204
- }
205
- link_text.push(next_ch);
206
- } else {
207
- link_text.push(next_ch);
208
- }
209
- }
210
-
211
- if let Some(&'(') = chars.peek() {
212
- chars.next();
213
- let mut paren_depth = 1;
214
- while let Some(&next_ch) = chars.peek() {
215
- chars.next();
216
- if next_ch == '(' {
217
- paren_depth += 1;
218
- } else if next_ch == ')' {
219
- paren_depth -= 1;
220
- if paren_depth == 0 {
221
- break;
222
- }
223
- }
224
- }
225
- }
226
-
227
- result.push_str(&link_text);
228
- } else {
229
- result.push(ch);
230
- }
231
- }
232
-
233
- result
234
- }
235
-
236
- /// Fallback: strip HTML tags without using specialized libraries
237
- fn strip_html_tags(html: &str) -> String {
238
- let mut text = String::new();
239
- let mut in_tag = false;
240
- let mut in_script_style = false;
241
- let mut tag_name = String::new();
242
-
243
- for ch in html.chars() {
244
- if ch == '<' {
245
- in_tag = true;
246
- tag_name.clear();
247
- continue;
248
- }
249
-
250
- if ch == '>' {
251
- in_tag = false;
252
-
253
- let tag_lower = tag_name.to_lowercase();
254
- if tag_lower.contains("script") || tag_lower.contains("style") {
255
- in_script_style = !tag_name.starts_with('/');
256
- }
257
- continue;
258
- }
259
-
260
- if in_tag {
261
- tag_name.push(ch);
262
- continue;
263
- }
264
-
265
- if in_script_style {
266
- continue;
267
- }
268
-
269
- if ch == '\n' || ch == '\r' || ch == '\t' || ch == ' ' {
270
- if !text.is_empty() && !text.ends_with(' ') {
271
- text.push(' ');
272
- }
273
- } else {
274
- text.push(ch);
275
- }
276
- }
277
-
278
- let mut result = String::new();
279
- let mut prev_space = false;
280
- for ch in text.chars() {
281
- if ch == ' ' {
282
- if !prev_space {
283
- result.push(ch);
284
- }
285
- prev_space = true;
286
- } else {
287
- result.push(ch);
288
- prev_space = false;
289
- }
290
- }
291
-
292
- result.trim().to_string()
293
- }
294
-
295
- /// Extract metadata from EPUB OPF file
296
- fn extract_metadata(opf_xml: &str) -> Result<BTreeMap<String, serde_json::Value>> {
297
- let mut metadata = BTreeMap::new();
298
-
299
- let (epub_metadata, _) = Self::parse_opf(opf_xml)?;
300
-
301
- if let Some(title) = epub_metadata.title {
302
- metadata.insert("title".to_string(), serde_json::json!(title));
303
- }
304
-
305
- if let Some(creator) = epub_metadata.creator {
306
- metadata.insert("creator".to_string(), serde_json::json!(creator.clone()));
307
- metadata.insert("authors".to_string(), serde_json::json!(vec![creator]));
308
- }
309
-
310
- if let Some(date) = epub_metadata.date {
311
- metadata.insert("date".to_string(), serde_json::json!(date));
312
- }
313
-
314
- if let Some(language) = epub_metadata.language {
315
- metadata.insert("language".to_string(), serde_json::json!(language));
316
- }
317
-
318
- if let Some(identifier) = epub_metadata.identifier {
319
- metadata.insert("identifier".to_string(), serde_json::json!(identifier));
320
- }
321
-
322
- if let Some(publisher) = epub_metadata.publisher {
323
- metadata.insert("publisher".to_string(), serde_json::json!(publisher));
324
- }
325
-
326
- if let Some(subject) = epub_metadata.subject {
327
- metadata.insert("subject".to_string(), serde_json::json!(subject));
328
- }
329
-
330
- if let Some(description) = epub_metadata.description {
331
- metadata.insert("description".to_string(), serde_json::json!(description));
332
- }
333
-
334
- if let Some(rights) = epub_metadata.rights {
335
- metadata.insert("rights".to_string(), serde_json::json!(rights));
336
- }
337
-
338
- Ok(metadata)
339
- }
340
-
341
- /// Parse container.xml to find the OPF file path
342
- fn parse_container_xml(xml: &str) -> Result<String> {
343
- match roxmltree::Document::parse(xml) {
344
- Ok(doc) => {
345
- for node in doc.descendants() {
346
- if node.tag_name().name() == "rootfile"
347
- && let Some(full_path) = node.attribute("full-path")
348
- {
349
- return Ok(full_path.to_string());
350
- }
351
- }
352
- Err(crate::KreuzbergError::Parsing {
353
- message: "No rootfile found in container.xml".to_string(),
354
- source: None,
355
- })
356
- }
357
- Err(e) => Err(crate::KreuzbergError::Parsing {
358
- message: format!("Failed to parse container.xml: {}", e),
359
- source: None,
360
- }),
361
- }
362
- }
363
-
364
- /// Parse OPF file and extract metadata and spine order
365
- fn parse_opf(xml: &str) -> Result<(OepbMetadata, Vec<String>)> {
366
- match roxmltree::Document::parse(xml) {
367
- Ok(doc) => {
368
- let root = doc.root();
369
-
370
- let mut metadata = OepbMetadata::default();
371
- let mut manifest: BTreeMap<String, String> = BTreeMap::new();
372
- let mut spine_order: Vec<String> = Vec::new();
373
-
374
- for node in root.descendants() {
375
- match node.tag_name().name() {
376
- "title" => {
377
- if let Some(text) = node.text() {
378
- metadata.title = Some(text.trim().to_string());
379
- }
380
- }
381
- "creator" => {
382
- if let Some(text) = node.text() {
383
- metadata.creator = Some(text.trim().to_string());
384
- }
385
- }
386
- "date" => {
387
- if let Some(text) = node.text() {
388
- metadata.date = Some(text.trim().to_string());
389
- }
390
- }
391
- "language" => {
392
- if let Some(text) = node.text() {
393
- metadata.language = Some(text.trim().to_string());
394
- }
395
- }
396
- "identifier" => {
397
- if let Some(text) = node.text() {
398
- metadata.identifier = Some(text.trim().to_string());
399
- }
400
- }
401
- "publisher" => {
402
- if let Some(text) = node.text() {
403
- metadata.publisher = Some(text.trim().to_string());
404
- }
405
- }
406
- "subject" => {
407
- if let Some(text) = node.text() {
408
- metadata.subject = Some(text.trim().to_string());
409
- }
410
- }
411
- "description" => {
412
- if let Some(text) = node.text() {
413
- metadata.description = Some(text.trim().to_string());
414
- }
415
- }
416
- "rights" => {
417
- if let Some(text) = node.text() {
418
- metadata.rights = Some(text.trim().to_string());
419
- }
420
- }
421
- "item" => {
422
- if let Some(id) = node.attribute("id")
423
- && let Some(href) = node.attribute("href")
424
- {
425
- manifest.insert(id.to_string(), href.to_string());
426
- }
427
- }
428
- _ => {}
429
- }
430
- }
431
-
432
- for node in root.descendants() {
433
- if node.tag_name().name() == "itemref"
434
- && let Some(idref) = node.attribute("idref")
435
- && let Some(href) = manifest.get(idref)
436
- {
437
- spine_order.push(href.clone());
438
- }
439
- }
440
-
441
- Ok((metadata, spine_order))
442
- }
443
- Err(e) => Err(crate::KreuzbergError::Parsing {
444
- message: format!("Failed to parse OPF file: {}", e),
445
- source: None,
446
- }),
447
- }
448
- }
449
-
450
- /// Read a file from the ZIP archive
451
- fn read_file_from_zip(archive: &mut ZipArchive<Cursor<Vec<u8>>>, path: &str) -> Result<String> {
452
- match archive.by_name(path) {
453
- Ok(mut file) => {
454
- let mut content = String::new();
455
- match std::io::Read::read_to_string(&mut file, &mut content) {
456
- Ok(_) => Ok(content),
457
- Err(e) => Err(crate::KreuzbergError::Parsing {
458
- message: format!("Failed to read file from EPUB: {}", e),
459
- source: None,
460
- }),
461
- }
462
- }
463
- Err(e) => Err(crate::KreuzbergError::Parsing {
464
- message: format!("File not found in EPUB: {} ({})", path, e),
465
- source: None,
466
- }),
467
- }
468
- }
469
-
470
- /// Resolve a relative path within the manifest directory
471
- fn resolve_path(base_dir: &str, relative_path: &str) -> String {
472
- if relative_path.starts_with('/') {
473
- relative_path.trim_start_matches('/').to_string()
474
- } else if base_dir.is_empty() || base_dir == "." {
475
- relative_path.to_string()
476
- } else {
477
- format!("{}/{}", base_dir.trim_end_matches('/'), relative_path)
478
- }
479
- }
480
- }
481
-
482
- /// Metadata extracted from OPF (Open Packaging Format) file
483
- #[derive(Debug, Default, Clone)]
484
- struct OepbMetadata {
485
- title: Option<String>,
486
- creator: Option<String>,
487
- date: Option<String>,
488
- language: Option<String>,
489
- identifier: Option<String>,
490
- publisher: Option<String>,
491
- subject: Option<String>,
492
- description: Option<String>,
493
- rights: Option<String>,
494
- }
495
-
496
- impl Default for EpubExtractor {
497
- fn default() -> Self {
498
- Self::new()
499
- }
500
- }
501
-
502
- impl Plugin for EpubExtractor {
503
- fn name(&self) -> &str {
504
- "epub-extractor"
505
- }
506
-
507
- fn version(&self) -> String {
508
- env!("CARGO_PKG_VERSION").to_string()
509
- }
510
-
511
- fn initialize(&self) -> Result<()> {
512
- Ok(())
513
- }
514
-
515
- fn shutdown(&self) -> Result<()> {
516
- Ok(())
517
- }
518
-
519
- fn description(&self) -> &str {
520
- "Extracts content and metadata from EPUB documents (native Rust implementation with permissive licenses)"
521
- }
522
-
523
- fn author(&self) -> &str {
524
- "Kreuzberg Team"
525
- }
526
- }
527
-
528
- #[cfg(feature = "office")]
529
- #[async_trait]
530
- impl DocumentExtractor for EpubExtractor {
531
- #[cfg_attr(
532
- feature = "otel",
533
- tracing::instrument(
534
- skip(self, content, _config),
535
- fields(
536
- extractor.name = self.name(),
537
- content.size_bytes = content.len(),
538
- )
539
- )
540
- )]
541
- async fn extract_bytes(
542
- &self,
543
- content: &[u8],
544
- mime_type: &str,
545
- _config: &ExtractionConfig,
546
- ) -> Result<ExtractionResult> {
547
- let cursor = Cursor::new(content.to_vec());
548
-
549
- let mut archive = ZipArchive::new(cursor).map_err(|e| crate::KreuzbergError::Parsing {
550
- message: format!("Failed to open EPUB as ZIP: {}", e),
551
- source: None,
552
- })?;
553
-
554
- let container_xml = Self::read_file_from_zip(&mut archive, "META-INF/container.xml")?;
555
- let opf_path = Self::parse_container_xml(&container_xml)?;
556
-
557
- let manifest_dir = if let Some(last_slash) = opf_path.rfind('/') {
558
- opf_path[..last_slash].to_string()
559
- } else {
560
- String::new()
561
- };
562
-
563
- let opf_xml = Self::read_file_from_zip(&mut archive, &opf_path)?;
564
-
565
- let extracted_content = Self::extract_content(&mut archive, &opf_path, &manifest_dir)?;
566
-
567
- let metadata_btree = Self::extract_metadata(&opf_xml)?;
568
- let metadata_map: std::collections::HashMap<String, serde_json::Value> = metadata_btree.into_iter().collect();
569
-
570
- Ok(ExtractionResult {
571
- content: extracted_content,
572
- mime_type: mime_type.to_string(),
573
- metadata: Metadata {
574
- additional: metadata_map,
575
- ..Default::default()
576
- },
577
- tables: vec![],
578
- detected_languages: None,
579
- chunks: None,
580
- images: None,
581
- })
582
- }
583
-
584
- fn supported_mime_types(&self) -> &[&str] {
585
- &[
586
- "application/epub+zip",
587
- "application/x-epub+zip",
588
- "application/vnd.epub+zip",
589
- ]
590
- }
591
-
592
- fn priority(&self) -> i32 {
593
- 60
594
- }
595
- }
596
-
597
- #[cfg(all(test, feature = "office"))]
598
- mod tests {
599
- use super::*;
600
-
601
- #[test]
602
- fn test_epub_extractor_plugin_interface() {
603
- let extractor = EpubExtractor::new();
604
- assert_eq!(extractor.name(), "epub-extractor");
605
- assert_eq!(extractor.version(), env!("CARGO_PKG_VERSION"));
606
- assert_eq!(extractor.priority(), 60);
607
- assert!(!extractor.supported_mime_types().is_empty());
608
- }
609
-
610
- #[test]
611
- fn test_epub_extractor_default() {
612
- let extractor = EpubExtractor;
613
- assert_eq!(extractor.name(), "epub-extractor");
614
- }
615
-
616
- #[tokio::test]
617
- async fn test_epub_extractor_initialize_shutdown() {
618
- let extractor = EpubExtractor::new();
619
- assert!(extractor.initialize().is_ok());
620
- assert!(extractor.shutdown().is_ok());
621
- }
622
-
623
- #[test]
624
- fn test_strip_html_tags_simple() {
625
- let html = "<html><body><p>Hello World</p></body></html>";
626
- let text = EpubExtractor::strip_html_tags(html);
627
- assert!(text.contains("Hello World"));
628
- }
629
-
630
- #[test]
631
- fn test_strip_html_tags_with_scripts() {
632
- let html = "<body><p>Text</p><script>alert('bad');</script><p>More</p></body>";
633
- let text = EpubExtractor::strip_html_tags(html);
634
- assert!(!text.contains("bad"));
635
- assert!(text.contains("Text"));
636
- assert!(text.contains("More"));
637
- }
638
-
639
- #[test]
640
- fn test_strip_html_tags_with_styles() {
641
- let html = "<body><p>Text</p><style>.class { color: red; }</style><p>More</p></body>";
642
- let text = EpubExtractor::strip_html_tags(html);
643
- assert!(!text.to_lowercase().contains("color"));
644
- assert!(text.contains("Text"));
645
- assert!(text.contains("More"));
646
- }
647
-
648
- #[test]
649
- fn test_strip_html_tags_normalizes_whitespace() {
650
- let html = "<p>Hello \n\t World</p>";
651
- let text = EpubExtractor::strip_html_tags(html);
652
- assert!(text.contains("Hello") && text.contains("World"));
653
- }
654
-
655
- #[test]
656
- fn test_remove_markdown_links() {
657
- let text = "This is a [link](http://example.com) in text";
658
- let result = EpubExtractor::remove_markdown_links(text);
659
- assert!(result.contains("link"));
660
- assert!(!result.contains("http://"));
661
- }
662
-
663
- #[test]
664
- fn test_resolve_path_with_base_dir() {
665
- let result = EpubExtractor::resolve_path("OEBPS", "chapter.xhtml");
666
- assert_eq!(result, "OEBPS/chapter.xhtml");
667
- }
668
-
669
- #[test]
670
- fn test_resolve_path_absolute() {
671
- let result = EpubExtractor::resolve_path("OEBPS", "/chapter.xhtml");
672
- assert_eq!(result, "chapter.xhtml");
673
- }
674
-
675
- #[test]
676
- fn test_resolve_path_empty_base() {
677
- let result = EpubExtractor::resolve_path("", "chapter.xhtml");
678
- assert_eq!(result, "chapter.xhtml");
679
- }
680
-
681
- #[test]
682
- fn test_epub_extractor_supported_mime_types() {
683
- let extractor = EpubExtractor::new();
684
- let supported = extractor.supported_mime_types();
685
- assert!(supported.contains(&"application/epub+zip"));
686
- assert!(supported.contains(&"application/x-epub+zip"));
687
- assert!(supported.contains(&"application/vnd.epub+zip"));
688
- }
689
-
690
- #[test]
691
- fn test_markdown_to_plain_text_removes_formatting() {
692
- let markdown = "# Heading\n\nThis is **bold** text with _italic_ emphasis.";
693
- let result = EpubExtractor::markdown_to_plain_text(markdown);
694
- assert!(result.contains("Heading"));
695
- assert!(result.contains("bold"));
696
- assert!(!result.contains("**"));
697
- }
698
-
699
- #[test]
700
- fn test_markdown_to_plain_text_removes_list_markers() {
701
- let markdown = "- Item 1\n- Item 2\n* Item 3";
702
- let result = EpubExtractor::markdown_to_plain_text(markdown);
703
- assert!(result.contains("Item 1"));
704
- assert!(result.contains("Item 2"));
705
- assert!(result.contains("Item 3"));
706
- }
707
- }