kreuzberg 4.0.0.rc2 → 4.0.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (446) hide show
  1. checksums.yaml +4 -4
  2. data/.gitignore +14 -14
  3. data/.rspec +3 -3
  4. data/.rubocop.yaml +1 -1
  5. data/.rubocop.yml +543 -538
  6. data/Gemfile +8 -8
  7. data/Gemfile.lock +194 -6
  8. data/README.md +391 -426
  9. data/Rakefile +34 -25
  10. data/Steepfile +51 -47
  11. data/examples/async_patterns.rb +283 -341
  12. data/ext/kreuzberg_rb/extconf.rb +65 -45
  13. data/ext/kreuzberg_rb/native/.cargo/config.toml +23 -0
  14. data/ext/kreuzberg_rb/native/Cargo.lock +7619 -6535
  15. data/ext/kreuzberg_rb/native/Cargo.toml +75 -44
  16. data/ext/kreuzberg_rb/native/README.md +425 -425
  17. data/ext/kreuzberg_rb/native/build.rs +15 -15
  18. data/ext/kreuzberg_rb/native/include/ieeefp.h +11 -11
  19. data/ext/kreuzberg_rb/native/include/msvc_compat/strings.h +14 -14
  20. data/ext/kreuzberg_rb/native/include/strings.h +20 -20
  21. data/ext/kreuzberg_rb/native/include/unistd.h +47 -47
  22. data/ext/kreuzberg_rb/native/src/lib.rs +3802 -2998
  23. data/extconf.rb +60 -28
  24. data/kreuzberg.gemspec +199 -148
  25. data/lib/kreuzberg/api_proxy.rb +126 -142
  26. data/lib/kreuzberg/cache_api.rb +67 -46
  27. data/lib/kreuzberg/cli.rb +47 -55
  28. data/lib/kreuzberg/cli_proxy.rb +117 -127
  29. data/lib/kreuzberg/config.rb +936 -691
  30. data/lib/kreuzberg/error_context.rb +136 -32
  31. data/lib/kreuzberg/errors.rb +116 -118
  32. data/lib/kreuzberg/extraction_api.rb +313 -85
  33. data/lib/kreuzberg/mcp_proxy.rb +177 -186
  34. data/lib/kreuzberg/ocr_backend_protocol.rb +40 -113
  35. data/lib/kreuzberg/post_processor_protocol.rb +15 -86
  36. data/lib/kreuzberg/result.rb +334 -216
  37. data/lib/kreuzberg/setup_lib_path.rb +99 -80
  38. data/lib/kreuzberg/types.rb +170 -0
  39. data/lib/kreuzberg/validator_protocol.rb +16 -89
  40. data/lib/kreuzberg/version.rb +5 -5
  41. data/lib/kreuzberg.rb +96 -103
  42. data/lib/libpdfium.so +0 -0
  43. data/sig/kreuzberg/internal.rbs +184 -184
  44. data/sig/kreuzberg.rbs +561 -520
  45. data/spec/binding/async_operations_spec.rb +473 -0
  46. data/spec/binding/batch_operations_spec.rb +595 -0
  47. data/spec/binding/batch_spec.rb +359 -0
  48. data/spec/binding/cache_spec.rb +227 -227
  49. data/spec/binding/cli_proxy_spec.rb +85 -85
  50. data/spec/binding/cli_spec.rb +55 -55
  51. data/spec/binding/config_result_spec.rb +377 -0
  52. data/spec/binding/config_spec.rb +419 -345
  53. data/spec/binding/config_validation_spec.rb +377 -283
  54. data/spec/binding/embeddings_spec.rb +816 -0
  55. data/spec/binding/error_handling_spec.rb +399 -213
  56. data/spec/binding/error_recovery_spec.rb +488 -0
  57. data/spec/binding/errors_spec.rb +66 -66
  58. data/spec/binding/font_config_spec.rb +220 -0
  59. data/spec/binding/images_spec.rb +738 -0
  60. data/spec/binding/keywords_extraction_spec.rb +600 -0
  61. data/spec/binding/metadata_types_spec.rb +1228 -0
  62. data/spec/binding/pages_extraction_spec.rb +471 -0
  63. data/spec/binding/plugins/ocr_backend_spec.rb +307 -307
  64. data/spec/binding/plugins/postprocessor_spec.rb +269 -269
  65. data/spec/binding/plugins/validator_spec.rb +273 -274
  66. data/spec/binding/tables_spec.rb +641 -0
  67. data/spec/fixtures/config.toml +38 -39
  68. data/spec/fixtures/config.yaml +41 -41
  69. data/spec/fixtures/invalid_config.toml +3 -4
  70. data/spec/smoke/package_spec.rb +177 -178
  71. data/spec/spec_helper.rb +40 -42
  72. data/spec/unit/config/chunking_config_spec.rb +213 -0
  73. data/spec/unit/config/embedding_config_spec.rb +343 -0
  74. data/spec/unit/config/extraction_config_spec.rb +438 -0
  75. data/spec/unit/config/font_config_spec.rb +285 -0
  76. data/spec/unit/config/hierarchy_config_spec.rb +314 -0
  77. data/spec/unit/config/image_extraction_config_spec.rb +209 -0
  78. data/spec/unit/config/image_preprocessing_config_spec.rb +249 -0
  79. data/spec/unit/config/keyword_config_spec.rb +229 -0
  80. data/spec/unit/config/language_detection_config_spec.rb +258 -0
  81. data/spec/unit/config/ocr_config_spec.rb +171 -0
  82. data/spec/unit/config/page_config_spec.rb +221 -0
  83. data/spec/unit/config/pdf_config_spec.rb +267 -0
  84. data/spec/unit/config/postprocessor_config_spec.rb +290 -0
  85. data/spec/unit/config/tesseract_config_spec.rb +181 -0
  86. data/spec/unit/config/token_reduction_config_spec.rb +251 -0
  87. data/test/metadata_types_test.rb +959 -0
  88. data/vendor/Cargo.toml +61 -0
  89. data/vendor/kreuzberg/Cargo.toml +259 -204
  90. data/vendor/kreuzberg/README.md +263 -175
  91. data/vendor/kreuzberg/build.rs +782 -474
  92. data/vendor/kreuzberg/examples/bench_fixes.rs +71 -0
  93. data/vendor/kreuzberg/examples/test_pdfium_fork.rs +62 -0
  94. data/vendor/kreuzberg/src/api/error.rs +81 -81
  95. data/vendor/kreuzberg/src/api/handlers.rs +320 -199
  96. data/vendor/kreuzberg/src/api/mod.rs +94 -79
  97. data/vendor/kreuzberg/src/api/server.rs +518 -353
  98. data/vendor/kreuzberg/src/api/types.rs +206 -170
  99. data/vendor/kreuzberg/src/cache/mod.rs +1167 -1167
  100. data/vendor/kreuzberg/src/chunking/mod.rs +2303 -677
  101. data/vendor/kreuzberg/src/chunking/processor.rs +219 -0
  102. data/vendor/kreuzberg/src/core/batch_mode.rs +95 -95
  103. data/vendor/kreuzberg/src/core/batch_optimizations.rs +385 -0
  104. data/vendor/kreuzberg/src/core/config.rs +1914 -1032
  105. data/vendor/kreuzberg/src/core/config_validation.rs +949 -0
  106. data/vendor/kreuzberg/src/core/extractor.rs +1200 -1024
  107. data/vendor/kreuzberg/src/core/formats.rs +235 -0
  108. data/vendor/kreuzberg/src/core/io.rs +329 -329
  109. data/vendor/kreuzberg/src/core/mime.rs +605 -605
  110. data/vendor/kreuzberg/src/core/mod.rs +61 -45
  111. data/vendor/kreuzberg/src/core/pipeline.rs +1223 -984
  112. data/vendor/kreuzberg/src/core/server_config.rs +1220 -0
  113. data/vendor/kreuzberg/src/embeddings.rs +471 -432
  114. data/vendor/kreuzberg/src/error.rs +431 -431
  115. data/vendor/kreuzberg/src/extraction/archive.rs +959 -954
  116. data/vendor/kreuzberg/src/extraction/capacity.rs +263 -0
  117. data/vendor/kreuzberg/src/extraction/docx.rs +404 -40
  118. data/vendor/kreuzberg/src/extraction/email.rs +855 -854
  119. data/vendor/kreuzberg/src/extraction/excel.rs +697 -688
  120. data/vendor/kreuzberg/src/extraction/html.rs +1830 -553
  121. data/vendor/kreuzberg/src/extraction/image.rs +492 -368
  122. data/vendor/kreuzberg/src/extraction/libreoffice.rs +574 -563
  123. data/vendor/kreuzberg/src/extraction/markdown.rs +216 -213
  124. data/vendor/kreuzberg/src/extraction/mod.rs +93 -81
  125. data/vendor/kreuzberg/src/extraction/office_metadata/app_properties.rs +398 -398
  126. data/vendor/kreuzberg/src/extraction/office_metadata/core_properties.rs +247 -247
  127. data/vendor/kreuzberg/src/extraction/office_metadata/custom_properties.rs +240 -240
  128. data/vendor/kreuzberg/src/extraction/office_metadata/mod.rs +130 -130
  129. data/vendor/kreuzberg/src/extraction/office_metadata/odt_properties.rs +284 -287
  130. data/vendor/kreuzberg/src/extraction/pptx.rs +3102 -3000
  131. data/vendor/kreuzberg/src/extraction/structured.rs +491 -490
  132. data/vendor/kreuzberg/src/extraction/table.rs +329 -328
  133. data/vendor/kreuzberg/src/extraction/text.rs +277 -269
  134. data/vendor/kreuzberg/src/extraction/xml.rs +333 -333
  135. data/vendor/kreuzberg/src/extractors/archive.rs +447 -446
  136. data/vendor/kreuzberg/src/extractors/bibtex.rs +470 -469
  137. data/vendor/kreuzberg/src/extractors/docbook.rs +504 -502
  138. data/vendor/kreuzberg/src/extractors/docx.rs +400 -367
  139. data/vendor/kreuzberg/src/extractors/email.rs +157 -143
  140. data/vendor/kreuzberg/src/extractors/epub.rs +696 -707
  141. data/vendor/kreuzberg/src/extractors/excel.rs +385 -343
  142. data/vendor/kreuzberg/src/extractors/fictionbook.rs +492 -491
  143. data/vendor/kreuzberg/src/extractors/html.rs +419 -393
  144. data/vendor/kreuzberg/src/extractors/image.rs +219 -198
  145. data/vendor/kreuzberg/src/extractors/jats.rs +1054 -1051
  146. data/vendor/kreuzberg/src/extractors/jupyter.rs +368 -367
  147. data/vendor/kreuzberg/src/extractors/latex.rs +653 -652
  148. data/vendor/kreuzberg/src/extractors/markdown.rs +701 -700
  149. data/vendor/kreuzberg/src/extractors/mod.rs +429 -365
  150. data/vendor/kreuzberg/src/extractors/odt.rs +628 -628
  151. data/vendor/kreuzberg/src/extractors/opml.rs +635 -634
  152. data/vendor/kreuzberg/src/extractors/orgmode.rs +529 -528
  153. data/vendor/kreuzberg/src/extractors/pdf.rs +761 -493
  154. data/vendor/kreuzberg/src/extractors/pptx.rs +279 -248
  155. data/vendor/kreuzberg/src/extractors/rst.rs +577 -576
  156. data/vendor/kreuzberg/src/extractors/rtf.rs +809 -810
  157. data/vendor/kreuzberg/src/extractors/security.rs +484 -484
  158. data/vendor/kreuzberg/src/extractors/security_tests.rs +367 -367
  159. data/vendor/kreuzberg/src/extractors/structured.rs +142 -140
  160. data/vendor/kreuzberg/src/extractors/text.rs +265 -260
  161. data/vendor/kreuzberg/src/extractors/typst.rs +651 -650
  162. data/vendor/kreuzberg/src/extractors/xml.rs +147 -135
  163. data/vendor/kreuzberg/src/image/dpi.rs +164 -164
  164. data/vendor/kreuzberg/src/image/mod.rs +6 -6
  165. data/vendor/kreuzberg/src/image/preprocessing.rs +417 -417
  166. data/vendor/kreuzberg/src/image/resize.rs +89 -89
  167. data/vendor/kreuzberg/src/keywords/config.rs +154 -154
  168. data/vendor/kreuzberg/src/keywords/mod.rs +237 -237
  169. data/vendor/kreuzberg/src/keywords/processor.rs +275 -267
  170. data/vendor/kreuzberg/src/keywords/rake.rs +293 -293
  171. data/vendor/kreuzberg/src/keywords/types.rs +68 -68
  172. data/vendor/kreuzberg/src/keywords/yake.rs +163 -163
  173. data/vendor/kreuzberg/src/language_detection/mod.rs +985 -942
  174. data/vendor/kreuzberg/src/language_detection/processor.rs +218 -0
  175. data/vendor/kreuzberg/src/lib.rs +114 -105
  176. data/vendor/kreuzberg/src/mcp/mod.rs +35 -32
  177. data/vendor/kreuzberg/src/mcp/server.rs +2090 -1968
  178. data/vendor/kreuzberg/src/ocr/cache.rs +469 -469
  179. data/vendor/kreuzberg/src/ocr/error.rs +37 -37
  180. data/vendor/kreuzberg/src/ocr/hocr.rs +216 -216
  181. data/vendor/kreuzberg/src/ocr/language_registry.rs +520 -0
  182. data/vendor/kreuzberg/src/ocr/mod.rs +60 -58
  183. data/vendor/kreuzberg/src/ocr/processor.rs +858 -863
  184. data/vendor/kreuzberg/src/ocr/table/mod.rs +4 -4
  185. data/vendor/kreuzberg/src/ocr/table/tsv_parser.rs +144 -144
  186. data/vendor/kreuzberg/src/ocr/tesseract_backend.rs +456 -450
  187. data/vendor/kreuzberg/src/ocr/types.rs +393 -393
  188. data/vendor/kreuzberg/src/ocr/utils.rs +47 -47
  189. data/vendor/kreuzberg/src/ocr/validation.rs +206 -206
  190. data/vendor/kreuzberg/src/panic_context.rs +154 -154
  191. data/vendor/kreuzberg/src/pdf/bindings.rs +306 -0
  192. data/vendor/kreuzberg/src/pdf/bundled.rs +408 -0
  193. data/vendor/kreuzberg/src/pdf/error.rs +214 -122
  194. data/vendor/kreuzberg/src/pdf/fonts.rs +358 -0
  195. data/vendor/kreuzberg/src/pdf/hierarchy.rs +903 -0
  196. data/vendor/kreuzberg/src/pdf/images.rs +139 -139
  197. data/vendor/kreuzberg/src/pdf/metadata.rs +509 -346
  198. data/vendor/kreuzberg/src/pdf/mod.rs +81 -50
  199. data/vendor/kreuzberg/src/pdf/rendering.rs +369 -369
  200. data/vendor/kreuzberg/src/pdf/table.rs +417 -393
  201. data/vendor/kreuzberg/src/pdf/text.rs +553 -158
  202. data/vendor/kreuzberg/src/plugins/extractor.rs +1042 -1013
  203. data/vendor/kreuzberg/src/plugins/mod.rs +212 -209
  204. data/vendor/kreuzberg/src/plugins/ocr.rs +637 -620
  205. data/vendor/kreuzberg/src/plugins/processor.rs +650 -642
  206. data/vendor/kreuzberg/src/plugins/registry.rs +1339 -1337
  207. data/vendor/kreuzberg/src/plugins/traits.rs +258 -258
  208. data/vendor/kreuzberg/src/plugins/validator.rs +967 -956
  209. data/vendor/kreuzberg/src/stopwords/mod.rs +1470 -1470
  210. data/vendor/kreuzberg/src/text/mod.rs +27 -19
  211. data/vendor/kreuzberg/src/text/quality.rs +710 -697
  212. data/vendor/kreuzberg/src/text/quality_processor.rs +231 -0
  213. data/vendor/kreuzberg/src/text/string_utils.rs +229 -217
  214. data/vendor/kreuzberg/src/text/token_reduction/cjk_utils.rs +164 -164
  215. data/vendor/kreuzberg/src/text/token_reduction/config.rs +100 -100
  216. data/vendor/kreuzberg/src/text/token_reduction/core.rs +832 -796
  217. data/vendor/kreuzberg/src/text/token_reduction/filters.rs +923 -902
  218. data/vendor/kreuzberg/src/text/token_reduction/mod.rs +160 -160
  219. data/vendor/kreuzberg/src/text/token_reduction/semantic.rs +619 -619
  220. data/vendor/kreuzberg/src/text/token_reduction/simd_text.rs +148 -147
  221. data/vendor/kreuzberg/src/text/utf8_validation.rs +193 -0
  222. data/vendor/kreuzberg/src/types.rs +1713 -903
  223. data/vendor/kreuzberg/src/utils/mod.rs +31 -17
  224. data/vendor/kreuzberg/src/utils/pool.rs +503 -0
  225. data/vendor/kreuzberg/src/utils/pool_sizing.rs +364 -0
  226. data/vendor/kreuzberg/src/utils/quality.rs +968 -959
  227. data/vendor/kreuzberg/src/utils/string_pool.rs +761 -0
  228. data/vendor/kreuzberg/src/utils/string_utils.rs +381 -381
  229. data/vendor/kreuzberg/stopwords/af_stopwords.json +53 -53
  230. data/vendor/kreuzberg/stopwords/ar_stopwords.json +482 -482
  231. data/vendor/kreuzberg/stopwords/bg_stopwords.json +261 -261
  232. data/vendor/kreuzberg/stopwords/bn_stopwords.json +400 -400
  233. data/vendor/kreuzberg/stopwords/br_stopwords.json +1205 -1205
  234. data/vendor/kreuzberg/stopwords/ca_stopwords.json +280 -280
  235. data/vendor/kreuzberg/stopwords/cs_stopwords.json +425 -425
  236. data/vendor/kreuzberg/stopwords/da_stopwords.json +172 -172
  237. data/vendor/kreuzberg/stopwords/de_stopwords.json +622 -622
  238. data/vendor/kreuzberg/stopwords/el_stopwords.json +849 -849
  239. data/vendor/kreuzberg/stopwords/en_stopwords.json +1300 -1300
  240. data/vendor/kreuzberg/stopwords/eo_stopwords.json +175 -175
  241. data/vendor/kreuzberg/stopwords/es_stopwords.json +734 -734
  242. data/vendor/kreuzberg/stopwords/et_stopwords.json +37 -37
  243. data/vendor/kreuzberg/stopwords/eu_stopwords.json +100 -100
  244. data/vendor/kreuzberg/stopwords/fa_stopwords.json +801 -801
  245. data/vendor/kreuzberg/stopwords/fi_stopwords.json +849 -849
  246. data/vendor/kreuzberg/stopwords/fr_stopwords.json +693 -693
  247. data/vendor/kreuzberg/stopwords/ga_stopwords.json +111 -111
  248. data/vendor/kreuzberg/stopwords/gl_stopwords.json +162 -162
  249. data/vendor/kreuzberg/stopwords/gu_stopwords.json +226 -226
  250. data/vendor/kreuzberg/stopwords/ha_stopwords.json +41 -41
  251. data/vendor/kreuzberg/stopwords/he_stopwords.json +196 -196
  252. data/vendor/kreuzberg/stopwords/hi_stopwords.json +227 -227
  253. data/vendor/kreuzberg/stopwords/hr_stopwords.json +181 -181
  254. data/vendor/kreuzberg/stopwords/hu_stopwords.json +791 -791
  255. data/vendor/kreuzberg/stopwords/hy_stopwords.json +47 -47
  256. data/vendor/kreuzberg/stopwords/id_stopwords.json +760 -760
  257. data/vendor/kreuzberg/stopwords/it_stopwords.json +634 -634
  258. data/vendor/kreuzberg/stopwords/ja_stopwords.json +136 -136
  259. data/vendor/kreuzberg/stopwords/kn_stopwords.json +84 -84
  260. data/vendor/kreuzberg/stopwords/ko_stopwords.json +681 -681
  261. data/vendor/kreuzberg/stopwords/ku_stopwords.json +64 -64
  262. data/vendor/kreuzberg/stopwords/la_stopwords.json +51 -51
  263. data/vendor/kreuzberg/stopwords/lt_stopwords.json +476 -476
  264. data/vendor/kreuzberg/stopwords/lv_stopwords.json +163 -163
  265. data/vendor/kreuzberg/stopwords/ml_stopwords.json +1 -1
  266. data/vendor/kreuzberg/stopwords/mr_stopwords.json +101 -101
  267. data/vendor/kreuzberg/stopwords/ms_stopwords.json +477 -477
  268. data/vendor/kreuzberg/stopwords/ne_stopwords.json +490 -490
  269. data/vendor/kreuzberg/stopwords/nl_stopwords.json +415 -415
  270. data/vendor/kreuzberg/stopwords/no_stopwords.json +223 -223
  271. data/vendor/kreuzberg/stopwords/pl_stopwords.json +331 -331
  272. data/vendor/kreuzberg/stopwords/pt_stopwords.json +562 -562
  273. data/vendor/kreuzberg/stopwords/ro_stopwords.json +436 -436
  274. data/vendor/kreuzberg/stopwords/ru_stopwords.json +561 -561
  275. data/vendor/kreuzberg/stopwords/si_stopwords.json +193 -193
  276. data/vendor/kreuzberg/stopwords/sk_stopwords.json +420 -420
  277. data/vendor/kreuzberg/stopwords/sl_stopwords.json +448 -448
  278. data/vendor/kreuzberg/stopwords/so_stopwords.json +32 -32
  279. data/vendor/kreuzberg/stopwords/st_stopwords.json +33 -33
  280. data/vendor/kreuzberg/stopwords/sv_stopwords.json +420 -420
  281. data/vendor/kreuzberg/stopwords/sw_stopwords.json +76 -76
  282. data/vendor/kreuzberg/stopwords/ta_stopwords.json +129 -129
  283. data/vendor/kreuzberg/stopwords/te_stopwords.json +54 -54
  284. data/vendor/kreuzberg/stopwords/th_stopwords.json +118 -118
  285. data/vendor/kreuzberg/stopwords/tl_stopwords.json +149 -149
  286. data/vendor/kreuzberg/stopwords/tr_stopwords.json +506 -506
  287. data/vendor/kreuzberg/stopwords/uk_stopwords.json +75 -75
  288. data/vendor/kreuzberg/stopwords/ur_stopwords.json +519 -519
  289. data/vendor/kreuzberg/stopwords/vi_stopwords.json +647 -647
  290. data/vendor/kreuzberg/stopwords/yo_stopwords.json +62 -62
  291. data/vendor/kreuzberg/stopwords/zh_stopwords.json +796 -796
  292. data/vendor/kreuzberg/stopwords/zu_stopwords.json +31 -31
  293. data/vendor/kreuzberg/tests/api_embed.rs +360 -0
  294. data/vendor/kreuzberg/tests/api_extract_multipart.rs +52 -52
  295. data/vendor/kreuzberg/tests/api_large_pdf_extraction.rs +471 -0
  296. data/vendor/kreuzberg/tests/api_large_pdf_extraction_diagnostics.rs +289 -0
  297. data/vendor/kreuzberg/tests/api_tests.rs +1472 -966
  298. data/vendor/kreuzberg/tests/archive_integration.rs +545 -543
  299. data/vendor/kreuzberg/tests/batch_orchestration.rs +587 -556
  300. data/vendor/kreuzberg/tests/batch_pooling_benchmark.rs +154 -0
  301. data/vendor/kreuzberg/tests/batch_processing.rs +328 -316
  302. data/vendor/kreuzberg/tests/bibtex_parity_test.rs +421 -421
  303. data/vendor/kreuzberg/tests/concurrency_stress.rs +541 -525
  304. data/vendor/kreuzberg/tests/config_features.rs +612 -598
  305. data/vendor/kreuzberg/tests/config_integration_test.rs +753 -0
  306. data/vendor/kreuzberg/tests/config_loading_tests.rs +416 -415
  307. data/vendor/kreuzberg/tests/core_integration.rs +519 -510
  308. data/vendor/kreuzberg/tests/csv_integration.rs +414 -414
  309. data/vendor/kreuzberg/tests/data/hierarchy_ground_truth.json +294 -0
  310. data/vendor/kreuzberg/tests/docbook_extractor_tests.rs +500 -498
  311. data/vendor/kreuzberg/tests/docx_metadata_extraction_test.rs +122 -122
  312. data/vendor/kreuzberg/tests/docx_vs_pandoc_comparison.rs +370 -370
  313. data/vendor/kreuzberg/tests/email_integration.rs +327 -325
  314. data/vendor/kreuzberg/tests/epub_native_extractor_tests.rs +275 -275
  315. data/vendor/kreuzberg/tests/error_handling.rs +402 -393
  316. data/vendor/kreuzberg/tests/fictionbook_extractor_tests.rs +228 -228
  317. data/vendor/kreuzberg/tests/format_integration.rs +165 -159
  318. data/vendor/kreuzberg/tests/helpers/mod.rs +202 -142
  319. data/vendor/kreuzberg/tests/html_table_test.rs +551 -551
  320. data/vendor/kreuzberg/tests/image_integration.rs +255 -253
  321. data/vendor/kreuzberg/tests/instrumentation_test.rs +139 -139
  322. data/vendor/kreuzberg/tests/jats_extractor_tests.rs +639 -639
  323. data/vendor/kreuzberg/tests/jupyter_extractor_tests.rs +704 -704
  324. data/vendor/kreuzberg/tests/keywords_integration.rs +479 -479
  325. data/vendor/kreuzberg/tests/keywords_quality.rs +509 -509
  326. data/vendor/kreuzberg/tests/latex_extractor_tests.rs +496 -496
  327. data/vendor/kreuzberg/tests/markdown_extractor_tests.rs +490 -490
  328. data/vendor/kreuzberg/tests/mime_detection.rs +429 -428
  329. data/vendor/kreuzberg/tests/ocr_configuration.rs +514 -510
  330. data/vendor/kreuzberg/tests/ocr_errors.rs +698 -676
  331. data/vendor/kreuzberg/tests/ocr_language_registry.rs +191 -0
  332. data/vendor/kreuzberg/tests/ocr_quality.rs +629 -627
  333. data/vendor/kreuzberg/tests/ocr_stress.rs +469 -469
  334. data/vendor/kreuzberg/tests/odt_extractor_tests.rs +674 -695
  335. data/vendor/kreuzberg/tests/opml_extractor_tests.rs +616 -616
  336. data/vendor/kreuzberg/tests/orgmode_extractor_tests.rs +822 -822
  337. data/vendor/kreuzberg/tests/page_markers.rs +297 -0
  338. data/vendor/kreuzberg/tests/pdf_hierarchy_detection.rs +301 -0
  339. data/vendor/kreuzberg/tests/pdf_hierarchy_quality.rs +589 -0
  340. data/vendor/kreuzberg/tests/pdf_integration.rs +45 -43
  341. data/vendor/kreuzberg/tests/pdf_ocr_triggering.rs +301 -0
  342. data/vendor/kreuzberg/tests/pdf_text_merging.rs +475 -0
  343. data/vendor/kreuzberg/tests/pdfium_linking.rs +340 -0
  344. data/vendor/kreuzberg/tests/pipeline_integration.rs +1446 -1411
  345. data/vendor/kreuzberg/tests/plugin_ocr_backend_test.rs +776 -771
  346. data/vendor/kreuzberg/tests/plugin_postprocessor_test.rs +577 -560
  347. data/vendor/kreuzberg/tests/plugin_system.rs +927 -921
  348. data/vendor/kreuzberg/tests/plugin_validator_test.rs +783 -783
  349. data/vendor/kreuzberg/tests/registry_integration_tests.rs +587 -586
  350. data/vendor/kreuzberg/tests/rst_extractor_tests.rs +694 -692
  351. data/vendor/kreuzberg/tests/rtf_extractor_tests.rs +775 -776
  352. data/vendor/kreuzberg/tests/security_validation.rs +416 -415
  353. data/vendor/kreuzberg/tests/stopwords_integration_test.rs +888 -888
  354. data/vendor/kreuzberg/tests/test_fastembed.rs +631 -609
  355. data/vendor/kreuzberg/tests/typst_behavioral_tests.rs +1260 -1259
  356. data/vendor/kreuzberg/tests/typst_extractor_tests.rs +648 -647
  357. data/vendor/kreuzberg/tests/xlsx_metadata_extraction_test.rs +87 -87
  358. data/vendor/kreuzberg-ffi/Cargo.toml +67 -0
  359. data/vendor/kreuzberg-ffi/README.md +851 -0
  360. data/vendor/kreuzberg-ffi/benches/result_view_benchmark.rs +227 -0
  361. data/vendor/kreuzberg-ffi/build.rs +168 -0
  362. data/vendor/kreuzberg-ffi/cbindgen.toml +37 -0
  363. data/vendor/kreuzberg-ffi/kreuzberg-ffi.pc.in +12 -0
  364. data/vendor/kreuzberg-ffi/kreuzberg.h +3012 -0
  365. data/vendor/kreuzberg-ffi/src/batch_streaming.rs +588 -0
  366. data/vendor/kreuzberg-ffi/src/config.rs +1341 -0
  367. data/vendor/kreuzberg-ffi/src/error.rs +901 -0
  368. data/vendor/kreuzberg-ffi/src/extraction.rs +555 -0
  369. data/vendor/kreuzberg-ffi/src/helpers.rs +879 -0
  370. data/vendor/kreuzberg-ffi/src/lib.rs +977 -0
  371. data/vendor/kreuzberg-ffi/src/memory.rs +493 -0
  372. data/vendor/kreuzberg-ffi/src/mime.rs +329 -0
  373. data/vendor/kreuzberg-ffi/src/panic_shield.rs +265 -0
  374. data/vendor/kreuzberg-ffi/src/plugins/document_extractor.rs +442 -0
  375. data/vendor/kreuzberg-ffi/src/plugins/mod.rs +14 -0
  376. data/vendor/kreuzberg-ffi/src/plugins/ocr_backend.rs +628 -0
  377. data/vendor/kreuzberg-ffi/src/plugins/post_processor.rs +438 -0
  378. data/vendor/kreuzberg-ffi/src/plugins/validator.rs +329 -0
  379. data/vendor/kreuzberg-ffi/src/result.rs +510 -0
  380. data/vendor/kreuzberg-ffi/src/result_pool.rs +639 -0
  381. data/vendor/kreuzberg-ffi/src/result_view.rs +773 -0
  382. data/vendor/kreuzberg-ffi/src/string_intern.rs +568 -0
  383. data/vendor/kreuzberg-ffi/src/types.rs +363 -0
  384. data/vendor/kreuzberg-ffi/src/util.rs +210 -0
  385. data/vendor/kreuzberg-ffi/src/validation.rs +848 -0
  386. data/vendor/kreuzberg-ffi/tests.disabled/README.md +48 -0
  387. data/vendor/kreuzberg-ffi/tests.disabled/config_loading_tests.rs +299 -0
  388. data/vendor/kreuzberg-ffi/tests.disabled/config_tests.rs +346 -0
  389. data/vendor/kreuzberg-ffi/tests.disabled/extractor_tests.rs +232 -0
  390. data/vendor/kreuzberg-ffi/tests.disabled/plugin_registration_tests.rs +470 -0
  391. data/vendor/kreuzberg-tesseract/.commitlintrc.json +13 -0
  392. data/vendor/kreuzberg-tesseract/.crate-ignore +2 -0
  393. data/vendor/kreuzberg-tesseract/Cargo.lock +2933 -0
  394. data/vendor/kreuzberg-tesseract/Cargo.toml +57 -0
  395. data/vendor/{rb-sys/LICENSE-MIT → kreuzberg-tesseract/LICENSE} +22 -21
  396. data/vendor/kreuzberg-tesseract/README.md +399 -0
  397. data/vendor/kreuzberg-tesseract/build.rs +1127 -0
  398. data/vendor/kreuzberg-tesseract/patches/README.md +71 -0
  399. data/vendor/kreuzberg-tesseract/patches/tesseract.diff +199 -0
  400. data/vendor/kreuzberg-tesseract/src/api.rs +1371 -0
  401. data/vendor/kreuzberg-tesseract/src/choice_iterator.rs +77 -0
  402. data/vendor/kreuzberg-tesseract/src/enums.rs +297 -0
  403. data/vendor/kreuzberg-tesseract/src/error.rs +81 -0
  404. data/vendor/kreuzberg-tesseract/src/lib.rs +145 -0
  405. data/vendor/kreuzberg-tesseract/src/monitor.rs +57 -0
  406. data/vendor/kreuzberg-tesseract/src/mutable_iterator.rs +197 -0
  407. data/vendor/kreuzberg-tesseract/src/page_iterator.rs +253 -0
  408. data/vendor/kreuzberg-tesseract/src/result_iterator.rs +286 -0
  409. data/vendor/kreuzberg-tesseract/src/result_renderer.rs +183 -0
  410. data/vendor/kreuzberg-tesseract/tests/integration_test.rs +211 -0
  411. metadata +196 -45
  412. data/vendor/kreuzberg/benches/otel_overhead.rs +0 -48
  413. data/vendor/kreuzberg/src/extractors/fictionbook.rs.backup2 +0 -738
  414. data/vendor/rb-sys/.cargo-ok +0 -1
  415. data/vendor/rb-sys/.cargo_vcs_info.json +0 -6
  416. data/vendor/rb-sys/Cargo.lock +0 -393
  417. data/vendor/rb-sys/Cargo.toml +0 -70
  418. data/vendor/rb-sys/Cargo.toml.orig +0 -57
  419. data/vendor/rb-sys/LICENSE-APACHE +0 -190
  420. data/vendor/rb-sys/bin/release.sh +0 -21
  421. data/vendor/rb-sys/build/features.rs +0 -108
  422. data/vendor/rb-sys/build/main.rs +0 -246
  423. data/vendor/rb-sys/build/stable_api_config.rs +0 -153
  424. data/vendor/rb-sys/build/version.rs +0 -48
  425. data/vendor/rb-sys/readme.md +0 -36
  426. data/vendor/rb-sys/src/bindings.rs +0 -21
  427. data/vendor/rb-sys/src/hidden.rs +0 -11
  428. data/vendor/rb-sys/src/lib.rs +0 -34
  429. data/vendor/rb-sys/src/macros.rs +0 -371
  430. data/vendor/rb-sys/src/memory.rs +0 -53
  431. data/vendor/rb-sys/src/ruby_abi_version.rs +0 -38
  432. data/vendor/rb-sys/src/special_consts.rs +0 -31
  433. data/vendor/rb-sys/src/stable_api/compiled.c +0 -179
  434. data/vendor/rb-sys/src/stable_api/compiled.rs +0 -257
  435. data/vendor/rb-sys/src/stable_api/ruby_2_6.rs +0 -316
  436. data/vendor/rb-sys/src/stable_api/ruby_2_7.rs +0 -316
  437. data/vendor/rb-sys/src/stable_api/ruby_3_0.rs +0 -324
  438. data/vendor/rb-sys/src/stable_api/ruby_3_1.rs +0 -317
  439. data/vendor/rb-sys/src/stable_api/ruby_3_2.rs +0 -315
  440. data/vendor/rb-sys/src/stable_api/ruby_3_3.rs +0 -326
  441. data/vendor/rb-sys/src/stable_api/ruby_3_4.rs +0 -327
  442. data/vendor/rb-sys/src/stable_api.rs +0 -261
  443. data/vendor/rb-sys/src/symbol.rs +0 -31
  444. data/vendor/rb-sys/src/tracking_allocator.rs +0 -332
  445. data/vendor/rb-sys/src/utils.rs +0 -89
  446. data/vendor/rb-sys/src/value_type.rs +0 -7
@@ -1,556 +1,587 @@
1
- //! Batch processing orchestration tests.
2
- //!
3
- //! Validates efficient parallel processing at multiple levels:
4
- //! - Multiple documents in parallel
5
- //! - Multiple pages within PDFs
6
- //! - OCR across pages
7
- //! - File I/O optimization
8
- //! - Resource utilization (CPU cores)
9
-
10
- use kreuzberg::core::config::ExtractionConfig;
11
- use kreuzberg::core::extractor::{batch_extract_bytes, batch_extract_file};
12
- use std::time::{Duration, Instant};
13
-
14
- #[cfg(feature = "ocr")]
15
- use kreuzberg::core::config::OcrConfig;
16
-
17
- #[cfg(feature = "ocr")]
18
- use kreuzberg::core::extractor::extract_file_sync;
19
-
20
- mod helpers;
21
-
22
- fn trim_trailing_newlines(value: &str) -> &str {
23
- value.trim_end_matches(['\n', '\r'])
24
- }
25
-
26
- fn assert_text_content(actual: &str, expected: &str) {
27
- assert_eq!(
28
- trim_trailing_newlines(actual),
29
- expected,
30
- "Content mismatch after trimming trailing newlines"
31
- );
32
- }
33
-
34
- /// Test that batch extraction processes documents in parallel.
35
- ///
36
- /// Validates:
37
- /// - Multiple documents process concurrently
38
- /// - Parallel processing is faster than sequential
39
- /// - Results maintain correct order
40
- #[tokio::test]
41
- async fn test_batch_documents_parallel_execution() {
42
- use helpers::get_test_file_path;
43
- use std::path::PathBuf;
44
-
45
- let config = ExtractionConfig::default();
46
-
47
- let test_files = vec![
48
- "text/contract.txt",
49
- "json/sample_document.json",
50
- "xml/simple_note.xml",
51
- "text/readme.md",
52
- ];
53
-
54
- let mut paths: Vec<PathBuf> = Vec::new();
55
- for _ in 0..5 {
56
- for file in &test_files {
57
- paths.push(get_test_file_path(file));
58
- }
59
- }
60
-
61
- let parallel_start = Instant::now();
62
- let results = batch_extract_file(paths.clone(), &config).await;
63
- let parallel_duration = parallel_start.elapsed();
64
-
65
- assert!(results.is_ok(), "Batch extraction should succeed");
66
- let results = results.unwrap();
67
- assert_eq!(results.len(), 20, "Should process all 20 files");
68
-
69
- for result in &results {
70
- assert!(
71
- !result.content.is_empty() || result.metadata.error.is_some(),
72
- "Each result should have content or error"
73
- );
74
- }
75
-
76
- assert!(
77
- parallel_duration < Duration::from_secs(5),
78
- "Batch processing 20 files should take <5s, took: {:?}",
79
- parallel_duration
80
- );
81
- }
82
-
83
- /// Test concurrency limiting in batch processing.
84
- ///
85
- /// Validates that batch extraction respects max_concurrent_extractions config.
86
- #[tokio::test]
87
- async fn test_batch_documents_concurrency_limiting() {
88
- use helpers::get_test_file_path;
89
-
90
- let config = ExtractionConfig {
91
- max_concurrent_extractions: Some(2),
92
- ..Default::default()
93
- };
94
-
95
- let paths = vec![
96
- get_test_file_path("text/contract.txt"),
97
- get_test_file_path("json/sample_document.json"),
98
- get_test_file_path("xml/simple_note.xml"),
99
- get_test_file_path("text/readme.md"),
100
- ];
101
-
102
- let results = batch_extract_file(paths, &config).await;
103
-
104
- assert!(results.is_ok());
105
- let results = results.unwrap();
106
- assert_eq!(results.len(), 4);
107
- }
108
-
109
- /// Test batch extraction with CPU-bound limit (default: num_cpus * 2).
110
- #[tokio::test]
111
- async fn test_batch_documents_default_concurrency() {
112
- use helpers::get_test_file_path;
113
-
114
- let config = ExtractionConfig::default();
115
-
116
- let mut paths = Vec::new();
117
- for _ in 0..13 {
118
- paths.push(get_test_file_path("text/contract.txt"));
119
- paths.push(get_test_file_path("json/sample_document.json"));
120
- paths.push(get_test_file_path("xml/simple_note.xml"));
121
- paths.push(get_test_file_path("text/readme.md"));
122
- }
123
- let paths = paths.into_iter().take(50).collect::<Vec<_>>();
124
-
125
- let start = Instant::now();
126
- let results = batch_extract_file(paths, &config).await;
127
- let duration = start.elapsed();
128
-
129
- assert!(results.is_ok());
130
- let results = results.unwrap();
131
- assert_eq!(results.len(), 50);
132
-
133
- println!("Processed 50 files in {:?}", duration);
134
- assert!(
135
- duration < Duration::from_secs(10),
136
- "50 files should process in <10s with parallelism, took: {:?}",
137
- duration
138
- );
139
- }
140
-
141
- /// Test that batch processing maintains result order.
142
- #[cfg(feature = "xml")]
143
- #[tokio::test]
144
- async fn test_batch_documents_preserves_order() {
145
- use helpers::get_test_file_path;
146
-
147
- let config = ExtractionConfig::default();
148
-
149
- let paths = vec![
150
- get_test_file_path("text/contract.txt"),
151
- get_test_file_path("json/sample_document.json"),
152
- get_test_file_path("xml/simple_note.xml"),
153
- ];
154
-
155
- let results = batch_extract_file(paths, &config).await.unwrap();
156
-
157
- assert_eq!(results.len(), 3, "Should have 3 results");
158
-
159
- assert!(!results[0].content.is_empty(), "First result should have content");
160
- assert!(!results[1].content.is_empty(), "Second result should have content");
161
- assert!(!results[2].content.is_empty(), "Third result should have content");
162
-
163
- assert!(
164
- results[0].content.contains("contract"),
165
- "First result should be from contract.txt, got: '{}'",
166
- results[0].content
167
- );
168
- assert!(
169
- results[1].content.contains("Sample") || results[1].content.contains("author"),
170
- "Second result should be from JSON document, got: '{}'",
171
- results[1].content
172
- );
173
- assert!(
174
- results[2].content.contains("Tove") || results[2].content.contains("note"),
175
- "Third result should be from XML note, got: '{}'",
176
- results[2].content
177
- );
178
- }
179
-
180
- /// Test that multi-page PDF extraction is efficient.
181
- ///
182
- /// Validates:
183
- /// - Multiple pages are processed
184
- /// - OCR is applied to all pages if needed
185
- /// - Content from all pages is combined
186
- #[cfg(feature = "pdf")]
187
- #[tokio::test]
188
- async fn test_multipage_pdf_extraction() {
189
- use helpers::{get_test_file_path, skip_if_missing};
190
-
191
- if skip_if_missing("pdfs/multi_page.pdf") {
192
- tracing::debug!("Skipping multi-page PDF test: test file not available");
193
- return;
194
- }
195
-
196
- let config = ExtractionConfig::default();
197
- let pdf_path = get_test_file_path("pdfs/multi_page.pdf");
198
-
199
- let start = Instant::now();
200
- let result = kreuzberg::core::extractor::extract_file(&pdf_path, None, &config).await;
201
- let duration = start.elapsed();
202
-
203
- assert!(result.is_ok(), "Multi-page PDF extraction should succeed");
204
- let extraction = result.unwrap();
205
-
206
- assert!(!extraction.content.is_empty(), "Should extract text from all pages");
207
- println!("Extracted multi-page PDF in {:?}", duration);
208
- }
209
-
210
- /// Test concurrent PDF extractions (multiple PDFs at once).
211
- #[cfg(feature = "pdf")]
212
- #[tokio::test]
213
- async fn test_concurrent_pdf_extractions() {
214
- use helpers::{get_test_file_path, skip_if_missing};
215
-
216
- if skip_if_missing("pdfs/simple.pdf") {
217
- tracing::debug!("Skipping concurrent PDF test: test file not available");
218
- return;
219
- }
220
-
221
- let config = ExtractionConfig::default();
222
-
223
- let mut paths = Vec::new();
224
- for _ in 0..10 {
225
- paths.push(get_test_file_path("pdfs/simple.pdf"));
226
- }
227
-
228
- let start = Instant::now();
229
- let results = batch_extract_file(paths, &config).await;
230
- let duration = start.elapsed();
231
-
232
- assert!(results.is_ok());
233
- let results = results.unwrap();
234
- assert_eq!(results.len(), 10);
235
-
236
- println!("Processed 10 PDFs in {:?}", duration);
237
- }
238
-
239
- /// Test OCR on multi-page scanned document.
240
- ///
241
- /// Validates:
242
- /// - All pages are OCR'd
243
- /// - Results are combined correctly
244
- /// - Processing is efficient
245
- #[cfg(feature = "ocr")]
246
- #[test]
247
- fn test_ocr_multipage_efficiency() {
248
- use helpers::{get_test_file_path, skip_if_missing};
249
-
250
- if skip_if_missing("images/ocr_image.jpg") {
251
- tracing::debug!("Skipping OCR multi-page test: test file not available");
252
- return;
253
- }
254
-
255
- let config = ExtractionConfig {
256
- ocr: Some(OcrConfig {
257
- backend: "tesseract".to_string(),
258
- language: "eng".to_string(),
259
- tesseract_config: None,
260
- }),
261
- force_ocr: false,
262
- use_cache: true,
263
- ..Default::default()
264
- };
265
-
266
- let file_path = get_test_file_path("images/ocr_image.jpg");
267
-
268
- let start = Instant::now();
269
- let result1 = extract_file_sync(&file_path, None, &config);
270
- let first_duration = start.elapsed();
271
-
272
- assert!(result1.is_ok(), "First OCR should succeed");
273
-
274
- let start = Instant::now();
275
- let result2 = extract_file_sync(&file_path, None, &config);
276
- let second_duration = start.elapsed();
277
-
278
- assert!(result2.is_ok(), "Second OCR should succeed");
279
-
280
- println!(
281
- "OCR timing: first={:?}, cached={:?}, speedup={:.1}x",
282
- first_duration,
283
- second_duration,
284
- first_duration.as_secs_f64() / second_duration.as_secs_f64().max(0.001)
285
- );
286
-
287
- assert!(
288
- second_duration < first_duration / 2,
289
- "Cached OCR should be at least 2x faster. First: {:?}, Second: {:?}",
290
- first_duration,
291
- second_duration
292
- );
293
- }
294
-
295
- /// Test parallel processing of byte arrays.
296
- ///
297
- /// Validates that batch_extract_bytes processes data in parallel.
298
- #[tokio::test]
299
- async fn test_batch_bytes_parallel_processing() {
300
- let config = ExtractionConfig::default();
301
-
302
- let contents: Vec<(Vec<u8>, &str)> = (0..30)
303
- .map(|i| {
304
- let content = format!("Test content number {}", i);
305
- (content.into_bytes(), "text/plain")
306
- })
307
- .collect();
308
-
309
- let contents_ref: Vec<(&[u8], &str)> = contents.iter().map(|(bytes, mime)| (bytes.as_slice(), *mime)).collect();
310
-
311
- let start = Instant::now();
312
- let results = batch_extract_bytes(contents_ref, &config).await;
313
- let duration = start.elapsed();
314
-
315
- assert!(results.is_ok());
316
- let results = results.unwrap();
317
- assert_eq!(results.len(), 30);
318
-
319
- for (i, result) in results.iter().enumerate() {
320
- let expected = format!("Test content number {}", i);
321
- assert_text_content(&result.content, &expected);
322
- }
323
-
324
- println!("Batch processed 30 byte arrays in {:?}", duration);
325
- }
326
-
327
- /// Test error handling in batch bytes processing.
328
- #[tokio::test]
329
- async fn test_batch_bytes_mixed_valid_invalid() {
330
- let config = ExtractionConfig::default();
331
-
332
- let contents = vec![
333
- (b"valid content 1".as_slice(), "text/plain"),
334
- (b"invalid content".as_slice(), "invalid/mime"),
335
- (b"valid content 2".as_slice(), "text/plain"),
336
- (b"more invalid".as_slice(), "bad/type"),
337
- (b"valid content 3".as_slice(), "text/plain"),
338
- ];
339
-
340
- let results = batch_extract_bytes(contents, &config).await;
341
-
342
- assert!(results.is_ok());
343
- let results = results.unwrap();
344
- assert_eq!(results.len(), 5);
345
-
346
- assert_text_content(&results[0].content, "valid content 1");
347
- assert_text_content(&results[2].content, "valid content 2");
348
- assert_text_content(&results[4].content, "valid content 3");
349
-
350
- assert!(results[1].metadata.error.is_some());
351
- assert!(results[3].metadata.error.is_some());
352
- }
353
-
354
- /// Test that batch processing utilizes multiple CPU cores.
355
- ///
356
- /// Validates that parallel extraction actually runs in parallel,
357
- /// not just sequentially with fancy task management.
358
- #[tokio::test]
359
- async fn test_batch_utilizes_multiple_cores() {
360
- let config = ExtractionConfig {
361
- max_concurrent_extractions: Some(num_cpus::get()),
362
- ..Default::default()
363
- };
364
-
365
- let mut contents = Vec::new();
366
- for i in 0..20 {
367
- let json = format!(
368
- r#"{{"id": {}, "data": "{}", "nested": {{"value": "{}"}}}}"#,
369
- i,
370
- "x".repeat(100),
371
- "y".repeat(100)
372
- );
373
- contents.push((json.into_bytes(), "application/json"));
374
- }
375
-
376
- let contents_ref: Vec<(&[u8], &str)> = contents.iter().map(|(bytes, mime)| (bytes.as_slice(), *mime)).collect();
377
-
378
- let start = Instant::now();
379
- let results = batch_extract_bytes(contents_ref, &config).await;
380
- let duration = start.elapsed();
381
-
382
- assert!(results.is_ok());
383
- let results = results.unwrap();
384
- assert_eq!(results.len(), 20);
385
-
386
- println!(
387
- "Processed 20 JSON documents in {:?} with {} cores",
388
- duration,
389
- num_cpus::get()
390
- );
391
-
392
- assert!(
393
- duration < Duration::from_secs(2),
394
- "Batch processing should leverage parallelism, took: {:?}",
395
- duration
396
- );
397
- }
398
-
399
- /// Test batch processing under memory pressure.
400
- ///
401
- /// Validates that semaphore prevents resource exhaustion.
402
- #[tokio::test]
403
- async fn test_batch_memory_pressure_handling() {
404
- let config = ExtractionConfig {
405
- max_concurrent_extractions: Some(4),
406
- ..Default::default()
407
- };
408
-
409
- let mut contents = Vec::new();
410
- for i in 0..50 {
411
- let json = format!(r#"{{"id": {}, "large_data": "{}"}}"#, i, "x".repeat(10000));
412
- contents.push((json.into_bytes(), "application/json"));
413
- }
414
-
415
- let contents_ref: Vec<(&[u8], &str)> = contents.iter().map(|(bytes, mime)| (bytes.as_slice(), *mime)).collect();
416
-
417
- let start = Instant::now();
418
- let results = batch_extract_bytes(contents_ref, &config).await;
419
- let duration = start.elapsed();
420
-
421
- assert!(results.is_ok());
422
- let results = results.unwrap();
423
- assert_eq!(results.len(), 50);
424
-
425
- println!("Processed 50 large documents with concurrency limit in {:?}", duration);
426
-
427
- for result in &results {
428
- assert!(!result.content.is_empty());
429
- }
430
- }
431
-
432
- /// Test that batch processing scales with CPU count.
433
- #[tokio::test]
434
- async fn test_batch_scales_with_cpu_count() {
435
- let cpu_count = num_cpus::get();
436
-
437
- let contents: Vec<(Vec<u8>, &str)> = (0..30)
438
- .map(|i| (format!("Content {}", i).into_bytes(), "text/plain"))
439
- .collect();
440
-
441
- let config_1 = ExtractionConfig {
442
- max_concurrent_extractions: Some(1),
443
- ..Default::default()
444
- };
445
-
446
- let contents_ref: Vec<(&[u8], &str)> = contents.iter().map(|(bytes, mime)| (bytes.as_slice(), *mime)).collect();
447
-
448
- let start = Instant::now();
449
- let _ = batch_extract_bytes(contents_ref.clone(), &config_1).await.unwrap();
450
- let duration_1 = start.elapsed();
451
-
452
- let config_full = ExtractionConfig {
453
- max_concurrent_extractions: Some(cpu_count),
454
- ..Default::default()
455
- };
456
-
457
- let start = Instant::now();
458
- let _ = batch_extract_bytes(contents_ref, &config_full).await.unwrap();
459
- let duration_full = start.elapsed();
460
-
461
- println!(
462
- "Concurrency=1: {:?}, Concurrency={}: {:?}, Speedup: {:.2}x",
463
- duration_1,
464
- cpu_count,
465
- duration_full,
466
- duration_1.as_secs_f64() / duration_full.as_secs_f64()
467
- );
468
-
469
- if cpu_count > 1 {
470
- let slowdown_ratio = duration_full.as_secs_f64() / duration_1.as_secs_f64();
471
- assert!(
472
- slowdown_ratio <= 5.0,
473
- "Parallel execution should not be excessively slower (got {:.2}x slowdown)",
474
- slowdown_ratio
475
- );
476
- }
477
- }
478
-
479
- /// End-to-end test: batch process mixed document types.
480
- #[cfg(feature = "xml")]
481
- #[tokio::test]
482
- async fn test_batch_mixed_document_types() {
483
- use helpers::get_test_file_path;
484
-
485
- let config = ExtractionConfig::default();
486
-
487
- let paths = vec![
488
- get_test_file_path("text/contract.txt"),
489
- get_test_file_path("json/sample_document.json"),
490
- get_test_file_path("xml/simple_note.xml"),
491
- get_test_file_path("text/readme.md"),
492
- ];
493
-
494
- let results = batch_extract_file(paths, &config).await;
495
-
496
- assert!(results.is_ok());
497
- let results = results.unwrap();
498
- assert_eq!(results.len(), 4);
499
-
500
- for (i, result) in results.iter().enumerate() {
501
- assert!(
502
- !result.content.is_empty(),
503
- "Document {} should have extracted content",
504
- i
505
- );
506
- }
507
-
508
- assert!(
509
- results[0].content.contains("contract"),
510
- "First result should be from contract.txt, got: '{}'",
511
- results[0].content
512
- );
513
- assert!(
514
- results[1].content.contains("Sample") || results[1].content.contains("author"),
515
- "Second result should be from JSON document, got: '{}'",
516
- results[1].content
517
- );
518
- assert!(
519
- results[2].content.contains("Tove") || results[2].content.contains("note"),
520
- "Third result should be from XML, got: '{}'",
521
- results[2].content
522
- );
523
- assert!(
524
- !results[3].content.is_empty(),
525
- "Fourth result should be from markdown, got: '{}'",
526
- results[3].content
527
- );
528
- }
529
-
530
- /// Test batch processing maintains high accuracy under load.
531
- #[tokio::test]
532
- async fn test_batch_accuracy_under_load() {
533
- let config = ExtractionConfig::default();
534
-
535
- let mut contents = Vec::new();
536
- for i in 0..100 {
537
- let content = format!("Document number {} with unique content", i);
538
- contents.push((content.into_bytes(), "text/plain"));
539
- }
540
-
541
- let contents_ref: Vec<(&[u8], &str)> = contents.iter().map(|(bytes, mime)| (bytes.as_slice(), *mime)).collect();
542
-
543
- let results = batch_extract_bytes(contents_ref, &config).await.unwrap();
544
-
545
- assert_eq!(results.len(), 100);
546
-
547
- for (i, result) in results.iter().enumerate() {
548
- let expected = format!("Document number {} with unique content", i);
549
- assert_eq!(
550
- trim_trailing_newlines(&result.content),
551
- expected,
552
- "Document {} content mismatch - possible cross-contamination",
553
- i
554
- );
555
- }
556
- }
1
+ //! Batch processing orchestration tests.
2
+ //!
3
+ //! Validates efficient parallel processing at multiple levels:
4
+ //! - Multiple documents in parallel
5
+ //! - Multiple pages within PDFs
6
+ //! - OCR across pages
7
+ //! - File I/O optimization
8
+ //! - Resource utilization (CPU cores)
9
+
10
+ use kreuzberg::core::config::ExtractionConfig;
11
+ use kreuzberg::core::extractor::{batch_extract_bytes, batch_extract_file};
12
+ use std::time::{Duration, Instant};
13
+
14
+ #[cfg(feature = "ocr")]
15
+ use kreuzberg::core::config::OcrConfig;
16
+
17
+ #[cfg(feature = "ocr")]
18
+ use kreuzberg::core::extractor::extract_file_sync;
19
+
20
+ mod helpers;
21
+
22
+ fn trim_trailing_newlines(value: &str) -> &str {
23
+ value.trim_end_matches(['\n', '\r'])
24
+ }
25
+
26
+ fn assert_text_content(actual: &str, expected: &str) {
27
+ assert_eq!(
28
+ trim_trailing_newlines(actual),
29
+ expected,
30
+ "Content mismatch after trimming trailing newlines"
31
+ );
32
+ }
33
+
34
+ /// Test that batch extraction processes documents in parallel.
35
+ ///
36
+ /// Validates:
37
+ /// - Multiple documents process concurrently
38
+ /// - Parallel processing is faster than sequential
39
+ /// - Results maintain correct order
40
+ #[tokio::test]
41
+ async fn test_batch_documents_parallel_execution() {
42
+ use helpers::get_test_file_path;
43
+ use std::path::PathBuf;
44
+
45
+ let config = ExtractionConfig::default();
46
+
47
+ let test_files = vec![
48
+ "text/contract.txt",
49
+ "json/sample_document.json",
50
+ "xml/simple_note.xml",
51
+ "text/readme.md",
52
+ ];
53
+
54
+ let mut paths: Vec<PathBuf> = Vec::new();
55
+ for _ in 0..5 {
56
+ for file in &test_files {
57
+ paths.push(get_test_file_path(file));
58
+ }
59
+ }
60
+
61
+ let parallel_start = Instant::now();
62
+ let results = batch_extract_file(paths.clone(), &config).await;
63
+ let parallel_duration = parallel_start.elapsed();
64
+
65
+ assert!(results.is_ok(), "Batch extraction should succeed");
66
+ let results = results.unwrap();
67
+ assert_eq!(results.len(), 20, "Should process all 20 files");
68
+
69
+ for result in &results {
70
+ assert!(
71
+ !result.content.is_empty() || result.metadata.error.is_some(),
72
+ "Each result should have content or error"
73
+ );
74
+ }
75
+
76
+ assert!(
77
+ parallel_duration < Duration::from_secs(5),
78
+ "Batch processing 20 files should take <5s, took: {:?}",
79
+ parallel_duration
80
+ );
81
+ }
82
+
83
+ /// Test concurrency limiting in batch processing.
84
+ ///
85
+ /// Validates that batch extraction respects max_concurrent_extractions config.
86
+ #[tokio::test]
87
+ async fn test_batch_documents_concurrency_limiting() {
88
+ use helpers::get_test_file_path;
89
+
90
+ let config = ExtractionConfig {
91
+ max_concurrent_extractions: Some(2),
92
+ ..Default::default()
93
+ };
94
+
95
+ let paths = vec![
96
+ get_test_file_path("text/contract.txt"),
97
+ get_test_file_path("json/sample_document.json"),
98
+ get_test_file_path("xml/simple_note.xml"),
99
+ get_test_file_path("text/readme.md"),
100
+ ];
101
+
102
+ let results = batch_extract_file(paths, &config).await;
103
+
104
+ assert!(results.is_ok());
105
+ let results = results.unwrap();
106
+ assert_eq!(results.len(), 4);
107
+ }
108
+
109
+ /// Test batch extraction with CPU-bound limit (default: num_cpus * 2).
110
+ #[tokio::test]
111
+ async fn test_batch_documents_default_concurrency() {
112
+ use helpers::get_test_file_path;
113
+
114
+ let config = ExtractionConfig::default();
115
+
116
+ let mut paths = Vec::new();
117
+ for _ in 0..13 {
118
+ paths.push(get_test_file_path("text/contract.txt"));
119
+ paths.push(get_test_file_path("json/sample_document.json"));
120
+ paths.push(get_test_file_path("xml/simple_note.xml"));
121
+ paths.push(get_test_file_path("text/readme.md"));
122
+ }
123
+ let paths = paths.into_iter().take(50).collect::<Vec<_>>();
124
+
125
+ let start = Instant::now();
126
+ let results = batch_extract_file(paths, &config).await;
127
+ let duration = start.elapsed();
128
+
129
+ assert!(results.is_ok());
130
+ let results = results.unwrap();
131
+ assert_eq!(results.len(), 50);
132
+
133
+ println!("Processed 50 files in {:?}", duration);
134
+ assert!(
135
+ duration < Duration::from_secs(10),
136
+ "50 files should process in <10s with parallelism, took: {:?}",
137
+ duration
138
+ );
139
+ }
140
+
141
+ /// Test that batch processing maintains result order.
142
+ #[cfg(feature = "xml")]
143
+ #[tokio::test]
144
+ async fn test_batch_documents_preserves_order() {
145
+ use helpers::get_test_file_path;
146
+
147
+ let config = ExtractionConfig::default();
148
+
149
+ let paths = vec![
150
+ get_test_file_path("text/contract.txt"),
151
+ get_test_file_path("json/sample_document.json"),
152
+ get_test_file_path("xml/simple_note.xml"),
153
+ ];
154
+
155
+ let results = batch_extract_file(paths, &config).await.unwrap();
156
+
157
+ assert_eq!(results.len(), 3, "Should have 3 results");
158
+
159
+ assert!(!results[0].content.is_empty(), "First result should have content");
160
+ assert!(!results[1].content.is_empty(), "Second result should have content");
161
+ assert!(!results[2].content.is_empty(), "Third result should have content");
162
+
163
+ assert!(
164
+ results[0].content.contains("contract"),
165
+ "First result should be from contract.txt, got: '{}'",
166
+ results[0].content
167
+ );
168
+ assert!(
169
+ results[1].content.contains("Sample") || results[1].content.contains("author"),
170
+ "Second result should be from JSON document, got: '{}'",
171
+ results[1].content
172
+ );
173
+ assert!(
174
+ results[2].content.contains("Tove") || results[2].content.contains("note"),
175
+ "Third result should be from XML note, got: '{}'",
176
+ results[2].content
177
+ );
178
+ }
179
+
180
+ /// Test that multi-page PDF extraction is efficient.
181
+ ///
182
+ /// Validates:
183
+ /// - Multiple pages are processed
184
+ /// - OCR is applied to all pages if needed
185
+ /// - Content from all pages is combined
186
+ #[cfg(feature = "pdf")]
187
+ #[tokio::test]
188
+ async fn test_multipage_pdf_extraction() {
189
+ use helpers::{get_test_file_path, skip_if_missing};
190
+
191
+ if skip_if_missing("pdfs/multi_page.pdf") {
192
+ tracing::debug!("Skipping multi-page PDF test: test file not available");
193
+ return;
194
+ }
195
+
196
+ let config = ExtractionConfig::default();
197
+ let pdf_path = get_test_file_path("pdfs/multi_page.pdf");
198
+
199
+ let start = Instant::now();
200
+ let result = kreuzberg::core::extractor::extract_file(&pdf_path, None, &config).await;
201
+ let duration = start.elapsed();
202
+
203
+ assert!(result.is_ok(), "Multi-page PDF extraction should succeed");
204
+ let extraction = result.unwrap();
205
+
206
+ assert!(!extraction.content.is_empty(), "Should extract text from all pages");
207
+ println!("Extracted multi-page PDF in {:?}", duration);
208
+ }
209
+
210
+ /// Test concurrent PDF extractions (multiple PDFs at once).
211
+ #[cfg(feature = "pdf")]
212
+ #[tokio::test]
213
+ async fn test_concurrent_pdf_extractions() {
214
+ use helpers::{get_test_file_path, skip_if_missing};
215
+
216
+ if skip_if_missing("pdfs/simple.pdf") {
217
+ tracing::debug!("Skipping concurrent PDF test: test file not available");
218
+ return;
219
+ }
220
+
221
+ let config = ExtractionConfig::default();
222
+
223
+ let mut paths = Vec::new();
224
+ for _ in 0..10 {
225
+ paths.push(get_test_file_path("pdfs/simple.pdf"));
226
+ }
227
+
228
+ let start = Instant::now();
229
+ let results = batch_extract_file(paths, &config).await;
230
+ let duration = start.elapsed();
231
+
232
+ assert!(results.is_ok());
233
+ let results = results.unwrap();
234
+ assert_eq!(results.len(), 10);
235
+
236
+ println!("Processed 10 PDFs in {:?}", duration);
237
+ }
238
+
239
+ /// Test OCR on multi-page scanned document.
240
+ ///
241
+ /// Validates:
242
+ /// - All pages are OCR'd
243
+ /// - Results are combined correctly
244
+ /// - Processing is efficient
245
+ #[cfg(feature = "ocr")]
246
+ #[test]
247
+ fn test_ocr_multipage_efficiency() {
248
+ use helpers::{get_test_file_path, skip_if_missing};
249
+
250
+ if skip_if_missing("images/ocr_image.jpg") {
251
+ tracing::debug!("Skipping OCR multi-page test: test file not available");
252
+ return;
253
+ }
254
+
255
+ let config = ExtractionConfig {
256
+ ocr: Some(OcrConfig {
257
+ backend: "tesseract".to_string(),
258
+ language: "eng".to_string(),
259
+ tesseract_config: None,
260
+ }),
261
+ force_ocr: false,
262
+ use_cache: true,
263
+ ..Default::default()
264
+ };
265
+
266
+ let file_path = get_test_file_path("images/ocr_image.jpg");
267
+
268
+ let start = Instant::now();
269
+ let result1 = extract_file_sync(&file_path, None, &config);
270
+ let first_duration = start.elapsed();
271
+
272
+ assert!(result1.is_ok(), "First OCR should succeed");
273
+
274
+ let start = Instant::now();
275
+ let result2 = extract_file_sync(&file_path, None, &config);
276
+ let second_duration = start.elapsed();
277
+
278
+ assert!(result2.is_ok(), "Second OCR should succeed");
279
+
280
+ println!(
281
+ "OCR timing: first={:?}, cached={:?}, speedup={:.1}x",
282
+ first_duration,
283
+ second_duration,
284
+ first_duration.as_secs_f64() / second_duration.as_secs_f64().max(0.001)
285
+ );
286
+
287
+ assert!(
288
+ second_duration < first_duration / 2,
289
+ "Cached OCR should be at least 2x faster. First: {:?}, Second: {:?}",
290
+ first_duration,
291
+ second_duration
292
+ );
293
+ }
294
+
295
+ /// Test parallel processing of byte arrays.
296
+ ///
297
+ /// Validates that batch_extract_bytes processes data in parallel.
298
+ #[tokio::test]
299
+ async fn test_batch_bytes_parallel_processing() {
300
+ let config = ExtractionConfig::default();
301
+
302
+ let contents: Vec<(Vec<u8>, &str)> = (0..30)
303
+ .map(|i| {
304
+ let content = format!("Test content number {}", i);
305
+ (content.into_bytes(), "text/plain")
306
+ })
307
+ .collect();
308
+
309
+ let contents_ref: Vec<(&[u8], &str)> = contents.iter().map(|(bytes, mime)| (bytes.as_slice(), *mime)).collect();
310
+ let owned_contents: Vec<(Vec<u8>, String)> = contents_ref
311
+ .into_iter()
312
+ .map(|(bytes, mime)| (bytes.to_vec(), mime.to_string()))
313
+ .collect();
314
+
315
+ let start = Instant::now();
316
+ let results = batch_extract_bytes(owned_contents, &config).await;
317
+ let duration = start.elapsed();
318
+
319
+ assert!(results.is_ok());
320
+ let results = results.unwrap();
321
+ assert_eq!(results.len(), 30);
322
+
323
+ for (i, result) in results.iter().enumerate() {
324
+ let expected = format!("Test content number {}", i);
325
+ assert_text_content(&result.content, &expected);
326
+ }
327
+
328
+ println!("Batch processed 30 byte arrays in {:?}", duration);
329
+ }
330
+
331
+ /// Test error handling in batch bytes processing.
332
+ #[tokio::test]
333
+ async fn test_batch_bytes_mixed_valid_invalid() {
334
+ let config = ExtractionConfig::default();
335
+
336
+ let contents = vec![
337
+ (b"valid content 1".as_slice(), "text/plain"),
338
+ (b"invalid content".as_slice(), "invalid/mime"),
339
+ (b"valid content 2".as_slice(), "text/plain"),
340
+ (b"more invalid".as_slice(), "bad/type"),
341
+ (b"valid content 3".as_slice(), "text/plain"),
342
+ ];
343
+
344
+ let owned_contents: Vec<(Vec<u8>, String)> = contents
345
+ .into_iter()
346
+ .map(|(bytes, mime)| (bytes.to_vec(), mime.to_string()))
347
+ .collect();
348
+
349
+ let results = batch_extract_bytes(owned_contents, &config).await;
350
+
351
+ assert!(results.is_ok());
352
+ let results = results.unwrap();
353
+ assert_eq!(results.len(), 5);
354
+
355
+ assert_text_content(&results[0].content, "valid content 1");
356
+ assert_text_content(&results[2].content, "valid content 2");
357
+ assert_text_content(&results[4].content, "valid content 3");
358
+
359
+ assert!(results[1].metadata.error.is_some());
360
+ assert!(results[3].metadata.error.is_some());
361
+ }
362
+
363
+ /// Test that batch processing utilizes multiple CPU cores.
364
+ ///
365
+ /// Validates that parallel extraction actually runs in parallel,
366
+ /// not just sequentially with fancy task management.
367
+ #[tokio::test]
368
+ async fn test_batch_utilizes_multiple_cores() {
369
+ let config = ExtractionConfig {
370
+ max_concurrent_extractions: Some(num_cpus::get()),
371
+ ..Default::default()
372
+ };
373
+
374
+ let mut contents = Vec::new();
375
+ for i in 0..20 {
376
+ let json = format!(
377
+ r#"{{"id": {}, "data": "{}", "nested": {{"value": "{}"}}}}"#,
378
+ i,
379
+ "x".repeat(100),
380
+ "y".repeat(100)
381
+ );
382
+ contents.push((json.into_bytes(), "application/json"));
383
+ }
384
+
385
+ let contents_ref: Vec<(&[u8], &str)> = contents.iter().map(|(bytes, mime)| (bytes.as_slice(), *mime)).collect();
386
+ let owned_contents: Vec<(Vec<u8>, String)> = contents_ref
387
+ .into_iter()
388
+ .map(|(bytes, mime)| (bytes.to_vec(), mime.to_string()))
389
+ .collect();
390
+
391
+ let start = Instant::now();
392
+ let results = batch_extract_bytes(owned_contents, &config).await;
393
+ let duration = start.elapsed();
394
+
395
+ assert!(results.is_ok());
396
+ let results = results.unwrap();
397
+ assert_eq!(results.len(), 20);
398
+
399
+ println!(
400
+ "Processed 20 JSON documents in {:?} with {} cores",
401
+ duration,
402
+ num_cpus::get()
403
+ );
404
+
405
+ assert!(
406
+ duration < Duration::from_secs(2),
407
+ "Batch processing should leverage parallelism, took: {:?}",
408
+ duration
409
+ );
410
+ }
411
+
412
+ /// Test batch processing under memory pressure.
413
+ ///
414
+ /// Validates that semaphore prevents resource exhaustion.
415
+ #[tokio::test]
416
+ async fn test_batch_memory_pressure_handling() {
417
+ let config = ExtractionConfig {
418
+ max_concurrent_extractions: Some(4),
419
+ ..Default::default()
420
+ };
421
+
422
+ let mut contents = Vec::new();
423
+ for i in 0..50 {
424
+ let json = format!(r#"{{"id": {}, "large_data": "{}"}}"#, i, "x".repeat(10000));
425
+ contents.push((json.into_bytes(), "application/json"));
426
+ }
427
+
428
+ let contents_ref: Vec<(&[u8], &str)> = contents.iter().map(|(bytes, mime)| (bytes.as_slice(), *mime)).collect();
429
+ let owned_contents: Vec<(Vec<u8>, String)> = contents_ref
430
+ .into_iter()
431
+ .map(|(bytes, mime)| (bytes.to_vec(), mime.to_string()))
432
+ .collect();
433
+
434
+ let start = Instant::now();
435
+ let results = batch_extract_bytes(owned_contents, &config).await;
436
+ let duration = start.elapsed();
437
+
438
+ assert!(results.is_ok());
439
+ let results = results.unwrap();
440
+ assert_eq!(results.len(), 50);
441
+
442
+ println!("Processed 50 large documents with concurrency limit in {:?}", duration);
443
+
444
+ for result in &results {
445
+ assert!(!result.content.is_empty());
446
+ }
447
+ }
448
+
449
+ /// Test that batch processing scales with CPU count.
450
+ #[tokio::test]
451
+ async fn test_batch_scales_with_cpu_count() {
452
+ let cpu_count = num_cpus::get();
453
+
454
+ let contents: Vec<(Vec<u8>, &str)> = (0..30)
455
+ .map(|i| (format!("Content {}", i).into_bytes(), "text/plain"))
456
+ .collect();
457
+
458
+ let config_1 = ExtractionConfig {
459
+ max_concurrent_extractions: Some(1),
460
+ ..Default::default()
461
+ };
462
+
463
+ let contents_ref: Vec<(&[u8], &str)> = contents.iter().map(|(bytes, mime)| (bytes.as_slice(), *mime)).collect();
464
+
465
+ let owned_contents_1: Vec<(Vec<u8>, String)> = contents_ref
466
+ .iter()
467
+ .map(|(bytes, mime)| (bytes.to_vec(), mime.to_string()))
468
+ .collect();
469
+
470
+ let start = Instant::now();
471
+ let _ = batch_extract_bytes(owned_contents_1, &config_1).await.unwrap();
472
+ let duration_1 = start.elapsed();
473
+
474
+ let config_full = ExtractionConfig {
475
+ max_concurrent_extractions: Some(cpu_count),
476
+ ..Default::default()
477
+ };
478
+
479
+ let owned_contents_full: Vec<(Vec<u8>, String)> = contents_ref
480
+ .into_iter()
481
+ .map(|(bytes, mime)| (bytes.to_vec(), mime.to_string()))
482
+ .collect();
483
+
484
+ let start = Instant::now();
485
+ let _ = batch_extract_bytes(owned_contents_full, &config_full).await.unwrap();
486
+ let duration_full = start.elapsed();
487
+
488
+ println!(
489
+ "Concurrency=1: {:?}, Concurrency={}: {:?}, Speedup: {:.2}x",
490
+ duration_1,
491
+ cpu_count,
492
+ duration_full,
493
+ duration_1.as_secs_f64() / duration_full.as_secs_f64()
494
+ );
495
+
496
+ if cpu_count > 1 {
497
+ let slowdown_ratio = duration_full.as_secs_f64() / duration_1.as_secs_f64();
498
+ assert!(
499
+ slowdown_ratio <= 5.0,
500
+ "Parallel execution should not be excessively slower (got {:.2}x slowdown)",
501
+ slowdown_ratio
502
+ );
503
+ }
504
+ }
505
+
506
+ /// End-to-end test: batch process mixed document types.
507
+ #[cfg(feature = "xml")]
508
+ #[tokio::test]
509
+ async fn test_batch_mixed_document_types() {
510
+ use helpers::get_test_file_path;
511
+
512
+ let config = ExtractionConfig::default();
513
+
514
+ let paths = vec![
515
+ get_test_file_path("text/contract.txt"),
516
+ get_test_file_path("json/sample_document.json"),
517
+ get_test_file_path("xml/simple_note.xml"),
518
+ get_test_file_path("text/readme.md"),
519
+ ];
520
+
521
+ let results = batch_extract_file(paths, &config).await;
522
+
523
+ assert!(results.is_ok());
524
+ let results = results.unwrap();
525
+ assert_eq!(results.len(), 4);
526
+
527
+ for (i, result) in results.iter().enumerate() {
528
+ assert!(
529
+ !result.content.is_empty(),
530
+ "Document {} should have extracted content",
531
+ i
532
+ );
533
+ }
534
+
535
+ assert!(
536
+ results[0].content.contains("contract"),
537
+ "First result should be from contract.txt, got: '{}'",
538
+ results[0].content
539
+ );
540
+ assert!(
541
+ results[1].content.contains("Sample") || results[1].content.contains("author"),
542
+ "Second result should be from JSON document, got: '{}'",
543
+ results[1].content
544
+ );
545
+ assert!(
546
+ results[2].content.contains("Tove") || results[2].content.contains("note"),
547
+ "Third result should be from XML, got: '{}'",
548
+ results[2].content
549
+ );
550
+ assert!(
551
+ !results[3].content.is_empty(),
552
+ "Fourth result should be from markdown, got: '{}'",
553
+ results[3].content
554
+ );
555
+ }
556
+
557
+ /// Test batch processing maintains high accuracy under load.
558
+ #[tokio::test]
559
+ async fn test_batch_accuracy_under_load() {
560
+ let config = ExtractionConfig::default();
561
+
562
+ let mut contents = Vec::new();
563
+ for i in 0..100 {
564
+ let content = format!("Document number {} with unique content", i);
565
+ contents.push((content.into_bytes(), "text/plain"));
566
+ }
567
+
568
+ let contents_ref: Vec<(&[u8], &str)> = contents.iter().map(|(bytes, mime)| (bytes.as_slice(), *mime)).collect();
569
+ let owned_contents: Vec<(Vec<u8>, String)> = contents_ref
570
+ .into_iter()
571
+ .map(|(bytes, mime)| (bytes.to_vec(), mime.to_string()))
572
+ .collect();
573
+
574
+ let results = batch_extract_bytes(owned_contents, &config).await.unwrap();
575
+
576
+ assert_eq!(results.len(), 100);
577
+
578
+ for (i, result) in results.iter().enumerate() {
579
+ let expected = format!("Document number {} with unique content", i);
580
+ assert_eq!(
581
+ trim_trailing_newlines(&result.content),
582
+ expected,
583
+ "Document {} content mismatch - possible cross-contamination",
584
+ i
585
+ );
586
+ }
587
+ }