kreuzberg 4.0.0.rc1 → 4.0.0.rc2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (342) hide show
  1. checksums.yaml +4 -4
  2. data/.gitignore +14 -8
  3. data/.rspec +3 -3
  4. data/.rubocop.yaml +1 -534
  5. data/.rubocop.yml +538 -0
  6. data/Gemfile +8 -9
  7. data/Gemfile.lock +9 -109
  8. data/README.md +426 -421
  9. data/Rakefile +25 -25
  10. data/Steepfile +47 -47
  11. data/examples/async_patterns.rb +341 -340
  12. data/ext/kreuzberg_rb/extconf.rb +45 -35
  13. data/ext/kreuzberg_rb/native/Cargo.lock +6535 -0
  14. data/ext/kreuzberg_rb/native/Cargo.toml +44 -36
  15. data/ext/kreuzberg_rb/native/README.md +425 -425
  16. data/ext/kreuzberg_rb/native/build.rs +15 -17
  17. data/ext/kreuzberg_rb/native/include/ieeefp.h +11 -11
  18. data/ext/kreuzberg_rb/native/include/msvc_compat/strings.h +14 -14
  19. data/ext/kreuzberg_rb/native/include/strings.h +20 -20
  20. data/ext/kreuzberg_rb/native/include/unistd.h +47 -47
  21. data/ext/kreuzberg_rb/native/src/lib.rs +2998 -2939
  22. data/extconf.rb +28 -28
  23. data/kreuzberg.gemspec +148 -105
  24. data/lib/kreuzberg/api_proxy.rb +142 -142
  25. data/lib/kreuzberg/cache_api.rb +46 -45
  26. data/lib/kreuzberg/cli.rb +55 -55
  27. data/lib/kreuzberg/cli_proxy.rb +127 -127
  28. data/lib/kreuzberg/config.rb +691 -684
  29. data/lib/kreuzberg/error_context.rb +32 -0
  30. data/lib/kreuzberg/errors.rb +118 -50
  31. data/lib/kreuzberg/extraction_api.rb +85 -84
  32. data/lib/kreuzberg/mcp_proxy.rb +186 -186
  33. data/lib/kreuzberg/ocr_backend_protocol.rb +113 -113
  34. data/lib/kreuzberg/post_processor_protocol.rb +86 -86
  35. data/lib/kreuzberg/result.rb +216 -216
  36. data/lib/kreuzberg/setup_lib_path.rb +80 -79
  37. data/lib/kreuzberg/validator_protocol.rb +89 -89
  38. data/lib/kreuzberg/version.rb +5 -5
  39. data/lib/kreuzberg.rb +103 -82
  40. data/sig/kreuzberg/internal.rbs +184 -184
  41. data/sig/kreuzberg.rbs +520 -468
  42. data/spec/binding/cache_spec.rb +227 -227
  43. data/spec/binding/cli_proxy_spec.rb +85 -87
  44. data/spec/binding/cli_spec.rb +55 -54
  45. data/spec/binding/config_spec.rb +345 -345
  46. data/spec/binding/config_validation_spec.rb +283 -283
  47. data/spec/binding/error_handling_spec.rb +213 -213
  48. data/spec/binding/errors_spec.rb +66 -66
  49. data/spec/binding/plugins/ocr_backend_spec.rb +307 -307
  50. data/spec/binding/plugins/postprocessor_spec.rb +269 -269
  51. data/spec/binding/plugins/validator_spec.rb +274 -274
  52. data/spec/fixtures/config.toml +39 -39
  53. data/spec/fixtures/config.yaml +41 -42
  54. data/spec/fixtures/invalid_config.toml +4 -4
  55. data/spec/smoke/package_spec.rb +178 -178
  56. data/spec/spec_helper.rb +42 -42
  57. data/vendor/kreuzberg/Cargo.toml +204 -134
  58. data/vendor/kreuzberg/README.md +175 -175
  59. data/vendor/kreuzberg/benches/otel_overhead.rs +48 -0
  60. data/vendor/kreuzberg/build.rs +474 -460
  61. data/vendor/kreuzberg/src/api/error.rs +81 -81
  62. data/vendor/kreuzberg/src/api/handlers.rs +199 -199
  63. data/vendor/kreuzberg/src/api/mod.rs +79 -79
  64. data/vendor/kreuzberg/src/api/server.rs +353 -353
  65. data/vendor/kreuzberg/src/api/types.rs +170 -170
  66. data/vendor/kreuzberg/src/cache/mod.rs +1167 -1143
  67. data/vendor/kreuzberg/src/chunking/mod.rs +677 -677
  68. data/vendor/kreuzberg/src/core/batch_mode.rs +95 -35
  69. data/vendor/kreuzberg/src/core/config.rs +1032 -1032
  70. data/vendor/kreuzberg/src/core/extractor.rs +1024 -903
  71. data/vendor/kreuzberg/src/core/io.rs +329 -327
  72. data/vendor/kreuzberg/src/core/mime.rs +605 -615
  73. data/vendor/kreuzberg/src/core/mod.rs +45 -42
  74. data/vendor/kreuzberg/src/core/pipeline.rs +984 -906
  75. data/vendor/kreuzberg/src/embeddings.rs +432 -323
  76. data/vendor/kreuzberg/src/error.rs +431 -431
  77. data/vendor/kreuzberg/src/extraction/archive.rs +954 -954
  78. data/vendor/kreuzberg/src/extraction/docx.rs +40 -40
  79. data/vendor/kreuzberg/src/extraction/email.rs +854 -854
  80. data/vendor/kreuzberg/src/extraction/excel.rs +688 -688
  81. data/vendor/kreuzberg/src/extraction/html.rs +553 -553
  82. data/vendor/kreuzberg/src/extraction/image.rs +368 -368
  83. data/vendor/kreuzberg/src/extraction/libreoffice.rs +563 -564
  84. data/vendor/kreuzberg/src/extraction/markdown.rs +213 -0
  85. data/vendor/kreuzberg/src/extraction/mod.rs +81 -77
  86. data/vendor/kreuzberg/src/extraction/office_metadata/app_properties.rs +398 -398
  87. data/vendor/kreuzberg/src/extraction/office_metadata/core_properties.rs +247 -247
  88. data/vendor/kreuzberg/src/extraction/office_metadata/custom_properties.rs +240 -240
  89. data/vendor/kreuzberg/src/extraction/office_metadata/mod.rs +130 -128
  90. data/vendor/kreuzberg/src/extraction/office_metadata/odt_properties.rs +287 -0
  91. data/vendor/kreuzberg/src/extraction/pptx.rs +3000 -3000
  92. data/vendor/kreuzberg/src/extraction/structured.rs +490 -490
  93. data/vendor/kreuzberg/src/extraction/table.rs +328 -328
  94. data/vendor/kreuzberg/src/extraction/text.rs +269 -269
  95. data/vendor/kreuzberg/src/extraction/xml.rs +333 -333
  96. data/vendor/kreuzberg/src/extractors/archive.rs +446 -425
  97. data/vendor/kreuzberg/src/extractors/bibtex.rs +469 -0
  98. data/vendor/kreuzberg/src/extractors/docbook.rs +502 -0
  99. data/vendor/kreuzberg/src/extractors/docx.rs +367 -479
  100. data/vendor/kreuzberg/src/extractors/email.rs +143 -129
  101. data/vendor/kreuzberg/src/extractors/epub.rs +707 -0
  102. data/vendor/kreuzberg/src/extractors/excel.rs +343 -344
  103. data/vendor/kreuzberg/src/extractors/fictionbook.rs +491 -0
  104. data/vendor/kreuzberg/src/extractors/fictionbook.rs.backup2 +738 -0
  105. data/vendor/kreuzberg/src/extractors/html.rs +393 -410
  106. data/vendor/kreuzberg/src/extractors/image.rs +198 -195
  107. data/vendor/kreuzberg/src/extractors/jats.rs +1051 -0
  108. data/vendor/kreuzberg/src/extractors/jupyter.rs +367 -0
  109. data/vendor/kreuzberg/src/extractors/latex.rs +652 -0
  110. data/vendor/kreuzberg/src/extractors/markdown.rs +700 -0
  111. data/vendor/kreuzberg/src/extractors/mod.rs +365 -268
  112. data/vendor/kreuzberg/src/extractors/odt.rs +628 -0
  113. data/vendor/kreuzberg/src/extractors/opml.rs +634 -0
  114. data/vendor/kreuzberg/src/extractors/orgmode.rs +528 -0
  115. data/vendor/kreuzberg/src/extractors/pdf.rs +493 -496
  116. data/vendor/kreuzberg/src/extractors/pptx.rs +248 -234
  117. data/vendor/kreuzberg/src/extractors/rst.rs +576 -0
  118. data/vendor/kreuzberg/src/extractors/rtf.rs +810 -0
  119. data/vendor/kreuzberg/src/extractors/security.rs +484 -0
  120. data/vendor/kreuzberg/src/extractors/security_tests.rs +367 -0
  121. data/vendor/kreuzberg/src/extractors/structured.rs +140 -126
  122. data/vendor/kreuzberg/src/extractors/text.rs +260 -242
  123. data/vendor/kreuzberg/src/extractors/typst.rs +650 -0
  124. data/vendor/kreuzberg/src/extractors/xml.rs +135 -128
  125. data/vendor/kreuzberg/src/image/dpi.rs +164 -164
  126. data/vendor/kreuzberg/src/image/mod.rs +6 -6
  127. data/vendor/kreuzberg/src/image/preprocessing.rs +417 -417
  128. data/vendor/kreuzberg/src/image/resize.rs +89 -89
  129. data/vendor/kreuzberg/src/keywords/config.rs +154 -154
  130. data/vendor/kreuzberg/src/keywords/mod.rs +237 -237
  131. data/vendor/kreuzberg/src/keywords/processor.rs +267 -267
  132. data/vendor/kreuzberg/src/keywords/rake.rs +293 -294
  133. data/vendor/kreuzberg/src/keywords/types.rs +68 -68
  134. data/vendor/kreuzberg/src/keywords/yake.rs +163 -163
  135. data/vendor/kreuzberg/src/language_detection/mod.rs +942 -942
  136. data/vendor/kreuzberg/src/lib.rs +105 -102
  137. data/vendor/kreuzberg/src/mcp/mod.rs +32 -32
  138. data/vendor/kreuzberg/src/mcp/server.rs +1968 -1966
  139. data/vendor/kreuzberg/src/ocr/cache.rs +469 -469
  140. data/vendor/kreuzberg/src/ocr/error.rs +37 -37
  141. data/vendor/kreuzberg/src/ocr/hocr.rs +216 -216
  142. data/vendor/kreuzberg/src/ocr/mod.rs +58 -58
  143. data/vendor/kreuzberg/src/ocr/processor.rs +863 -847
  144. data/vendor/kreuzberg/src/ocr/table/mod.rs +4 -4
  145. data/vendor/kreuzberg/src/ocr/table/tsv_parser.rs +144 -144
  146. data/vendor/kreuzberg/src/ocr/tesseract_backend.rs +450 -450
  147. data/vendor/kreuzberg/src/ocr/types.rs +393 -393
  148. data/vendor/kreuzberg/src/ocr/utils.rs +47 -47
  149. data/vendor/kreuzberg/src/ocr/validation.rs +206 -206
  150. data/vendor/kreuzberg/src/panic_context.rs +154 -0
  151. data/vendor/kreuzberg/src/pdf/error.rs +122 -122
  152. data/vendor/kreuzberg/src/pdf/images.rs +139 -139
  153. data/vendor/kreuzberg/src/pdf/metadata.rs +346 -346
  154. data/vendor/kreuzberg/src/pdf/mod.rs +50 -50
  155. data/vendor/kreuzberg/src/pdf/rendering.rs +369 -369
  156. data/vendor/kreuzberg/src/pdf/table.rs +393 -420
  157. data/vendor/kreuzberg/src/pdf/text.rs +158 -161
  158. data/vendor/kreuzberg/src/plugins/extractor.rs +1013 -1010
  159. data/vendor/kreuzberg/src/plugins/mod.rs +209 -209
  160. data/vendor/kreuzberg/src/plugins/ocr.rs +620 -629
  161. data/vendor/kreuzberg/src/plugins/processor.rs +642 -641
  162. data/vendor/kreuzberg/src/plugins/registry.rs +1337 -1324
  163. data/vendor/kreuzberg/src/plugins/traits.rs +258 -258
  164. data/vendor/kreuzberg/src/plugins/validator.rs +956 -955
  165. data/vendor/kreuzberg/src/stopwords/mod.rs +1470 -1470
  166. data/vendor/kreuzberg/src/text/mod.rs +19 -19
  167. data/vendor/kreuzberg/src/text/quality.rs +697 -697
  168. data/vendor/kreuzberg/src/text/string_utils.rs +217 -217
  169. data/vendor/kreuzberg/src/text/token_reduction/cjk_utils.rs +164 -164
  170. data/vendor/kreuzberg/src/text/token_reduction/config.rs +100 -100
  171. data/vendor/kreuzberg/src/text/token_reduction/core.rs +796 -796
  172. data/vendor/kreuzberg/src/text/token_reduction/filters.rs +902 -902
  173. data/vendor/kreuzberg/src/text/token_reduction/mod.rs +160 -160
  174. data/vendor/kreuzberg/src/text/token_reduction/semantic.rs +619 -619
  175. data/vendor/kreuzberg/src/text/token_reduction/simd_text.rs +147 -147
  176. data/vendor/kreuzberg/src/types.rs +903 -873
  177. data/vendor/kreuzberg/src/utils/mod.rs +17 -17
  178. data/vendor/kreuzberg/src/utils/quality.rs +959 -959
  179. data/vendor/kreuzberg/src/utils/string_utils.rs +381 -381
  180. data/vendor/kreuzberg/stopwords/af_stopwords.json +53 -53
  181. data/vendor/kreuzberg/stopwords/ar_stopwords.json +482 -482
  182. data/vendor/kreuzberg/stopwords/bg_stopwords.json +261 -261
  183. data/vendor/kreuzberg/stopwords/bn_stopwords.json +400 -400
  184. data/vendor/kreuzberg/stopwords/br_stopwords.json +1205 -1205
  185. data/vendor/kreuzberg/stopwords/ca_stopwords.json +280 -280
  186. data/vendor/kreuzberg/stopwords/cs_stopwords.json +425 -425
  187. data/vendor/kreuzberg/stopwords/da_stopwords.json +172 -172
  188. data/vendor/kreuzberg/stopwords/de_stopwords.json +622 -622
  189. data/vendor/kreuzberg/stopwords/el_stopwords.json +849 -849
  190. data/vendor/kreuzberg/stopwords/en_stopwords.json +1300 -1300
  191. data/vendor/kreuzberg/stopwords/eo_stopwords.json +175 -175
  192. data/vendor/kreuzberg/stopwords/es_stopwords.json +734 -734
  193. data/vendor/kreuzberg/stopwords/et_stopwords.json +37 -37
  194. data/vendor/kreuzberg/stopwords/eu_stopwords.json +100 -100
  195. data/vendor/kreuzberg/stopwords/fa_stopwords.json +801 -801
  196. data/vendor/kreuzberg/stopwords/fi_stopwords.json +849 -849
  197. data/vendor/kreuzberg/stopwords/fr_stopwords.json +693 -693
  198. data/vendor/kreuzberg/stopwords/ga_stopwords.json +111 -111
  199. data/vendor/kreuzberg/stopwords/gl_stopwords.json +162 -162
  200. data/vendor/kreuzberg/stopwords/gu_stopwords.json +226 -226
  201. data/vendor/kreuzberg/stopwords/ha_stopwords.json +41 -41
  202. data/vendor/kreuzberg/stopwords/he_stopwords.json +196 -196
  203. data/vendor/kreuzberg/stopwords/hi_stopwords.json +227 -227
  204. data/vendor/kreuzberg/stopwords/hr_stopwords.json +181 -181
  205. data/vendor/kreuzberg/stopwords/hu_stopwords.json +791 -791
  206. data/vendor/kreuzberg/stopwords/hy_stopwords.json +47 -47
  207. data/vendor/kreuzberg/stopwords/id_stopwords.json +760 -760
  208. data/vendor/kreuzberg/stopwords/it_stopwords.json +634 -634
  209. data/vendor/kreuzberg/stopwords/ja_stopwords.json +136 -136
  210. data/vendor/kreuzberg/stopwords/kn_stopwords.json +84 -84
  211. data/vendor/kreuzberg/stopwords/ko_stopwords.json +681 -681
  212. data/vendor/kreuzberg/stopwords/ku_stopwords.json +64 -64
  213. data/vendor/kreuzberg/stopwords/la_stopwords.json +51 -51
  214. data/vendor/kreuzberg/stopwords/lt_stopwords.json +476 -476
  215. data/vendor/kreuzberg/stopwords/lv_stopwords.json +163 -163
  216. data/vendor/kreuzberg/stopwords/ml_stopwords.json +1 -1
  217. data/vendor/kreuzberg/stopwords/mr_stopwords.json +101 -101
  218. data/vendor/kreuzberg/stopwords/ms_stopwords.json +477 -477
  219. data/vendor/kreuzberg/stopwords/ne_stopwords.json +490 -490
  220. data/vendor/kreuzberg/stopwords/nl_stopwords.json +415 -415
  221. data/vendor/kreuzberg/stopwords/no_stopwords.json +223 -223
  222. data/vendor/kreuzberg/stopwords/pl_stopwords.json +331 -331
  223. data/vendor/kreuzberg/stopwords/pt_stopwords.json +562 -562
  224. data/vendor/kreuzberg/stopwords/ro_stopwords.json +436 -436
  225. data/vendor/kreuzberg/stopwords/ru_stopwords.json +561 -561
  226. data/vendor/kreuzberg/stopwords/si_stopwords.json +193 -193
  227. data/vendor/kreuzberg/stopwords/sk_stopwords.json +420 -420
  228. data/vendor/kreuzberg/stopwords/sl_stopwords.json +448 -448
  229. data/vendor/kreuzberg/stopwords/so_stopwords.json +32 -32
  230. data/vendor/kreuzberg/stopwords/st_stopwords.json +33 -33
  231. data/vendor/kreuzberg/stopwords/sv_stopwords.json +420 -420
  232. data/vendor/kreuzberg/stopwords/sw_stopwords.json +76 -76
  233. data/vendor/kreuzberg/stopwords/ta_stopwords.json +129 -129
  234. data/vendor/kreuzberg/stopwords/te_stopwords.json +54 -54
  235. data/vendor/kreuzberg/stopwords/th_stopwords.json +118 -118
  236. data/vendor/kreuzberg/stopwords/tl_stopwords.json +149 -149
  237. data/vendor/kreuzberg/stopwords/tr_stopwords.json +506 -506
  238. data/vendor/kreuzberg/stopwords/uk_stopwords.json +75 -75
  239. data/vendor/kreuzberg/stopwords/ur_stopwords.json +519 -519
  240. data/vendor/kreuzberg/stopwords/vi_stopwords.json +647 -647
  241. data/vendor/kreuzberg/stopwords/yo_stopwords.json +62 -62
  242. data/vendor/kreuzberg/stopwords/zh_stopwords.json +796 -796
  243. data/vendor/kreuzberg/stopwords/zu_stopwords.json +31 -31
  244. data/vendor/kreuzberg/tests/api_extract_multipart.rs +52 -0
  245. data/vendor/kreuzberg/tests/api_tests.rs +966 -966
  246. data/vendor/kreuzberg/tests/archive_integration.rs +543 -543
  247. data/vendor/kreuzberg/tests/batch_orchestration.rs +556 -542
  248. data/vendor/kreuzberg/tests/batch_processing.rs +316 -304
  249. data/vendor/kreuzberg/tests/bibtex_parity_test.rs +421 -0
  250. data/vendor/kreuzberg/tests/concurrency_stress.rs +525 -509
  251. data/vendor/kreuzberg/tests/config_features.rs +598 -580
  252. data/vendor/kreuzberg/tests/config_loading_tests.rs +415 -439
  253. data/vendor/kreuzberg/tests/core_integration.rs +510 -493
  254. data/vendor/kreuzberg/tests/csv_integration.rs +414 -424
  255. data/vendor/kreuzberg/tests/docbook_extractor_tests.rs +498 -0
  256. data/vendor/kreuzberg/tests/docx_metadata_extraction_test.rs +122 -124
  257. data/vendor/kreuzberg/tests/docx_vs_pandoc_comparison.rs +370 -0
  258. data/vendor/kreuzberg/tests/email_integration.rs +325 -325
  259. data/vendor/kreuzberg/tests/epub_native_extractor_tests.rs +275 -0
  260. data/vendor/kreuzberg/tests/error_handling.rs +393 -393
  261. data/vendor/kreuzberg/tests/fictionbook_extractor_tests.rs +228 -0
  262. data/vendor/kreuzberg/tests/format_integration.rs +159 -159
  263. data/vendor/kreuzberg/tests/helpers/mod.rs +142 -142
  264. data/vendor/kreuzberg/tests/html_table_test.rs +551 -0
  265. data/vendor/kreuzberg/tests/image_integration.rs +253 -253
  266. data/vendor/kreuzberg/tests/instrumentation_test.rs +139 -0
  267. data/vendor/kreuzberg/tests/jats_extractor_tests.rs +639 -0
  268. data/vendor/kreuzberg/tests/jupyter_extractor_tests.rs +704 -0
  269. data/vendor/kreuzberg/tests/keywords_integration.rs +479 -479
  270. data/vendor/kreuzberg/tests/keywords_quality.rs +509 -509
  271. data/vendor/kreuzberg/tests/latex_extractor_tests.rs +496 -0
  272. data/vendor/kreuzberg/tests/markdown_extractor_tests.rs +490 -0
  273. data/vendor/kreuzberg/tests/mime_detection.rs +428 -428
  274. data/vendor/kreuzberg/tests/ocr_configuration.rs +510 -510
  275. data/vendor/kreuzberg/tests/ocr_errors.rs +676 -676
  276. data/vendor/kreuzberg/tests/ocr_quality.rs +627 -627
  277. data/vendor/kreuzberg/tests/ocr_stress.rs +469 -469
  278. data/vendor/kreuzberg/tests/odt_extractor_tests.rs +695 -0
  279. data/vendor/kreuzberg/tests/opml_extractor_tests.rs +616 -0
  280. data/vendor/kreuzberg/tests/orgmode_extractor_tests.rs +822 -0
  281. data/vendor/kreuzberg/tests/pdf_integration.rs +43 -43
  282. data/vendor/kreuzberg/tests/pipeline_integration.rs +1411 -1412
  283. data/vendor/kreuzberg/tests/plugin_ocr_backend_test.rs +771 -771
  284. data/vendor/kreuzberg/tests/plugin_postprocessor_test.rs +560 -561
  285. data/vendor/kreuzberg/tests/plugin_system.rs +921 -921
  286. data/vendor/kreuzberg/tests/plugin_validator_test.rs +783 -783
  287. data/vendor/kreuzberg/tests/registry_integration_tests.rs +586 -607
  288. data/vendor/kreuzberg/tests/rst_extractor_tests.rs +692 -0
  289. data/vendor/kreuzberg/tests/rtf_extractor_tests.rs +776 -0
  290. data/vendor/kreuzberg/tests/security_validation.rs +415 -404
  291. data/vendor/kreuzberg/tests/stopwords_integration_test.rs +888 -888
  292. data/vendor/kreuzberg/tests/test_fastembed.rs +609 -609
  293. data/vendor/kreuzberg/tests/typst_behavioral_tests.rs +1259 -0
  294. data/vendor/kreuzberg/tests/typst_extractor_tests.rs +647 -0
  295. data/vendor/kreuzberg/tests/xlsx_metadata_extraction_test.rs +87 -87
  296. data/vendor/rb-sys/.cargo-ok +1 -0
  297. data/vendor/rb-sys/.cargo_vcs_info.json +6 -0
  298. data/vendor/rb-sys/Cargo.lock +393 -0
  299. data/vendor/rb-sys/Cargo.toml +70 -0
  300. data/vendor/rb-sys/Cargo.toml.orig +57 -0
  301. data/vendor/rb-sys/LICENSE-APACHE +190 -0
  302. data/vendor/rb-sys/LICENSE-MIT +21 -0
  303. data/vendor/rb-sys/bin/release.sh +21 -0
  304. data/vendor/rb-sys/build/features.rs +108 -0
  305. data/vendor/rb-sys/build/main.rs +246 -0
  306. data/vendor/rb-sys/build/stable_api_config.rs +153 -0
  307. data/vendor/rb-sys/build/version.rs +48 -0
  308. data/vendor/rb-sys/readme.md +36 -0
  309. data/vendor/rb-sys/src/bindings.rs +21 -0
  310. data/vendor/rb-sys/src/hidden.rs +11 -0
  311. data/vendor/rb-sys/src/lib.rs +34 -0
  312. data/vendor/rb-sys/src/macros.rs +371 -0
  313. data/vendor/rb-sys/src/memory.rs +53 -0
  314. data/vendor/rb-sys/src/ruby_abi_version.rs +38 -0
  315. data/vendor/rb-sys/src/special_consts.rs +31 -0
  316. data/vendor/rb-sys/src/stable_api/compiled.c +179 -0
  317. data/vendor/rb-sys/src/stable_api/compiled.rs +257 -0
  318. data/vendor/rb-sys/src/stable_api/ruby_2_6.rs +316 -0
  319. data/vendor/rb-sys/src/stable_api/ruby_2_7.rs +316 -0
  320. data/vendor/rb-sys/src/stable_api/ruby_3_0.rs +324 -0
  321. data/vendor/rb-sys/src/stable_api/ruby_3_1.rs +317 -0
  322. data/vendor/rb-sys/src/stable_api/ruby_3_2.rs +315 -0
  323. data/vendor/rb-sys/src/stable_api/ruby_3_3.rs +326 -0
  324. data/vendor/rb-sys/src/stable_api/ruby_3_4.rs +327 -0
  325. data/vendor/rb-sys/src/stable_api.rs +261 -0
  326. data/vendor/rb-sys/src/symbol.rs +31 -0
  327. data/vendor/rb-sys/src/tracking_allocator.rs +332 -0
  328. data/vendor/rb-sys/src/utils.rs +89 -0
  329. data/vendor/rb-sys/src/value_type.rs +7 -0
  330. metadata +90 -95
  331. data/pkg/kreuzberg-4.0.0.rc1.gem +0 -0
  332. data/spec/examples.txt +0 -104
  333. data/vendor/kreuzberg/src/bin/profile_extract.rs +0 -455
  334. data/vendor/kreuzberg/src/extraction/pandoc/batch.rs +0 -275
  335. data/vendor/kreuzberg/src/extraction/pandoc/mime_types.rs +0 -178
  336. data/vendor/kreuzberg/src/extraction/pandoc/mod.rs +0 -491
  337. data/vendor/kreuzberg/src/extraction/pandoc/server.rs +0 -496
  338. data/vendor/kreuzberg/src/extraction/pandoc/subprocess.rs +0 -1188
  339. data/vendor/kreuzberg/src/extraction/pandoc/version.rs +0 -162
  340. data/vendor/kreuzberg/src/extractors/pandoc.rs +0 -201
  341. data/vendor/kreuzberg/tests/chunking_offset_demo.rs +0 -92
  342. data/vendor/kreuzberg/tests/pandoc_integration.rs +0 -503
@@ -1,580 +1,598 @@
1
- //! Configuration features integration tests.
2
- //!
3
- //! Tests for chunking, language detection, caching, token reduction, and quality processing.
4
- //! Validates that configuration options work correctly end-to-end.
5
-
6
- use kreuzberg::core::config::{ChunkingConfig, ExtractionConfig, LanguageDetectionConfig, TokenReductionConfig};
7
- use kreuzberg::core::extractor::extract_bytes;
8
-
9
- mod helpers;
10
-
11
- /// Test chunking enabled - text split into chunks.
12
- #[tokio::test]
13
- async fn test_chunking_enabled() {
14
- let config = ExtractionConfig {
15
- chunking: Some(ChunkingConfig {
16
- max_chars: 50,
17
- max_overlap: 10,
18
- embedding: None,
19
- preset: None,
20
- }),
21
- ..Default::default()
22
- };
23
-
24
- let text = "This is a long text that should be split into multiple chunks. ".repeat(10);
25
- let text_bytes = text.as_bytes();
26
-
27
- let result = extract_bytes(text_bytes, "text/plain", &config)
28
- .await
29
- .expect("Should extract successfully");
30
-
31
- assert!(result.chunks.is_some(), "Chunks should be present");
32
- let chunks = result.chunks.unwrap();
33
- assert!(chunks.len() > 1, "Should have multiple chunks");
34
-
35
- assert!(result.metadata.additional.contains_key("chunk_count"));
36
- let chunk_count = result.metadata.additional.get("chunk_count").unwrap();
37
- assert_eq!(
38
- chunks.len(),
39
- chunk_count.as_u64().unwrap() as usize,
40
- "Chunks length should match chunk_count metadata"
41
- );
42
-
43
- for chunk in &chunks {
44
- assert!(!chunk.content.is_empty(), "Chunk should not be empty");
45
- assert!(
46
- chunk.content.len() <= 50 + 10,
47
- "Chunk length {} exceeds max_chars + overlap",
48
- chunk.content.len()
49
- );
50
- }
51
- }
52
-
53
- /// Test chunking with overlap - overlap preserved between chunks.
54
- #[tokio::test]
55
- async fn test_chunking_with_overlap() {
56
- let config = ExtractionConfig {
57
- chunking: Some(ChunkingConfig {
58
- max_chars: 100,
59
- max_overlap: 20,
60
- embedding: None,
61
- preset: None,
62
- }),
63
- ..Default::default()
64
- };
65
-
66
- let text = "a".repeat(250);
67
- let text_bytes = text.as_bytes();
68
-
69
- let result = extract_bytes(text_bytes, "text/plain", &config)
70
- .await
71
- .expect("Should extract successfully");
72
-
73
- assert!(result.chunks.is_some(), "Chunks should be present");
74
- let chunks = result.chunks.unwrap();
75
- assert!(chunks.len() >= 2, "Should have at least 2 chunks");
76
-
77
- assert!(result.metadata.additional.contains_key("chunk_count"));
78
-
79
- if chunks.len() >= 2 {
80
- let chunk1 = &chunks[0];
81
- let chunk2 = &chunks[1];
82
-
83
- let chunk1_end = &chunk1.content[chunk1.content.len().saturating_sub(20)..];
84
- assert!(
85
- chunk2.content.starts_with(chunk1_end)
86
- || chunk1_end.starts_with(&chunk2.content[..chunk1_end.len().min(chunk2.content.len())]),
87
- "Chunks should have overlap"
88
- );
89
- }
90
- }
91
-
92
- /// Test chunking with custom sizes - custom chunk size and overlap.
93
- #[tokio::test]
94
- async fn test_chunking_custom_sizes() {
95
- let config = ExtractionConfig {
96
- chunking: Some(ChunkingConfig {
97
- max_chars: 200,
98
- max_overlap: 50,
99
- embedding: None,
100
- preset: None,
101
- }),
102
- ..Default::default()
103
- };
104
-
105
- let text = "Custom chunk test. ".repeat(50);
106
- let text_bytes = text.as_bytes();
107
-
108
- let result = extract_bytes(text_bytes, "text/plain", &config)
109
- .await
110
- .expect("Should extract successfully");
111
-
112
- assert!(result.chunks.is_some(), "Chunks should be present");
113
- let chunks = result.chunks.unwrap();
114
- assert!(!chunks.is_empty(), "Should have at least 1 chunk");
115
-
116
- assert!(result.metadata.additional.contains_key("chunk_count"));
117
-
118
- for chunk in &chunks {
119
- assert!(
120
- chunk.content.len() <= 200 + 50,
121
- "Chunk length {} exceeds custom max_chars + overlap",
122
- chunk.content.len()
123
- );
124
- }
125
- }
126
-
127
- /// Test chunking disabled - no chunking when disabled.
128
- #[tokio::test]
129
- async fn test_chunking_disabled() {
130
- let config = ExtractionConfig {
131
- chunking: None,
132
- ..Default::default()
133
- };
134
-
135
- let text = "This is a long text that should NOT be split into chunks. ".repeat(10);
136
- let text_bytes = text.as_bytes();
137
-
138
- let result = extract_bytes(text_bytes, "text/plain", &config)
139
- .await
140
- .expect("Should extract successfully");
141
-
142
- assert!(result.chunks.is_none(), "Should not have chunks when chunking disabled");
143
- assert!(
144
- !result.metadata.additional.contains_key("chunk_count"),
145
- "Should not have chunk_count when chunking disabled"
146
- );
147
-
148
- assert!(!result.content.is_empty(), "Content should be extracted");
149
- assert!(result.content.contains("long text"), "Should contain original text");
150
- }
151
-
152
- /// Test language detection for single language document.
153
- #[tokio::test]
154
- async fn test_language_detection_single() {
155
- let config = ExtractionConfig {
156
- language_detection: Some(LanguageDetectionConfig {
157
- enabled: true,
158
- min_confidence: 0.8,
159
- detect_multiple: false,
160
- }),
161
- ..Default::default()
162
- };
163
-
164
- let text = "Hello world! This is English text. It should be detected as English language.";
165
- let text_bytes = text.as_bytes();
166
-
167
- let result = extract_bytes(text_bytes, "text/plain", &config)
168
- .await
169
- .expect("Should extract successfully");
170
-
171
- assert!(result.detected_languages.is_some(), "Should detect language");
172
- let languages = result.detected_languages.unwrap();
173
- assert!(!languages.is_empty(), "Should detect at least one language");
174
- assert_eq!(languages[0], "eng", "Should detect English");
175
- }
176
-
177
- /// Test language detection for multi-language document.
178
- #[cfg_attr(coverage, ignore = "coverage instrumentation affects multi-language heuristics")]
179
- #[tokio::test]
180
- async fn test_language_detection_multiple() {
181
- let config = ExtractionConfig {
182
- language_detection: Some(LanguageDetectionConfig {
183
- enabled: true,
184
- min_confidence: 0.7,
185
- detect_multiple: true,
186
- }),
187
- ..Default::default()
188
- };
189
-
190
- let text = "Hello world! This is English. ".repeat(10) + "Hola mundo! Este es español. ".repeat(10).as_str();
191
- let text_bytes = text.as_bytes();
192
-
193
- let result = extract_bytes(text_bytes, "text/plain", &config)
194
- .await
195
- .expect("Should extract successfully");
196
-
197
- assert!(result.detected_languages.is_some(), "Should detect languages");
198
- let languages = result.detected_languages.unwrap();
199
- assert!(!languages.is_empty(), "Should detect at least one language");
200
- }
201
-
202
- /// Test language detection with confidence threshold.
203
- #[tokio::test]
204
- async fn test_language_detection_confidence() {
205
- let config = ExtractionConfig {
206
- language_detection: Some(LanguageDetectionConfig {
207
- enabled: true,
208
- min_confidence: 0.9,
209
- detect_multiple: false,
210
- }),
211
- ..Default::default()
212
- };
213
-
214
- let text = "This is clear English text that should have high confidence.";
215
- let text_bytes = text.as_bytes();
216
-
217
- let result = extract_bytes(text_bytes, "text/plain", &config)
218
- .await
219
- .expect("Should extract successfully");
220
-
221
- if let Some(languages) = result.detected_languages {
222
- assert!(!languages.is_empty());
223
- }
224
- }
225
-
226
- /// Test language detection disabled.
227
- #[tokio::test]
228
- async fn test_language_detection_disabled() {
229
- let config = ExtractionConfig {
230
- language_detection: Some(LanguageDetectionConfig {
231
- enabled: false,
232
- min_confidence: 0.8,
233
- detect_multiple: false,
234
- }),
235
- ..Default::default()
236
- };
237
-
238
- let text = "Hello world! This is English text.";
239
- let text_bytes = text.as_bytes();
240
-
241
- let result = extract_bytes(text_bytes, "text/plain", &config)
242
- .await
243
- .expect("Should extract successfully");
244
-
245
- assert!(
246
- result.detected_languages.is_none(),
247
- "Should not detect language when disabled"
248
- );
249
- }
250
-
251
- /// Test cache hit behavior - second extraction from cache.
252
- #[tokio::test]
253
- async fn test_cache_hit_behavior() {
254
- let config = ExtractionConfig {
255
- use_cache: true,
256
- ..Default::default()
257
- };
258
-
259
- let text = "Test text for caching behavior.";
260
- let text_bytes = text.as_bytes();
261
-
262
- let result1 = extract_bytes(text_bytes, "text/plain", &config)
263
- .await
264
- .expect("First extraction should succeed");
265
-
266
- let result2 = extract_bytes(text_bytes, "text/plain", &config)
267
- .await
268
- .expect("Second extraction should succeed");
269
-
270
- assert_eq!(result1.content, result2.content);
271
- }
272
-
273
- /// Test cache miss and invalidation.
274
- #[tokio::test]
275
- async fn test_cache_miss_invalidation() {
276
- let config = ExtractionConfig {
277
- use_cache: true,
278
- ..Default::default()
279
- };
280
-
281
- let text1 = "First text for cache test.";
282
- let text2 = "Second different text.";
283
-
284
- let result1 = extract_bytes(text1.as_bytes(), "text/plain", &config)
285
- .await
286
- .expect("First extraction should succeed");
287
-
288
- let result2 = extract_bytes(text2.as_bytes(), "text/plain", &config)
289
- .await
290
- .expect("Second extraction should succeed");
291
-
292
- assert_ne!(result1.content, result2.content);
293
- }
294
-
295
- /// Test custom cache directory (Note: OCR cache uses hardcoded directory).
296
- #[tokio::test]
297
- async fn test_custom_cache_directory() {
298
- let config = ExtractionConfig {
299
- use_cache: true,
300
- ..Default::default()
301
- };
302
-
303
- let text = "Test text for cache directory test.";
304
- let text_bytes = text.as_bytes();
305
-
306
- let result = extract_bytes(text_bytes, "text/plain", &config)
307
- .await
308
- .expect("Should extract successfully");
309
-
310
- assert!(!result.content.is_empty());
311
- }
312
-
313
- /// Test cache disabled - bypass cache.
314
- #[tokio::test]
315
- async fn test_cache_disabled() {
316
- let config = ExtractionConfig {
317
- use_cache: false,
318
- ..Default::default()
319
- };
320
-
321
- let text = "Test text without caching.";
322
- let text_bytes = text.as_bytes();
323
-
324
- let result1 = extract_bytes(text_bytes, "text/plain", &config)
325
- .await
326
- .expect("First extraction should succeed");
327
-
328
- let result2 = extract_bytes(text_bytes, "text/plain", &config)
329
- .await
330
- .expect("Second extraction should succeed");
331
-
332
- assert_eq!(result1.content, result2.content);
333
- }
334
-
335
- /// Test token reduction in aggressive mode.
336
- #[tokio::test]
337
- async fn test_token_reduction_aggressive() {
338
- let config = ExtractionConfig {
339
- token_reduction: Some(TokenReductionConfig {
340
- mode: "aggressive".to_string(),
341
- preserve_important_words: true,
342
- }),
343
- ..Default::default()
344
- };
345
-
346
- let text = "This is a very long sentence with many unnecessary words that could be reduced. ".repeat(5);
347
- let text_bytes = text.as_bytes();
348
-
349
- let result = extract_bytes(text_bytes, "text/plain", &config)
350
- .await
351
- .expect("Should extract successfully");
352
-
353
- assert!(!result.content.is_empty());
354
- }
355
-
356
- /// Test token reduction in conservative mode.
357
- #[tokio::test]
358
- async fn test_token_reduction_conservative() {
359
- let config = ExtractionConfig {
360
- token_reduction: Some(TokenReductionConfig {
361
- mode: "light".to_string(),
362
- preserve_important_words: true,
363
- }),
364
- ..Default::default()
365
- };
366
-
367
- let text = "Conservative token reduction test with moderate text length.";
368
- let text_bytes = text.as_bytes();
369
-
370
- let result = extract_bytes(text_bytes, "text/plain", &config)
371
- .await
372
- .expect("Should extract successfully");
373
-
374
- assert!(!result.content.is_empty());
375
- }
376
-
377
- /// Test token reduction disabled.
378
- #[tokio::test]
379
- async fn test_token_reduction_disabled() {
380
- let config = ExtractionConfig {
381
- token_reduction: Some(TokenReductionConfig {
382
- mode: "off".to_string(),
383
- preserve_important_words: false,
384
- }),
385
- ..Default::default()
386
- };
387
-
388
- let text = "Text without token reduction applied.";
389
- let text_bytes = text.as_bytes();
390
-
391
- let result = extract_bytes(text_bytes, "text/plain", &config)
392
- .await
393
- .expect("Should extract successfully");
394
-
395
- assert!(result.content.contains("without token reduction"));
396
- }
397
-
398
- /// Test quality processing enabled - quality scoring applied.
399
- #[tokio::test]
400
- async fn test_quality_processing_enabled() {
401
- let config = ExtractionConfig {
402
- enable_quality_processing: true,
403
- ..Default::default()
404
- };
405
-
406
- let text = "This is well-structured text. It has multiple sentences. And proper punctuation.";
407
- let text_bytes = text.as_bytes();
408
-
409
- let result = extract_bytes(text_bytes, "text/plain", &config)
410
- .await
411
- .expect("Should extract successfully");
412
-
413
- if let Some(score) = result.metadata.additional.get("quality_score") {
414
- let score_value = score.as_f64().unwrap();
415
- assert!((0.0..=1.0).contains(&score_value));
416
- }
417
-
418
- assert!(!result.content.is_empty());
419
- }
420
-
421
- /// Test quality processing calculates score for different text quality.
422
- #[tokio::test]
423
- async fn test_quality_threshold_filtering() {
424
- let config = ExtractionConfig {
425
- enable_quality_processing: true,
426
- ..Default::default()
427
- };
428
-
429
- let high_quality = "This is a well-structured document. It has proper sentences. And good formatting.";
430
- let result_high = extract_bytes(high_quality.as_bytes(), "text/plain", &config)
431
- .await
432
- .expect("Should extract successfully");
433
-
434
- let low_quality = "a b c d ....... word123mixed . . ";
435
- let result_low = extract_bytes(low_quality.as_bytes(), "text/plain", &config)
436
- .await
437
- .expect("Should extract successfully");
438
-
439
- assert!(
440
- result_high.metadata.additional.contains_key("quality_score"),
441
- "High quality should have score"
442
- );
443
- assert!(
444
- result_low.metadata.additional.contains_key("quality_score"),
445
- "Low quality should have score"
446
- );
447
-
448
- let score_high = result_high
449
- .metadata
450
- .additional
451
- .get("quality_score")
452
- .unwrap()
453
- .as_f64()
454
- .unwrap();
455
- let score_low = result_low
456
- .metadata
457
- .additional
458
- .get("quality_score")
459
- .unwrap()
460
- .as_f64()
461
- .unwrap();
462
-
463
- assert!((0.0..=1.0).contains(&score_high));
464
- assert!((0.0..=1.0).contains(&score_low));
465
- }
466
-
467
- /// Test quality processing disabled.
468
- #[tokio::test]
469
- async fn test_quality_processing_disabled() {
470
- let config = ExtractionConfig {
471
- enable_quality_processing: false,
472
- ..Default::default()
473
- };
474
-
475
- let text = "Text without quality processing.";
476
- let text_bytes = text.as_bytes();
477
-
478
- let result = extract_bytes(text_bytes, "text/plain", &config)
479
- .await
480
- .expect("Should extract successfully");
481
-
482
- assert!(!result.metadata.additional.contains_key("quality_score"));
483
- assert!(!result.content.is_empty());
484
- }
485
-
486
- /// Test chunking with embeddings using balanced preset.
487
- #[tokio::test]
488
- #[cfg(feature = "embeddings")]
489
- async fn test_chunking_with_embeddings() {
490
- use kreuzberg::core::config::EmbeddingConfig;
491
-
492
- let config = ExtractionConfig {
493
- chunking: Some(ChunkingConfig {
494
- max_chars: 100,
495
- max_overlap: 20,
496
- embedding: Some(EmbeddingConfig::default()),
497
- preset: None,
498
- }),
499
- ..Default::default()
500
- };
501
-
502
- let text = "This is a test document for embedding generation. ".repeat(10);
503
- let text_bytes = text.as_bytes();
504
-
505
- let result = extract_bytes(text_bytes, "text/plain", &config)
506
- .await
507
- .expect("Should extract successfully");
508
-
509
- assert!(result.chunks.is_some(), "Chunks should be present");
510
- let chunks = result.chunks.unwrap();
511
- assert!(chunks.len() > 1, "Should have multiple chunks");
512
-
513
- println!("Metadata: {:?}", result.metadata.additional);
514
-
515
- if let Some(error) = result.metadata.additional.get("embedding_error") {
516
- panic!("Embedding generation failed: {}", error);
517
- }
518
-
519
- assert!(
520
- result.metadata.additional.contains_key("embeddings_generated"),
521
- "Should have embeddings_generated metadata"
522
- );
523
- assert_eq!(
524
- result.metadata.additional.get("embeddings_generated").unwrap(),
525
- &serde_json::Value::Bool(true)
526
- );
527
-
528
- for chunk in &chunks {
529
- assert!(chunk.embedding.is_some(), "Each chunk should have an embedding");
530
- let embedding = chunk.embedding.as_ref().unwrap();
531
- assert_eq!(
532
- embedding.len(),
533
- 768,
534
- "Embedding should have 768 dimensions for balanced preset"
535
- );
536
-
537
- let magnitude: f32 = embedding.iter().map(|x| x * x).sum::<f32>().sqrt();
538
- assert!(
539
- (magnitude - 1.0).abs() < 0.01,
540
- "Embedding should be normalized (magnitude ~= 1.0)"
541
- );
542
- }
543
- }
544
-
545
- /// Test chunking with fast embedding preset.
546
- #[tokio::test]
547
- #[cfg(feature = "embeddings")]
548
- async fn test_chunking_with_fast_embeddings() {
549
- use kreuzberg::core::config::{EmbeddingConfig, EmbeddingModelType};
550
-
551
- let config = ExtractionConfig {
552
- chunking: Some(ChunkingConfig {
553
- max_chars: 100,
554
- max_overlap: 20,
555
- embedding: Some(EmbeddingConfig {
556
- model: EmbeddingModelType::Preset {
557
- name: "fast".to_string(),
558
- },
559
- ..Default::default()
560
- }),
561
- preset: None,
562
- }),
563
- ..Default::default()
564
- };
565
-
566
- let text = "Fast embedding test. ".repeat(10);
567
- let text_bytes = text.as_bytes();
568
-
569
- let result = extract_bytes(text_bytes, "text/plain", &config)
570
- .await
571
- .expect("Should extract successfully");
572
-
573
- let chunks = result.chunks.expect("Should have chunks");
574
- assert!(!chunks.is_empty(), "Should have at least one chunk");
575
-
576
- for chunk in &chunks {
577
- let embedding = chunk.embedding.as_ref().expect("Should have embedding");
578
- assert_eq!(embedding.len(), 384, "Fast preset should produce 384-dim embeddings");
579
- }
580
- }
1
+ //! Configuration features integration tests.
2
+ //!
3
+ //! Tests for chunking, language detection, caching, token reduction, and quality processing.
4
+ //! Validates that configuration options work correctly end-to-end.
5
+
6
+ use kreuzberg::core::config::{ChunkingConfig, ExtractionConfig, LanguageDetectionConfig, TokenReductionConfig};
7
+ use kreuzberg::core::extractor::extract_bytes;
8
+
9
+ mod helpers;
10
+
11
+ /// Test chunking enabled - text split into chunks.
12
+ #[tokio::test]
13
+ async fn test_chunking_enabled() {
14
+ let config = ExtractionConfig {
15
+ chunking: Some(ChunkingConfig {
16
+ max_chars: 50,
17
+ max_overlap: 10,
18
+ embedding: None,
19
+ preset: None,
20
+ }),
21
+ ..Default::default()
22
+ };
23
+
24
+ let text = "This is a long text that should be split into multiple chunks. ".repeat(10);
25
+ let text_bytes = text.as_bytes();
26
+
27
+ let result = extract_bytes(text_bytes, "text/plain", &config)
28
+ .await
29
+ .expect("Should extract successfully");
30
+
31
+ assert!(result.chunks.is_some(), "Chunks should be present");
32
+ let chunks = result.chunks.unwrap();
33
+ assert!(chunks.len() > 1, "Should have multiple chunks");
34
+
35
+ assert!(result.metadata.additional.contains_key("chunk_count"));
36
+ let chunk_count = result.metadata.additional.get("chunk_count").unwrap();
37
+ assert_eq!(
38
+ chunks.len(),
39
+ chunk_count.as_u64().unwrap() as usize,
40
+ "Chunks length should match chunk_count metadata"
41
+ );
42
+
43
+ for chunk in &chunks {
44
+ assert!(!chunk.content.is_empty(), "Chunk should not be empty");
45
+ assert!(
46
+ chunk.content.len() <= 50 + 10,
47
+ "Chunk length {} exceeds max_chars + overlap",
48
+ chunk.content.len()
49
+ );
50
+ }
51
+ }
52
+
53
+ /// Test chunking with overlap - overlap preserved between chunks.
54
+ #[tokio::test]
55
+ async fn test_chunking_with_overlap() {
56
+ let config = ExtractionConfig {
57
+ chunking: Some(ChunkingConfig {
58
+ max_chars: 100,
59
+ max_overlap: 20,
60
+ embedding: None,
61
+ preset: None,
62
+ }),
63
+ ..Default::default()
64
+ };
65
+
66
+ let text = "a".repeat(250);
67
+ let text_bytes = text.as_bytes();
68
+
69
+ let result = extract_bytes(text_bytes, "text/plain", &config)
70
+ .await
71
+ .expect("Should extract successfully");
72
+
73
+ assert!(result.chunks.is_some(), "Chunks should be present");
74
+ let chunks = result.chunks.unwrap();
75
+ assert!(chunks.len() >= 2, "Should have at least 2 chunks");
76
+
77
+ assert!(result.metadata.additional.contains_key("chunk_count"));
78
+
79
+ if chunks.len() >= 2 {
80
+ let chunk1 = &chunks[0];
81
+ let chunk2 = &chunks[1];
82
+
83
+ let chunk1_end = &chunk1.content[chunk1.content.len().saturating_sub(20)..];
84
+ assert!(
85
+ chunk2.content.starts_with(chunk1_end)
86
+ || chunk1_end.starts_with(&chunk2.content[..chunk1_end.len().min(chunk2.content.len())]),
87
+ "Chunks should have overlap"
88
+ );
89
+ }
90
+ }
91
+
92
+ /// Test chunking with custom sizes - custom chunk size and overlap.
93
+ #[tokio::test]
94
+ async fn test_chunking_custom_sizes() {
95
+ let config = ExtractionConfig {
96
+ chunking: Some(ChunkingConfig {
97
+ max_chars: 200,
98
+ max_overlap: 50,
99
+ embedding: None,
100
+ preset: None,
101
+ }),
102
+ ..Default::default()
103
+ };
104
+
105
+ let text = "Custom chunk test. ".repeat(50);
106
+ let text_bytes = text.as_bytes();
107
+
108
+ let result = extract_bytes(text_bytes, "text/plain", &config)
109
+ .await
110
+ .expect("Should extract successfully");
111
+
112
+ assert!(result.chunks.is_some(), "Chunks should be present");
113
+ let chunks = result.chunks.unwrap();
114
+ assert!(!chunks.is_empty(), "Should have at least 1 chunk");
115
+
116
+ assert!(result.metadata.additional.contains_key("chunk_count"));
117
+
118
+ for chunk in &chunks {
119
+ assert!(
120
+ chunk.content.len() <= 200 + 50,
121
+ "Chunk length {} exceeds custom max_chars + overlap",
122
+ chunk.content.len()
123
+ );
124
+ }
125
+ }
126
+
127
+ /// Test chunking disabled - no chunking when disabled.
128
+ #[tokio::test]
129
+ async fn test_chunking_disabled() {
130
+ let config = ExtractionConfig {
131
+ chunking: None,
132
+ ..Default::default()
133
+ };
134
+
135
+ let text = "This is a long text that should NOT be split into chunks. ".repeat(10);
136
+ let text_bytes = text.as_bytes();
137
+
138
+ let result = extract_bytes(text_bytes, "text/plain", &config)
139
+ .await
140
+ .expect("Should extract successfully");
141
+
142
+ assert!(result.chunks.is_none(), "Should not have chunks when chunking disabled");
143
+ assert!(
144
+ !result.metadata.additional.contains_key("chunk_count"),
145
+ "Should not have chunk_count when chunking disabled"
146
+ );
147
+
148
+ assert!(!result.content.is_empty(), "Content should be extracted");
149
+ assert!(result.content.contains("long text"), "Should contain original text");
150
+ }
151
+
152
+ /// Test language detection for single language document.
153
+ #[tokio::test]
154
+ async fn test_language_detection_single() {
155
+ let config = ExtractionConfig {
156
+ language_detection: Some(LanguageDetectionConfig {
157
+ enabled: true,
158
+ min_confidence: 0.8,
159
+ detect_multiple: false,
160
+ }),
161
+ ..Default::default()
162
+ };
163
+
164
+ let text = "Hello world! This is English text. It should be detected as English language.";
165
+ let text_bytes = text.as_bytes();
166
+
167
+ let result = extract_bytes(text_bytes, "text/plain", &config)
168
+ .await
169
+ .expect("Should extract successfully");
170
+
171
+ assert!(result.detected_languages.is_some(), "Should detect language");
172
+ let languages = result.detected_languages.unwrap();
173
+ assert!(!languages.is_empty(), "Should detect at least one language");
174
+ assert_eq!(languages[0], "eng", "Should detect English");
175
+ }
176
+
177
+ /// Test language detection for multi-language document.
178
+ #[cfg_attr(coverage, ignore = "coverage instrumentation affects multi-language heuristics")]
179
+ #[tokio::test]
180
+ async fn test_language_detection_multiple() {
181
+ let config = ExtractionConfig {
182
+ language_detection: Some(LanguageDetectionConfig {
183
+ enabled: true,
184
+ min_confidence: 0.7,
185
+ detect_multiple: true,
186
+ }),
187
+ ..Default::default()
188
+ };
189
+
190
+ let text = "Hello world! This is English. ".repeat(10) + "Hola mundo! Este es español. ".repeat(10).as_str();
191
+ let text_bytes = text.as_bytes();
192
+
193
+ let result = extract_bytes(text_bytes, "text/plain", &config)
194
+ .await
195
+ .expect("Should extract successfully");
196
+
197
+ assert!(result.detected_languages.is_some(), "Should detect languages");
198
+ let languages = result.detected_languages.unwrap();
199
+ assert!(!languages.is_empty(), "Should detect at least one language");
200
+ }
201
+
202
+ /// Test language detection with confidence threshold.
203
+ #[tokio::test]
204
+ async fn test_language_detection_confidence() {
205
+ let config = ExtractionConfig {
206
+ language_detection: Some(LanguageDetectionConfig {
207
+ enabled: true,
208
+ min_confidence: 0.9,
209
+ detect_multiple: false,
210
+ }),
211
+ ..Default::default()
212
+ };
213
+
214
+ let text = "This is clear English text that should have high confidence.";
215
+ let text_bytes = text.as_bytes();
216
+
217
+ let result = extract_bytes(text_bytes, "text/plain", &config)
218
+ .await
219
+ .expect("Should extract successfully");
220
+
221
+ if let Some(languages) = result.detected_languages {
222
+ assert!(!languages.is_empty());
223
+ }
224
+ }
225
+
226
+ /// Test language detection disabled.
227
+ #[tokio::test]
228
+ async fn test_language_detection_disabled() {
229
+ let config = ExtractionConfig {
230
+ language_detection: Some(LanguageDetectionConfig {
231
+ enabled: false,
232
+ min_confidence: 0.8,
233
+ detect_multiple: false,
234
+ }),
235
+ ..Default::default()
236
+ };
237
+
238
+ let text = "Hello world! This is English text.";
239
+ let text_bytes = text.as_bytes();
240
+
241
+ let result = extract_bytes(text_bytes, "text/plain", &config)
242
+ .await
243
+ .expect("Should extract successfully");
244
+
245
+ assert!(
246
+ result.detected_languages.is_none(),
247
+ "Should not detect language when disabled"
248
+ );
249
+ }
250
+
251
+ /// Test cache hit behavior - second extraction from cache.
252
+ #[tokio::test]
253
+ async fn test_cache_hit_behavior() {
254
+ let config = ExtractionConfig {
255
+ use_cache: true,
256
+ ..Default::default()
257
+ };
258
+
259
+ let text = "Test text for caching behavior.";
260
+ let text_bytes = text.as_bytes();
261
+
262
+ let result1 = extract_bytes(text_bytes, "text/plain", &config)
263
+ .await
264
+ .expect("First extraction should succeed");
265
+
266
+ let result2 = extract_bytes(text_bytes, "text/plain", &config)
267
+ .await
268
+ .expect("Second extraction should succeed");
269
+
270
+ assert_eq!(result1.content, result2.content);
271
+ }
272
+
273
+ /// Test cache miss and invalidation.
274
+ #[tokio::test]
275
+ async fn test_cache_miss_invalidation() {
276
+ let config = ExtractionConfig {
277
+ use_cache: true,
278
+ ..Default::default()
279
+ };
280
+
281
+ let text1 = "First text for cache test.";
282
+ let text2 = "Second different text.";
283
+
284
+ let result1 = extract_bytes(text1.as_bytes(), "text/plain", &config)
285
+ .await
286
+ .expect("First extraction should succeed");
287
+
288
+ let result2 = extract_bytes(text2.as_bytes(), "text/plain", &config)
289
+ .await
290
+ .expect("Second extraction should succeed");
291
+
292
+ assert_ne!(result1.content, result2.content);
293
+ }
294
+
295
+ /// Test custom cache directory (Note: OCR cache uses hardcoded directory).
296
+ #[tokio::test]
297
+ async fn test_custom_cache_directory() {
298
+ let config = ExtractionConfig {
299
+ use_cache: true,
300
+ ..Default::default()
301
+ };
302
+
303
+ let text = "Test text for cache directory test.";
304
+ let text_bytes = text.as_bytes();
305
+
306
+ let result = extract_bytes(text_bytes, "text/plain", &config)
307
+ .await
308
+ .expect("Should extract successfully");
309
+
310
+ assert!(!result.content.is_empty());
311
+ }
312
+
313
+ /// Test cache disabled - bypass cache.
314
+ #[tokio::test]
315
+ async fn test_cache_disabled() {
316
+ let config = ExtractionConfig {
317
+ use_cache: false,
318
+ ..Default::default()
319
+ };
320
+
321
+ let text = "Test text without caching.";
322
+ let text_bytes = text.as_bytes();
323
+
324
+ let result1 = extract_bytes(text_bytes, "text/plain", &config)
325
+ .await
326
+ .expect("First extraction should succeed");
327
+
328
+ let result2 = extract_bytes(text_bytes, "text/plain", &config)
329
+ .await
330
+ .expect("Second extraction should succeed");
331
+
332
+ assert_eq!(result1.content, result2.content);
333
+ }
334
+
335
+ /// Test token reduction in aggressive mode.
336
+ #[tokio::test]
337
+ async fn test_token_reduction_aggressive() {
338
+ let config = ExtractionConfig {
339
+ token_reduction: Some(TokenReductionConfig {
340
+ mode: "aggressive".to_string(),
341
+ preserve_important_words: true,
342
+ }),
343
+ ..Default::default()
344
+ };
345
+
346
+ let text = "This is a very long sentence with many unnecessary words that could be reduced. ".repeat(5);
347
+ let text_bytes = text.as_bytes();
348
+
349
+ let result = extract_bytes(text_bytes, "text/plain", &config)
350
+ .await
351
+ .expect("Should extract successfully");
352
+
353
+ assert!(!result.content.is_empty());
354
+ }
355
+
356
+ /// Test token reduction in conservative mode.
357
+ #[tokio::test]
358
+ async fn test_token_reduction_conservative() {
359
+ let config = ExtractionConfig {
360
+ token_reduction: Some(TokenReductionConfig {
361
+ mode: "light".to_string(),
362
+ preserve_important_words: true,
363
+ }),
364
+ ..Default::default()
365
+ };
366
+
367
+ let text = "Conservative token reduction test with moderate text length.";
368
+ let text_bytes = text.as_bytes();
369
+
370
+ let result = extract_bytes(text_bytes, "text/plain", &config)
371
+ .await
372
+ .expect("Should extract successfully");
373
+
374
+ assert!(!result.content.is_empty());
375
+ }
376
+
377
+ /// Test token reduction disabled.
378
+ #[tokio::test]
379
+ async fn test_token_reduction_disabled() {
380
+ let config = ExtractionConfig {
381
+ token_reduction: Some(TokenReductionConfig {
382
+ mode: "off".to_string(),
383
+ preserve_important_words: false,
384
+ }),
385
+ ..Default::default()
386
+ };
387
+
388
+ let text = "Text without token reduction applied.";
389
+ let text_bytes = text.as_bytes();
390
+
391
+ let result = extract_bytes(text_bytes, "text/plain", &config)
392
+ .await
393
+ .expect("Should extract successfully");
394
+
395
+ assert!(result.content.contains("without token reduction"));
396
+ }
397
+
398
+ /// Test quality processing enabled - quality scoring applied.
399
+ #[tokio::test]
400
+ async fn test_quality_processing_enabled() {
401
+ let config = ExtractionConfig {
402
+ enable_quality_processing: true,
403
+ ..Default::default()
404
+ };
405
+
406
+ let text = "This is well-structured text. It has multiple sentences. And proper punctuation.";
407
+ let text_bytes = text.as_bytes();
408
+
409
+ let result = extract_bytes(text_bytes, "text/plain", &config)
410
+ .await
411
+ .expect("Should extract successfully");
412
+
413
+ if let Some(score) = result.metadata.additional.get("quality_score") {
414
+ let score_value = score.as_f64().unwrap();
415
+ assert!((0.0..=1.0).contains(&score_value));
416
+ }
417
+
418
+ assert!(!result.content.is_empty());
419
+ }
420
+
421
+ /// Test quality processing calculates score for different text quality.
422
+ #[tokio::test]
423
+ async fn test_quality_threshold_filtering() {
424
+ let config = ExtractionConfig {
425
+ enable_quality_processing: true,
426
+ ..Default::default()
427
+ };
428
+
429
+ let high_quality = "This is a well-structured document. It has proper sentences. And good formatting.";
430
+ let result_high = extract_bytes(high_quality.as_bytes(), "text/plain", &config)
431
+ .await
432
+ .expect("Should extract successfully");
433
+
434
+ let low_quality = "a b c d ....... word123mixed . . ";
435
+ let result_low = extract_bytes(low_quality.as_bytes(), "text/plain", &config)
436
+ .await
437
+ .expect("Should extract successfully");
438
+
439
+ assert!(
440
+ result_high.metadata.additional.contains_key("quality_score"),
441
+ "High quality should have score"
442
+ );
443
+ assert!(
444
+ result_low.metadata.additional.contains_key("quality_score"),
445
+ "Low quality should have score"
446
+ );
447
+
448
+ let score_high = result_high
449
+ .metadata
450
+ .additional
451
+ .get("quality_score")
452
+ .unwrap()
453
+ .as_f64()
454
+ .unwrap();
455
+ let score_low = result_low
456
+ .metadata
457
+ .additional
458
+ .get("quality_score")
459
+ .unwrap()
460
+ .as_f64()
461
+ .unwrap();
462
+
463
+ assert!((0.0..=1.0).contains(&score_high));
464
+ assert!((0.0..=1.0).contains(&score_low));
465
+ }
466
+
467
+ /// Test quality processing disabled.
468
+ #[tokio::test]
469
+ async fn test_quality_processing_disabled() {
470
+ let config = ExtractionConfig {
471
+ enable_quality_processing: false,
472
+ ..Default::default()
473
+ };
474
+
475
+ let text = "Text without quality processing.";
476
+ let text_bytes = text.as_bytes();
477
+
478
+ let result = extract_bytes(text_bytes, "text/plain", &config)
479
+ .await
480
+ .expect("Should extract successfully");
481
+
482
+ assert!(!result.metadata.additional.contains_key("quality_score"));
483
+ assert!(!result.content.is_empty());
484
+ }
485
+
486
+ /// Test chunking with embeddings using balanced preset.
487
+ ///
488
+ /// This test requires ONNX Runtime to be installed as a system dependency.
489
+ /// On macOS with Homebrew: `brew install onnxruntime`
490
+ /// On Linux: Install via your package manager or download from https://github.com/microsoft/onnxruntime/releases
491
+ /// On Windows: Download from https://github.com/microsoft/onnxruntime/releases
492
+ #[tokio::test]
493
+ #[cfg(feature = "embeddings")]
494
+ #[cfg_attr(target_os = "macos", ignore = "ONNX models not cached on macOS")]
495
+ #[cfg_attr(target_os = "windows", ignore = "ONNX models not cached on Windows")]
496
+ async fn test_chunking_with_embeddings() {
497
+ use kreuzberg::core::config::EmbeddingConfig;
498
+
499
+ let config = ExtractionConfig {
500
+ chunking: Some(ChunkingConfig {
501
+ max_chars: 100,
502
+ max_overlap: 20,
503
+ embedding: Some(EmbeddingConfig::default()),
504
+ preset: None,
505
+ }),
506
+ ..Default::default()
507
+ };
508
+
509
+ let text = "This is a test document for embedding generation. ".repeat(10);
510
+ let text_bytes = text.as_bytes();
511
+
512
+ let result = extract_bytes(text_bytes, "text/plain", &config)
513
+ .await
514
+ .expect("Should extract successfully");
515
+
516
+ assert!(result.chunks.is_some(), "Chunks should be present");
517
+ let chunks = result.chunks.unwrap();
518
+ assert!(chunks.len() > 1, "Should have multiple chunks");
519
+
520
+ println!("Metadata: {:?}", result.metadata.additional);
521
+
522
+ if let Some(error) = result.metadata.additional.get("embedding_error") {
523
+ panic!("Embedding generation failed: {}", error);
524
+ }
525
+
526
+ assert!(
527
+ result.metadata.additional.contains_key("embeddings_generated"),
528
+ "Should have embeddings_generated metadata"
529
+ );
530
+ assert_eq!(
531
+ result.metadata.additional.get("embeddings_generated").unwrap(),
532
+ &serde_json::Value::Bool(true)
533
+ );
534
+
535
+ for chunk in &chunks {
536
+ assert!(chunk.embedding.is_some(), "Each chunk should have an embedding");
537
+ let embedding = chunk.embedding.as_ref().unwrap();
538
+ assert_eq!(
539
+ embedding.len(),
540
+ 768,
541
+ "Embedding should have 768 dimensions for balanced preset"
542
+ );
543
+
544
+ let magnitude: f32 = embedding.iter().map(|x| x * x).sum::<f32>().sqrt();
545
+ assert!(
546
+ (magnitude - 1.0).abs() < 0.01,
547
+ "Embedding should be normalized (magnitude ~= 1.0)"
548
+ );
549
+ }
550
+ }
551
+
552
+ /// Test chunking with fast embedding preset.
553
+ ///
554
+ /// This test requires ONNX Runtime to be installed as a system dependency.
555
+ /// On macOS with Homebrew: `brew install onnxruntime`
556
+ /// On Linux: Install via your package manager or download from https://github.com/microsoft/onnxruntime/releases
557
+ /// On Windows: Download from https://github.com/microsoft/onnxruntime/releases
558
+ #[tokio::test]
559
+ #[cfg(feature = "embeddings")]
560
+ #[cfg_attr(target_os = "macos", ignore = "ONNX models not cached on macOS")]
561
+ #[cfg_attr(target_os = "windows", ignore = "ONNX models not cached on Windows")]
562
+ async fn test_chunking_with_fast_embeddings() {
563
+ use kreuzberg::core::config::{EmbeddingConfig, EmbeddingModelType};
564
+
565
+ let config = ExtractionConfig {
566
+ chunking: Some(ChunkingConfig {
567
+ max_chars: 100,
568
+ max_overlap: 20,
569
+ embedding: Some(EmbeddingConfig {
570
+ model: EmbeddingModelType::Preset {
571
+ name: "fast".to_string(),
572
+ },
573
+ ..Default::default()
574
+ }),
575
+ preset: None,
576
+ }),
577
+ ..Default::default()
578
+ };
579
+
580
+ let text = "Fast embedding test. ".repeat(10);
581
+ let text_bytes = text.as_bytes();
582
+
583
+ let result = extract_bytes(text_bytes, "text/plain", &config)
584
+ .await
585
+ .expect("Should extract successfully");
586
+
587
+ let chunks = result.chunks.expect("Should have chunks");
588
+ assert!(!chunks.is_empty(), "Should have at least one chunk");
589
+
590
+ if let Some(error) = result.metadata.additional.get("embedding_error") {
591
+ panic!("Embedding generation failed: {}", error);
592
+ }
593
+
594
+ for chunk in &chunks {
595
+ let embedding = chunk.embedding.as_ref().expect("Should have embedding");
596
+ assert_eq!(embedding.len(), 384, "Fast preset should produce 384-dim embeddings");
597
+ }
598
+ }