kreuzberg 4.0.0.pre.rc.8 → 4.0.0.pre.rc.13

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (370) hide show
  1. checksums.yaml +4 -4
  2. data/.gitignore +14 -14
  3. data/.rspec +3 -3
  4. data/.rubocop.yaml +1 -1
  5. data/.rubocop.yml +538 -538
  6. data/Gemfile +8 -8
  7. data/Gemfile.lock +4 -104
  8. data/README.md +454 -432
  9. data/Rakefile +25 -25
  10. data/Steepfile +47 -47
  11. data/examples/async_patterns.rb +341 -341
  12. data/ext/kreuzberg_rb/extconf.rb +45 -45
  13. data/ext/kreuzberg_rb/native/.cargo/config.toml +2 -2
  14. data/ext/kreuzberg_rb/native/Cargo.lock +6941 -6721
  15. data/ext/kreuzberg_rb/native/Cargo.toml +54 -54
  16. data/ext/kreuzberg_rb/native/README.md +425 -425
  17. data/ext/kreuzberg_rb/native/build.rs +15 -15
  18. data/ext/kreuzberg_rb/native/include/ieeefp.h +11 -11
  19. data/ext/kreuzberg_rb/native/include/msvc_compat/strings.h +14 -14
  20. data/ext/kreuzberg_rb/native/include/strings.h +20 -20
  21. data/ext/kreuzberg_rb/native/include/unistd.h +47 -47
  22. data/ext/kreuzberg_rb/native/src/lib.rs +3158 -3135
  23. data/extconf.rb +28 -28
  24. data/kreuzberg.gemspec +214 -182
  25. data/lib/kreuzberg/api_proxy.rb +142 -142
  26. data/lib/kreuzberg/cache_api.rb +81 -46
  27. data/lib/kreuzberg/cli.rb +55 -55
  28. data/lib/kreuzberg/cli_proxy.rb +127 -127
  29. data/lib/kreuzberg/config.rb +724 -724
  30. data/lib/kreuzberg/error_context.rb +80 -32
  31. data/lib/kreuzberg/errors.rb +118 -118
  32. data/lib/kreuzberg/extraction_api.rb +340 -85
  33. data/lib/kreuzberg/mcp_proxy.rb +186 -186
  34. data/lib/kreuzberg/ocr_backend_protocol.rb +113 -113
  35. data/lib/kreuzberg/post_processor_protocol.rb +86 -86
  36. data/lib/kreuzberg/result.rb +279 -279
  37. data/lib/kreuzberg/setup_lib_path.rb +80 -80
  38. data/lib/kreuzberg/validator_protocol.rb +89 -89
  39. data/lib/kreuzberg/version.rb +5 -5
  40. data/lib/kreuzberg.rb +109 -103
  41. data/lib/pdfium.dll +0 -0
  42. data/sig/kreuzberg/internal.rbs +184 -184
  43. data/sig/kreuzberg.rbs +546 -537
  44. data/spec/binding/cache_spec.rb +227 -227
  45. data/spec/binding/cli_proxy_spec.rb +85 -85
  46. data/spec/binding/cli_spec.rb +55 -55
  47. data/spec/binding/config_spec.rb +345 -345
  48. data/spec/binding/config_validation_spec.rb +283 -283
  49. data/spec/binding/error_handling_spec.rb +213 -213
  50. data/spec/binding/errors_spec.rb +66 -66
  51. data/spec/binding/plugins/ocr_backend_spec.rb +307 -307
  52. data/spec/binding/plugins/postprocessor_spec.rb +269 -269
  53. data/spec/binding/plugins/validator_spec.rb +274 -274
  54. data/spec/fixtures/config.toml +39 -39
  55. data/spec/fixtures/config.yaml +41 -41
  56. data/spec/fixtures/invalid_config.toml +4 -4
  57. data/spec/smoke/package_spec.rb +178 -178
  58. data/spec/spec_helper.rb +42 -42
  59. data/vendor/Cargo.toml +45 -0
  60. data/vendor/kreuzberg/Cargo.toml +61 -38
  61. data/vendor/kreuzberg/README.md +230 -221
  62. data/vendor/kreuzberg/benches/otel_overhead.rs +48 -48
  63. data/vendor/kreuzberg/build.rs +843 -891
  64. data/vendor/kreuzberg/src/api/error.rs +81 -81
  65. data/vendor/kreuzberg/src/api/handlers.rs +199 -199
  66. data/vendor/kreuzberg/src/api/mod.rs +79 -79
  67. data/vendor/kreuzberg/src/api/server.rs +353 -353
  68. data/vendor/kreuzberg/src/api/types.rs +170 -170
  69. data/vendor/kreuzberg/src/cache/mod.rs +1167 -1167
  70. data/vendor/kreuzberg/src/chunking/mod.rs +1877 -1877
  71. data/vendor/kreuzberg/src/chunking/processor.rs +220 -220
  72. data/vendor/kreuzberg/src/core/batch_mode.rs +95 -95
  73. data/vendor/kreuzberg/src/core/config.rs +1080 -1080
  74. data/vendor/kreuzberg/src/core/extractor.rs +1156 -1156
  75. data/vendor/kreuzberg/src/core/io.rs +329 -329
  76. data/vendor/kreuzberg/src/core/mime.rs +605 -605
  77. data/vendor/kreuzberg/src/core/mod.rs +47 -47
  78. data/vendor/kreuzberg/src/core/pipeline.rs +1184 -1171
  79. data/vendor/kreuzberg/src/embeddings.rs +500 -432
  80. data/vendor/kreuzberg/src/error.rs +431 -431
  81. data/vendor/kreuzberg/src/extraction/archive.rs +954 -954
  82. data/vendor/kreuzberg/src/extraction/docx.rs +398 -398
  83. data/vendor/kreuzberg/src/extraction/email.rs +854 -854
  84. data/vendor/kreuzberg/src/extraction/excel.rs +688 -688
  85. data/vendor/kreuzberg/src/extraction/html.rs +601 -569
  86. data/vendor/kreuzberg/src/extraction/image.rs +491 -491
  87. data/vendor/kreuzberg/src/extraction/libreoffice.rs +574 -562
  88. data/vendor/kreuzberg/src/extraction/markdown.rs +213 -213
  89. data/vendor/kreuzberg/src/extraction/mod.rs +81 -81
  90. data/vendor/kreuzberg/src/extraction/office_metadata/app_properties.rs +398 -398
  91. data/vendor/kreuzberg/src/extraction/office_metadata/core_properties.rs +247 -247
  92. data/vendor/kreuzberg/src/extraction/office_metadata/custom_properties.rs +240 -240
  93. data/vendor/kreuzberg/src/extraction/office_metadata/mod.rs +130 -130
  94. data/vendor/kreuzberg/src/extraction/office_metadata/odt_properties.rs +284 -284
  95. data/vendor/kreuzberg/src/extraction/pptx.rs +3100 -3100
  96. data/vendor/kreuzberg/src/extraction/structured.rs +490 -490
  97. data/vendor/kreuzberg/src/extraction/table.rs +328 -328
  98. data/vendor/kreuzberg/src/extraction/text.rs +269 -269
  99. data/vendor/kreuzberg/src/extraction/xml.rs +333 -333
  100. data/vendor/kreuzberg/src/extractors/archive.rs +447 -447
  101. data/vendor/kreuzberg/src/extractors/bibtex.rs +470 -470
  102. data/vendor/kreuzberg/src/extractors/docbook.rs +504 -504
  103. data/vendor/kreuzberg/src/extractors/docx.rs +400 -400
  104. data/vendor/kreuzberg/src/extractors/email.rs +157 -157
  105. data/vendor/kreuzberg/src/extractors/epub.rs +708 -708
  106. data/vendor/kreuzberg/src/extractors/excel.rs +345 -345
  107. data/vendor/kreuzberg/src/extractors/fictionbook.rs +492 -492
  108. data/vendor/kreuzberg/src/extractors/html.rs +407 -407
  109. data/vendor/kreuzberg/src/extractors/image.rs +219 -219
  110. data/vendor/kreuzberg/src/extractors/jats.rs +1054 -1054
  111. data/vendor/kreuzberg/src/extractors/jupyter.rs +368 -368
  112. data/vendor/kreuzberg/src/extractors/latex.rs +653 -653
  113. data/vendor/kreuzberg/src/extractors/markdown.rs +701 -701
  114. data/vendor/kreuzberg/src/extractors/mod.rs +429 -429
  115. data/vendor/kreuzberg/src/extractors/odt.rs +628 -628
  116. data/vendor/kreuzberg/src/extractors/opml.rs +635 -635
  117. data/vendor/kreuzberg/src/extractors/orgmode.rs +529 -529
  118. data/vendor/kreuzberg/src/extractors/pdf.rs +749 -673
  119. data/vendor/kreuzberg/src/extractors/pptx.rs +267 -267
  120. data/vendor/kreuzberg/src/extractors/rst.rs +577 -577
  121. data/vendor/kreuzberg/src/extractors/rtf.rs +809 -809
  122. data/vendor/kreuzberg/src/extractors/security.rs +484 -484
  123. data/vendor/kreuzberg/src/extractors/security_tests.rs +367 -367
  124. data/vendor/kreuzberg/src/extractors/structured.rs +142 -142
  125. data/vendor/kreuzberg/src/extractors/text.rs +265 -265
  126. data/vendor/kreuzberg/src/extractors/typst.rs +651 -651
  127. data/vendor/kreuzberg/src/extractors/xml.rs +147 -147
  128. data/vendor/kreuzberg/src/image/dpi.rs +164 -164
  129. data/vendor/kreuzberg/src/image/mod.rs +6 -6
  130. data/vendor/kreuzberg/src/image/preprocessing.rs +417 -417
  131. data/vendor/kreuzberg/src/image/resize.rs +89 -89
  132. data/vendor/kreuzberg/src/keywords/config.rs +154 -154
  133. data/vendor/kreuzberg/src/keywords/mod.rs +237 -237
  134. data/vendor/kreuzberg/src/keywords/processor.rs +275 -275
  135. data/vendor/kreuzberg/src/keywords/rake.rs +293 -293
  136. data/vendor/kreuzberg/src/keywords/types.rs +68 -68
  137. data/vendor/kreuzberg/src/keywords/yake.rs +163 -163
  138. data/vendor/kreuzberg/src/language_detection/mod.rs +985 -985
  139. data/vendor/kreuzberg/src/language_detection/processor.rs +219 -219
  140. data/vendor/kreuzberg/src/lib.rs +113 -113
  141. data/vendor/kreuzberg/src/mcp/mod.rs +35 -35
  142. data/vendor/kreuzberg/src/mcp/server.rs +2076 -2076
  143. data/vendor/kreuzberg/src/ocr/cache.rs +469 -469
  144. data/vendor/kreuzberg/src/ocr/error.rs +37 -37
  145. data/vendor/kreuzberg/src/ocr/hocr.rs +216 -216
  146. data/vendor/kreuzberg/src/ocr/mod.rs +58 -58
  147. data/vendor/kreuzberg/src/ocr/processor.rs +863 -863
  148. data/vendor/kreuzberg/src/ocr/table/mod.rs +4 -4
  149. data/vendor/kreuzberg/src/ocr/table/tsv_parser.rs +144 -144
  150. data/vendor/kreuzberg/src/ocr/tesseract_backend.rs +452 -452
  151. data/vendor/kreuzberg/src/ocr/types.rs +393 -393
  152. data/vendor/kreuzberg/src/ocr/utils.rs +47 -47
  153. data/vendor/kreuzberg/src/ocr/validation.rs +206 -206
  154. data/vendor/kreuzberg/src/panic_context.rs +154 -154
  155. data/vendor/kreuzberg/src/pdf/bindings.rs +44 -0
  156. data/vendor/kreuzberg/src/pdf/bundled.rs +346 -328
  157. data/vendor/kreuzberg/src/pdf/error.rs +130 -130
  158. data/vendor/kreuzberg/src/pdf/images.rs +139 -139
  159. data/vendor/kreuzberg/src/pdf/metadata.rs +489 -489
  160. data/vendor/kreuzberg/src/pdf/mod.rs +68 -66
  161. data/vendor/kreuzberg/src/pdf/rendering.rs +368 -368
  162. data/vendor/kreuzberg/src/pdf/table.rs +420 -417
  163. data/vendor/kreuzberg/src/pdf/text.rs +240 -240
  164. data/vendor/kreuzberg/src/plugins/extractor.rs +1044 -1044
  165. data/vendor/kreuzberg/src/plugins/mod.rs +212 -212
  166. data/vendor/kreuzberg/src/plugins/ocr.rs +639 -639
  167. data/vendor/kreuzberg/src/plugins/processor.rs +650 -650
  168. data/vendor/kreuzberg/src/plugins/registry.rs +1339 -1339
  169. data/vendor/kreuzberg/src/plugins/traits.rs +258 -258
  170. data/vendor/kreuzberg/src/plugins/validator.rs +967 -967
  171. data/vendor/kreuzberg/src/stopwords/mod.rs +1470 -1470
  172. data/vendor/kreuzberg/src/text/mod.rs +25 -25
  173. data/vendor/kreuzberg/src/text/quality.rs +697 -697
  174. data/vendor/kreuzberg/src/text/quality_processor.rs +219 -219
  175. data/vendor/kreuzberg/src/text/string_utils.rs +217 -217
  176. data/vendor/kreuzberg/src/text/token_reduction/cjk_utils.rs +164 -164
  177. data/vendor/kreuzberg/src/text/token_reduction/config.rs +100 -100
  178. data/vendor/kreuzberg/src/text/token_reduction/core.rs +796 -796
  179. data/vendor/kreuzberg/src/text/token_reduction/filters.rs +902 -902
  180. data/vendor/kreuzberg/src/text/token_reduction/mod.rs +160 -160
  181. data/vendor/kreuzberg/src/text/token_reduction/semantic.rs +619 -619
  182. data/vendor/kreuzberg/src/text/token_reduction/simd_text.rs +147 -147
  183. data/vendor/kreuzberg/src/types.rs +1055 -1055
  184. data/vendor/kreuzberg/src/utils/mod.rs +17 -17
  185. data/vendor/kreuzberg/src/utils/quality.rs +959 -959
  186. data/vendor/kreuzberg/src/utils/string_utils.rs +381 -381
  187. data/vendor/kreuzberg/stopwords/af_stopwords.json +53 -53
  188. data/vendor/kreuzberg/stopwords/ar_stopwords.json +482 -482
  189. data/vendor/kreuzberg/stopwords/bg_stopwords.json +261 -261
  190. data/vendor/kreuzberg/stopwords/bn_stopwords.json +400 -400
  191. data/vendor/kreuzberg/stopwords/br_stopwords.json +1205 -1205
  192. data/vendor/kreuzberg/stopwords/ca_stopwords.json +280 -280
  193. data/vendor/kreuzberg/stopwords/cs_stopwords.json +425 -425
  194. data/vendor/kreuzberg/stopwords/da_stopwords.json +172 -172
  195. data/vendor/kreuzberg/stopwords/de_stopwords.json +622 -622
  196. data/vendor/kreuzberg/stopwords/el_stopwords.json +849 -849
  197. data/vendor/kreuzberg/stopwords/en_stopwords.json +1300 -1300
  198. data/vendor/kreuzberg/stopwords/eo_stopwords.json +175 -175
  199. data/vendor/kreuzberg/stopwords/es_stopwords.json +734 -734
  200. data/vendor/kreuzberg/stopwords/et_stopwords.json +37 -37
  201. data/vendor/kreuzberg/stopwords/eu_stopwords.json +100 -100
  202. data/vendor/kreuzberg/stopwords/fa_stopwords.json +801 -801
  203. data/vendor/kreuzberg/stopwords/fi_stopwords.json +849 -849
  204. data/vendor/kreuzberg/stopwords/fr_stopwords.json +693 -693
  205. data/vendor/kreuzberg/stopwords/ga_stopwords.json +111 -111
  206. data/vendor/kreuzberg/stopwords/gl_stopwords.json +162 -162
  207. data/vendor/kreuzberg/stopwords/gu_stopwords.json +226 -226
  208. data/vendor/kreuzberg/stopwords/ha_stopwords.json +41 -41
  209. data/vendor/kreuzberg/stopwords/he_stopwords.json +196 -196
  210. data/vendor/kreuzberg/stopwords/hi_stopwords.json +227 -227
  211. data/vendor/kreuzberg/stopwords/hr_stopwords.json +181 -181
  212. data/vendor/kreuzberg/stopwords/hu_stopwords.json +791 -791
  213. data/vendor/kreuzberg/stopwords/hy_stopwords.json +47 -47
  214. data/vendor/kreuzberg/stopwords/id_stopwords.json +760 -760
  215. data/vendor/kreuzberg/stopwords/it_stopwords.json +634 -634
  216. data/vendor/kreuzberg/stopwords/ja_stopwords.json +136 -136
  217. data/vendor/kreuzberg/stopwords/kn_stopwords.json +84 -84
  218. data/vendor/kreuzberg/stopwords/ko_stopwords.json +681 -681
  219. data/vendor/kreuzberg/stopwords/ku_stopwords.json +64 -64
  220. data/vendor/kreuzberg/stopwords/la_stopwords.json +51 -51
  221. data/vendor/kreuzberg/stopwords/lt_stopwords.json +476 -476
  222. data/vendor/kreuzberg/stopwords/lv_stopwords.json +163 -163
  223. data/vendor/kreuzberg/stopwords/ml_stopwords.json +1 -1
  224. data/vendor/kreuzberg/stopwords/mr_stopwords.json +101 -101
  225. data/vendor/kreuzberg/stopwords/ms_stopwords.json +477 -477
  226. data/vendor/kreuzberg/stopwords/ne_stopwords.json +490 -490
  227. data/vendor/kreuzberg/stopwords/nl_stopwords.json +415 -415
  228. data/vendor/kreuzberg/stopwords/no_stopwords.json +223 -223
  229. data/vendor/kreuzberg/stopwords/pl_stopwords.json +331 -331
  230. data/vendor/kreuzberg/stopwords/pt_stopwords.json +562 -562
  231. data/vendor/kreuzberg/stopwords/ro_stopwords.json +436 -436
  232. data/vendor/kreuzberg/stopwords/ru_stopwords.json +561 -561
  233. data/vendor/kreuzberg/stopwords/si_stopwords.json +193 -193
  234. data/vendor/kreuzberg/stopwords/sk_stopwords.json +420 -420
  235. data/vendor/kreuzberg/stopwords/sl_stopwords.json +448 -448
  236. data/vendor/kreuzberg/stopwords/so_stopwords.json +32 -32
  237. data/vendor/kreuzberg/stopwords/st_stopwords.json +33 -33
  238. data/vendor/kreuzberg/stopwords/sv_stopwords.json +420 -420
  239. data/vendor/kreuzberg/stopwords/sw_stopwords.json +76 -76
  240. data/vendor/kreuzberg/stopwords/ta_stopwords.json +129 -129
  241. data/vendor/kreuzberg/stopwords/te_stopwords.json +54 -54
  242. data/vendor/kreuzberg/stopwords/th_stopwords.json +118 -118
  243. data/vendor/kreuzberg/stopwords/tl_stopwords.json +149 -149
  244. data/vendor/kreuzberg/stopwords/tr_stopwords.json +506 -506
  245. data/vendor/kreuzberg/stopwords/uk_stopwords.json +75 -75
  246. data/vendor/kreuzberg/stopwords/ur_stopwords.json +519 -519
  247. data/vendor/kreuzberg/stopwords/vi_stopwords.json +647 -647
  248. data/vendor/kreuzberg/stopwords/yo_stopwords.json +62 -62
  249. data/vendor/kreuzberg/stopwords/zh_stopwords.json +796 -796
  250. data/vendor/kreuzberg/stopwords/zu_stopwords.json +31 -31
  251. data/vendor/kreuzberg/tests/api_extract_multipart.rs +52 -52
  252. data/vendor/kreuzberg/tests/api_tests.rs +966 -966
  253. data/vendor/kreuzberg/tests/archive_integration.rs +545 -545
  254. data/vendor/kreuzberg/tests/batch_orchestration.rs +556 -556
  255. data/vendor/kreuzberg/tests/batch_processing.rs +318 -318
  256. data/vendor/kreuzberg/tests/bibtex_parity_test.rs +421 -421
  257. data/vendor/kreuzberg/tests/concurrency_stress.rs +533 -533
  258. data/vendor/kreuzberg/tests/config_features.rs +612 -612
  259. data/vendor/kreuzberg/tests/config_loading_tests.rs +416 -416
  260. data/vendor/kreuzberg/tests/core_integration.rs +510 -510
  261. data/vendor/kreuzberg/tests/csv_integration.rs +414 -414
  262. data/vendor/kreuzberg/tests/docbook_extractor_tests.rs +500 -500
  263. data/vendor/kreuzberg/tests/docx_metadata_extraction_test.rs +122 -122
  264. data/vendor/kreuzberg/tests/docx_vs_pandoc_comparison.rs +370 -370
  265. data/vendor/kreuzberg/tests/email_integration.rs +327 -327
  266. data/vendor/kreuzberg/tests/epub_native_extractor_tests.rs +275 -275
  267. data/vendor/kreuzberg/tests/error_handling.rs +402 -402
  268. data/vendor/kreuzberg/tests/fictionbook_extractor_tests.rs +228 -228
  269. data/vendor/kreuzberg/tests/format_integration.rs +164 -161
  270. data/vendor/kreuzberg/tests/helpers/mod.rs +142 -142
  271. data/vendor/kreuzberg/tests/html_table_test.rs +551 -551
  272. data/vendor/kreuzberg/tests/image_integration.rs +255 -255
  273. data/vendor/kreuzberg/tests/instrumentation_test.rs +139 -139
  274. data/vendor/kreuzberg/tests/jats_extractor_tests.rs +639 -639
  275. data/vendor/kreuzberg/tests/jupyter_extractor_tests.rs +704 -704
  276. data/vendor/kreuzberg/tests/keywords_integration.rs +479 -479
  277. data/vendor/kreuzberg/tests/keywords_quality.rs +509 -509
  278. data/vendor/kreuzberg/tests/latex_extractor_tests.rs +496 -496
  279. data/vendor/kreuzberg/tests/markdown_extractor_tests.rs +490 -490
  280. data/vendor/kreuzberg/tests/mime_detection.rs +429 -429
  281. data/vendor/kreuzberg/tests/ocr_configuration.rs +514 -514
  282. data/vendor/kreuzberg/tests/ocr_errors.rs +698 -698
  283. data/vendor/kreuzberg/tests/ocr_quality.rs +629 -629
  284. data/vendor/kreuzberg/tests/ocr_stress.rs +469 -469
  285. data/vendor/kreuzberg/tests/odt_extractor_tests.rs +674 -674
  286. data/vendor/kreuzberg/tests/opml_extractor_tests.rs +616 -616
  287. data/vendor/kreuzberg/tests/orgmode_extractor_tests.rs +822 -822
  288. data/vendor/kreuzberg/tests/pdf_integration.rs +45 -45
  289. data/vendor/kreuzberg/tests/pdfium_linking.rs +374 -374
  290. data/vendor/kreuzberg/tests/pipeline_integration.rs +1436 -1436
  291. data/vendor/kreuzberg/tests/plugin_ocr_backend_test.rs +776 -776
  292. data/vendor/kreuzberg/tests/plugin_postprocessor_test.rs +560 -560
  293. data/vendor/kreuzberg/tests/plugin_system.rs +927 -927
  294. data/vendor/kreuzberg/tests/plugin_validator_test.rs +783 -783
  295. data/vendor/kreuzberg/tests/registry_integration_tests.rs +587 -587
  296. data/vendor/kreuzberg/tests/rst_extractor_tests.rs +694 -694
  297. data/vendor/kreuzberg/tests/rtf_extractor_tests.rs +775 -775
  298. data/vendor/kreuzberg/tests/security_validation.rs +416 -416
  299. data/vendor/kreuzberg/tests/stopwords_integration_test.rs +888 -888
  300. data/vendor/kreuzberg/tests/test_fastembed.rs +631 -631
  301. data/vendor/kreuzberg/tests/typst_behavioral_tests.rs +1260 -1260
  302. data/vendor/kreuzberg/tests/typst_extractor_tests.rs +648 -648
  303. data/vendor/kreuzberg/tests/xlsx_metadata_extraction_test.rs +87 -87
  304. data/vendor/kreuzberg-ffi/Cargo.toml +63 -0
  305. data/vendor/kreuzberg-ffi/README.md +851 -0
  306. data/vendor/kreuzberg-ffi/build.rs +176 -0
  307. data/vendor/kreuzberg-ffi/cbindgen.toml +27 -0
  308. data/vendor/kreuzberg-ffi/kreuzberg-ffi-install.pc +12 -0
  309. data/vendor/kreuzberg-ffi/kreuzberg-ffi.pc.in +12 -0
  310. data/vendor/kreuzberg-ffi/kreuzberg.h +1087 -0
  311. data/vendor/kreuzberg-ffi/src/lib.rs +3616 -0
  312. data/vendor/kreuzberg-ffi/src/panic_shield.rs +247 -0
  313. data/vendor/kreuzberg-ffi/tests.disabled/README.md +48 -0
  314. data/vendor/kreuzberg-ffi/tests.disabled/config_loading_tests.rs +299 -0
  315. data/vendor/kreuzberg-ffi/tests.disabled/config_tests.rs +346 -0
  316. data/vendor/kreuzberg-ffi/tests.disabled/extractor_tests.rs +232 -0
  317. data/vendor/kreuzberg-ffi/tests.disabled/plugin_registration_tests.rs +470 -0
  318. data/vendor/kreuzberg-tesseract/.commitlintrc.json +13 -0
  319. data/vendor/kreuzberg-tesseract/.crate-ignore +2 -0
  320. data/vendor/kreuzberg-tesseract/Cargo.lock +2933 -0
  321. data/vendor/kreuzberg-tesseract/Cargo.toml +48 -0
  322. data/vendor/kreuzberg-tesseract/LICENSE +22 -0
  323. data/vendor/kreuzberg-tesseract/README.md +399 -0
  324. data/vendor/kreuzberg-tesseract/build.rs +1354 -0
  325. data/vendor/kreuzberg-tesseract/patches/README.md +71 -0
  326. data/vendor/kreuzberg-tesseract/patches/tesseract.diff +199 -0
  327. data/vendor/kreuzberg-tesseract/src/api.rs +1371 -0
  328. data/vendor/kreuzberg-tesseract/src/choice_iterator.rs +77 -0
  329. data/vendor/kreuzberg-tesseract/src/enums.rs +297 -0
  330. data/vendor/kreuzberg-tesseract/src/error.rs +81 -0
  331. data/vendor/kreuzberg-tesseract/src/lib.rs +145 -0
  332. data/vendor/kreuzberg-tesseract/src/monitor.rs +57 -0
  333. data/vendor/kreuzberg-tesseract/src/mutable_iterator.rs +197 -0
  334. data/vendor/kreuzberg-tesseract/src/page_iterator.rs +253 -0
  335. data/vendor/kreuzberg-tesseract/src/result_iterator.rs +286 -0
  336. data/vendor/kreuzberg-tesseract/src/result_renderer.rs +183 -0
  337. data/vendor/kreuzberg-tesseract/tests/integration_test.rs +211 -0
  338. data/vendor/rb-sys/.cargo_vcs_info.json +5 -5
  339. data/vendor/rb-sys/Cargo.lock +393 -393
  340. data/vendor/rb-sys/Cargo.toml +70 -70
  341. data/vendor/rb-sys/Cargo.toml.orig +57 -57
  342. data/vendor/rb-sys/LICENSE-APACHE +190 -190
  343. data/vendor/rb-sys/LICENSE-MIT +21 -21
  344. data/vendor/rb-sys/build/features.rs +111 -111
  345. data/vendor/rb-sys/build/main.rs +286 -286
  346. data/vendor/rb-sys/build/stable_api_config.rs +155 -155
  347. data/vendor/rb-sys/build/version.rs +50 -50
  348. data/vendor/rb-sys/readme.md +36 -36
  349. data/vendor/rb-sys/src/bindings.rs +21 -21
  350. data/vendor/rb-sys/src/hidden.rs +11 -11
  351. data/vendor/rb-sys/src/lib.rs +35 -35
  352. data/vendor/rb-sys/src/macros.rs +371 -371
  353. data/vendor/rb-sys/src/memory.rs +53 -53
  354. data/vendor/rb-sys/src/ruby_abi_version.rs +38 -38
  355. data/vendor/rb-sys/src/special_consts.rs +31 -31
  356. data/vendor/rb-sys/src/stable_api/compiled.c +179 -179
  357. data/vendor/rb-sys/src/stable_api/compiled.rs +257 -257
  358. data/vendor/rb-sys/src/stable_api/ruby_2_7.rs +324 -324
  359. data/vendor/rb-sys/src/stable_api/ruby_3_0.rs +332 -332
  360. data/vendor/rb-sys/src/stable_api/ruby_3_1.rs +325 -325
  361. data/vendor/rb-sys/src/stable_api/ruby_3_2.rs +323 -323
  362. data/vendor/rb-sys/src/stable_api/ruby_3_3.rs +339 -339
  363. data/vendor/rb-sys/src/stable_api/ruby_3_4.rs +339 -339
  364. data/vendor/rb-sys/src/stable_api.rs +260 -260
  365. data/vendor/rb-sys/src/symbol.rs +31 -31
  366. data/vendor/rb-sys/src/tracking_allocator.rs +330 -330
  367. data/vendor/rb-sys/src/utils.rs +89 -89
  368. data/vendor/rb-sys/src/value_type.rs +7 -7
  369. metadata +44 -81
  370. data/vendor/rb-sys/bin/release.sh +0 -21
@@ -1,612 +1,612 @@
1
- //! Configuration features integration tests.
2
- //!
3
- //! Tests for chunking, language detection, caching, token reduction, and quality processing.
4
- //! Validates that configuration options work correctly end-to-end.
5
-
6
- #[cfg(feature = "chunking")]
7
- use kreuzberg::core::config::ChunkingConfig;
8
- use kreuzberg::core::config::ExtractionConfig;
9
- #[cfg(feature = "language-detection")]
10
- use kreuzberg::core::config::LanguageDetectionConfig;
11
- use kreuzberg::core::config::TokenReductionConfig;
12
- use kreuzberg::core::extractor::extract_bytes;
13
-
14
- mod helpers;
15
-
16
- /// Test chunking enabled - text split into chunks.
17
- #[tokio::test]
18
- #[cfg(feature = "chunking")]
19
- async fn test_chunking_enabled() {
20
- let config = ExtractionConfig {
21
- chunking: Some(ChunkingConfig {
22
- max_chars: 50,
23
- max_overlap: 10,
24
- embedding: None,
25
- preset: None,
26
- }),
27
- ..Default::default()
28
- };
29
-
30
- let text = "This is a long text that should be split into multiple chunks. ".repeat(10);
31
- let text_bytes = text.as_bytes();
32
-
33
- let result = extract_bytes(text_bytes, "text/plain", &config)
34
- .await
35
- .expect("Should extract successfully");
36
-
37
- assert!(result.chunks.is_some(), "Chunks should be present");
38
- let chunks = result.chunks.unwrap();
39
- assert!(chunks.len() > 1, "Should have multiple chunks");
40
-
41
- assert!(result.metadata.additional.contains_key("chunk_count"));
42
- let chunk_count = result.metadata.additional.get("chunk_count").unwrap();
43
- assert_eq!(
44
- chunks.len(),
45
- chunk_count.as_u64().unwrap() as usize,
46
- "Chunks length should match chunk_count metadata"
47
- );
48
-
49
- for chunk in &chunks {
50
- assert!(!chunk.content.is_empty(), "Chunk should not be empty");
51
- assert!(
52
- chunk.content.len() <= 50 + 10,
53
- "Chunk length {} exceeds max_chars + overlap",
54
- chunk.content.len()
55
- );
56
- }
57
- }
58
-
59
- /// Test chunking with overlap - overlap preserved between chunks.
60
- #[tokio::test]
61
- #[cfg(feature = "chunking")]
62
- async fn test_chunking_with_overlap() {
63
- let config = ExtractionConfig {
64
- chunking: Some(ChunkingConfig {
65
- max_chars: 100,
66
- max_overlap: 20,
67
- embedding: None,
68
- preset: None,
69
- }),
70
- ..Default::default()
71
- };
72
-
73
- let text = "a".repeat(250);
74
- let text_bytes = text.as_bytes();
75
-
76
- let result = extract_bytes(text_bytes, "text/plain", &config)
77
- .await
78
- .expect("Should extract successfully");
79
-
80
- assert!(result.chunks.is_some(), "Chunks should be present");
81
- let chunks = result.chunks.unwrap();
82
- assert!(chunks.len() >= 2, "Should have at least 2 chunks");
83
-
84
- assert!(result.metadata.additional.contains_key("chunk_count"));
85
-
86
- if chunks.len() >= 2 {
87
- let chunk1 = &chunks[0];
88
- let chunk2 = &chunks[1];
89
-
90
- let chunk1_end = &chunk1.content[chunk1.content.len().saturating_sub(20)..];
91
- assert!(
92
- chunk2.content.starts_with(chunk1_end)
93
- || chunk1_end.starts_with(&chunk2.content[..chunk1_end.len().min(chunk2.content.len())]),
94
- "Chunks should have overlap"
95
- );
96
- }
97
- }
98
-
99
- /// Test chunking with custom sizes - custom chunk size and overlap.
100
- #[tokio::test]
101
- #[cfg(feature = "chunking")]
102
- async fn test_chunking_custom_sizes() {
103
- let config = ExtractionConfig {
104
- chunking: Some(ChunkingConfig {
105
- max_chars: 200,
106
- max_overlap: 50,
107
- embedding: None,
108
- preset: None,
109
- }),
110
- ..Default::default()
111
- };
112
-
113
- let text = "Custom chunk test. ".repeat(50);
114
- let text_bytes = text.as_bytes();
115
-
116
- let result = extract_bytes(text_bytes, "text/plain", &config)
117
- .await
118
- .expect("Should extract successfully");
119
-
120
- assert!(result.chunks.is_some(), "Chunks should be present");
121
- let chunks = result.chunks.unwrap();
122
- assert!(!chunks.is_empty(), "Should have at least 1 chunk");
123
-
124
- assert!(result.metadata.additional.contains_key("chunk_count"));
125
-
126
- for chunk in &chunks {
127
- assert!(
128
- chunk.content.len() <= 200 + 50,
129
- "Chunk length {} exceeds custom max_chars + overlap",
130
- chunk.content.len()
131
- );
132
- }
133
- }
134
-
135
- /// Test chunking disabled - no chunking when disabled.
136
- #[tokio::test]
137
- async fn test_chunking_disabled() {
138
- let config = ExtractionConfig {
139
- chunking: None,
140
- ..Default::default()
141
- };
142
-
143
- let text = "This is a long text that should NOT be split into chunks. ".repeat(10);
144
- let text_bytes = text.as_bytes();
145
-
146
- let result = extract_bytes(text_bytes, "text/plain", &config)
147
- .await
148
- .expect("Should extract successfully");
149
-
150
- assert!(result.chunks.is_none(), "Should not have chunks when chunking disabled");
151
- assert!(
152
- !result.metadata.additional.contains_key("chunk_count"),
153
- "Should not have chunk_count when chunking disabled"
154
- );
155
-
156
- assert!(!result.content.is_empty(), "Content should be extracted");
157
- assert!(result.content.contains("long text"), "Should contain original text");
158
- }
159
-
160
- /// Test language detection for single language document.
161
- #[tokio::test]
162
- #[cfg(feature = "language-detection")]
163
- async fn test_language_detection_single() {
164
- let config = ExtractionConfig {
165
- language_detection: Some(LanguageDetectionConfig {
166
- enabled: true,
167
- min_confidence: 0.8,
168
- detect_multiple: false,
169
- }),
170
- ..Default::default()
171
- };
172
-
173
- let text = "Hello world! This is English text. It should be detected as English language.";
174
- let text_bytes = text.as_bytes();
175
-
176
- let result = extract_bytes(text_bytes, "text/plain", &config)
177
- .await
178
- .expect("Should extract successfully");
179
-
180
- assert!(result.detected_languages.is_some(), "Should detect language");
181
- let languages = result.detected_languages.unwrap();
182
- assert!(!languages.is_empty(), "Should detect at least one language");
183
- assert_eq!(languages[0], "eng", "Should detect English");
184
- }
185
-
186
- /// Test language detection for multi-language document.
187
- #[cfg_attr(coverage, ignore = "coverage instrumentation affects multi-language heuristics")]
188
- #[tokio::test]
189
- #[cfg(feature = "language-detection")]
190
- async fn test_language_detection_multiple() {
191
- let config = ExtractionConfig {
192
- language_detection: Some(LanguageDetectionConfig {
193
- enabled: true,
194
- min_confidence: 0.7,
195
- detect_multiple: true,
196
- }),
197
- ..Default::default()
198
- };
199
-
200
- let text = "Hello world! This is English. ".repeat(10) + "Hola mundo! Este es español. ".repeat(10).as_str();
201
- let text_bytes = text.as_bytes();
202
-
203
- let result = extract_bytes(text_bytes, "text/plain", &config)
204
- .await
205
- .expect("Should extract successfully");
206
-
207
- assert!(result.detected_languages.is_some(), "Should detect languages");
208
- let languages = result.detected_languages.unwrap();
209
- assert!(!languages.is_empty(), "Should detect at least one language");
210
- }
211
-
212
- /// Test language detection with confidence threshold.
213
- #[tokio::test]
214
- #[cfg(feature = "language-detection")]
215
- async fn test_language_detection_confidence() {
216
- let config = ExtractionConfig {
217
- language_detection: Some(LanguageDetectionConfig {
218
- enabled: true,
219
- min_confidence: 0.9,
220
- detect_multiple: false,
221
- }),
222
- ..Default::default()
223
- };
224
-
225
- let text = "This is clear English text that should have high confidence.";
226
- let text_bytes = text.as_bytes();
227
-
228
- let result = extract_bytes(text_bytes, "text/plain", &config)
229
- .await
230
- .expect("Should extract successfully");
231
-
232
- if let Some(languages) = result.detected_languages {
233
- assert!(!languages.is_empty());
234
- }
235
- }
236
-
237
- /// Test language detection disabled.
238
- #[tokio::test]
239
- #[cfg(feature = "language-detection")]
240
- async fn test_language_detection_disabled() {
241
- let config = ExtractionConfig {
242
- language_detection: Some(LanguageDetectionConfig {
243
- enabled: false,
244
- min_confidence: 0.8,
245
- detect_multiple: false,
246
- }),
247
- ..Default::default()
248
- };
249
-
250
- let text = "Hello world! This is English text.";
251
- let text_bytes = text.as_bytes();
252
-
253
- let result = extract_bytes(text_bytes, "text/plain", &config)
254
- .await
255
- .expect("Should extract successfully");
256
-
257
- assert!(
258
- result.detected_languages.is_none(),
259
- "Should not detect language when disabled"
260
- );
261
- }
262
-
263
- /// Test cache hit behavior - second extraction from cache.
264
- #[tokio::test]
265
- async fn test_cache_hit_behavior() {
266
- let config = ExtractionConfig {
267
- use_cache: true,
268
- ..Default::default()
269
- };
270
-
271
- let text = "Test text for caching behavior.";
272
- let text_bytes = text.as_bytes();
273
-
274
- let result1 = extract_bytes(text_bytes, "text/plain", &config)
275
- .await
276
- .expect("First extraction should succeed");
277
-
278
- let result2 = extract_bytes(text_bytes, "text/plain", &config)
279
- .await
280
- .expect("Second extraction should succeed");
281
-
282
- assert_eq!(result1.content, result2.content);
283
- }
284
-
285
- /// Test cache miss and invalidation.
286
- #[tokio::test]
287
- async fn test_cache_miss_invalidation() {
288
- let config = ExtractionConfig {
289
- use_cache: true,
290
- ..Default::default()
291
- };
292
-
293
- let text1 = "First text for cache test.";
294
- let text2 = "Second different text.";
295
-
296
- let result1 = extract_bytes(text1.as_bytes(), "text/plain", &config)
297
- .await
298
- .expect("First extraction should succeed");
299
-
300
- let result2 = extract_bytes(text2.as_bytes(), "text/plain", &config)
301
- .await
302
- .expect("Second extraction should succeed");
303
-
304
- assert_ne!(result1.content, result2.content);
305
- }
306
-
307
- /// Test custom cache directory (Note: OCR cache uses hardcoded directory).
308
- #[tokio::test]
309
- async fn test_custom_cache_directory() {
310
- let config = ExtractionConfig {
311
- use_cache: true,
312
- ..Default::default()
313
- };
314
-
315
- let text = "Test text for cache directory test.";
316
- let text_bytes = text.as_bytes();
317
-
318
- let result = extract_bytes(text_bytes, "text/plain", &config)
319
- .await
320
- .expect("Should extract successfully");
321
-
322
- assert!(!result.content.is_empty());
323
- }
324
-
325
- /// Test cache disabled - bypass cache.
326
- #[tokio::test]
327
- async fn test_cache_disabled() {
328
- let config = ExtractionConfig {
329
- use_cache: false,
330
- ..Default::default()
331
- };
332
-
333
- let text = "Test text without caching.";
334
- let text_bytes = text.as_bytes();
335
-
336
- let result1 = extract_bytes(text_bytes, "text/plain", &config)
337
- .await
338
- .expect("First extraction should succeed");
339
-
340
- let result2 = extract_bytes(text_bytes, "text/plain", &config)
341
- .await
342
- .expect("Second extraction should succeed");
343
-
344
- assert_eq!(result1.content, result2.content);
345
- }
346
-
347
- /// Test token reduction in aggressive mode.
348
- #[tokio::test]
349
- async fn test_token_reduction_aggressive() {
350
- let config = ExtractionConfig {
351
- token_reduction: Some(TokenReductionConfig {
352
- mode: "aggressive".to_string(),
353
- preserve_important_words: true,
354
- }),
355
- ..Default::default()
356
- };
357
-
358
- let text = "This is a very long sentence with many unnecessary words that could be reduced. ".repeat(5);
359
- let text_bytes = text.as_bytes();
360
-
361
- let result = extract_bytes(text_bytes, "text/plain", &config)
362
- .await
363
- .expect("Should extract successfully");
364
-
365
- assert!(!result.content.is_empty());
366
- }
367
-
368
- /// Test token reduction in conservative mode.
369
- #[tokio::test]
370
- async fn test_token_reduction_conservative() {
371
- let config = ExtractionConfig {
372
- token_reduction: Some(TokenReductionConfig {
373
- mode: "light".to_string(),
374
- preserve_important_words: true,
375
- }),
376
- ..Default::default()
377
- };
378
-
379
- let text = "Conservative token reduction test with moderate text length.";
380
- let text_bytes = text.as_bytes();
381
-
382
- let result = extract_bytes(text_bytes, "text/plain", &config)
383
- .await
384
- .expect("Should extract successfully");
385
-
386
- assert!(!result.content.is_empty());
387
- }
388
-
389
- /// Test token reduction disabled.
390
- #[tokio::test]
391
- async fn test_token_reduction_disabled() {
392
- let config = ExtractionConfig {
393
- token_reduction: Some(TokenReductionConfig {
394
- mode: "off".to_string(),
395
- preserve_important_words: false,
396
- }),
397
- ..Default::default()
398
- };
399
-
400
- let text = "Text without token reduction applied.";
401
- let text_bytes = text.as_bytes();
402
-
403
- let result = extract_bytes(text_bytes, "text/plain", &config)
404
- .await
405
- .expect("Should extract successfully");
406
-
407
- assert!(result.content.contains("without token reduction"));
408
- }
409
-
410
- /// Test quality processing enabled - quality scoring applied.
411
- #[tokio::test]
412
- #[cfg(feature = "quality")]
413
- async fn test_quality_processing_enabled() {
414
- let config = ExtractionConfig {
415
- enable_quality_processing: true,
416
- ..Default::default()
417
- };
418
-
419
- let text = "This is well-structured text. It has multiple sentences. And proper punctuation.";
420
- let text_bytes = text.as_bytes();
421
-
422
- let result = extract_bytes(text_bytes, "text/plain", &config)
423
- .await
424
- .expect("Should extract successfully");
425
-
426
- if let Some(score) = result.metadata.additional.get("quality_score") {
427
- let score_value = score.as_f64().unwrap();
428
- assert!((0.0..=1.0).contains(&score_value));
429
- }
430
-
431
- assert!(!result.content.is_empty());
432
- }
433
-
434
- /// Test quality processing calculates score for different text quality.
435
- #[tokio::test]
436
- #[cfg(feature = "quality")]
437
- async fn test_quality_threshold_filtering() {
438
- let config = ExtractionConfig {
439
- enable_quality_processing: true,
440
- ..Default::default()
441
- };
442
-
443
- let high_quality = "This is a well-structured document. It has proper sentences. And good formatting.";
444
- let result_high = extract_bytes(high_quality.as_bytes(), "text/plain", &config)
445
- .await
446
- .expect("Should extract successfully");
447
-
448
- let low_quality = "a b c d ....... word123mixed . . ";
449
- let result_low = extract_bytes(low_quality.as_bytes(), "text/plain", &config)
450
- .await
451
- .expect("Should extract successfully");
452
-
453
- assert!(
454
- result_high.metadata.additional.contains_key("quality_score"),
455
- "High quality should have score"
456
- );
457
- assert!(
458
- result_low.metadata.additional.contains_key("quality_score"),
459
- "Low quality should have score"
460
- );
461
-
462
- let score_high = result_high
463
- .metadata
464
- .additional
465
- .get("quality_score")
466
- .unwrap()
467
- .as_f64()
468
- .unwrap();
469
- let score_low = result_low
470
- .metadata
471
- .additional
472
- .get("quality_score")
473
- .unwrap()
474
- .as_f64()
475
- .unwrap();
476
-
477
- assert!((0.0..=1.0).contains(&score_high));
478
- assert!((0.0..=1.0).contains(&score_low));
479
- }
480
-
481
- /// Test quality processing disabled.
482
- #[tokio::test]
483
- async fn test_quality_processing_disabled() {
484
- let config = ExtractionConfig {
485
- enable_quality_processing: false,
486
- ..Default::default()
487
- };
488
-
489
- let text = "Text without quality processing.";
490
- let text_bytes = text.as_bytes();
491
-
492
- let result = extract_bytes(text_bytes, "text/plain", &config)
493
- .await
494
- .expect("Should extract successfully");
495
-
496
- assert!(!result.metadata.additional.contains_key("quality_score"));
497
- assert!(!result.content.is_empty());
498
- }
499
-
500
- /// Test chunking with embeddings using balanced preset.
501
- ///
502
- /// This test requires ONNX Runtime to be installed as a system dependency.
503
- /// On macOS with Homebrew: `brew install onnxruntime`
504
- /// On Linux: Install via your package manager or download from https://github.com/microsoft/onnxruntime/releases
505
- /// On Windows: Download from https://github.com/microsoft/onnxruntime/releases
506
- #[tokio::test]
507
- #[cfg(feature = "embeddings")]
508
- #[cfg_attr(target_os = "macos", ignore = "ONNX models not cached on macOS")]
509
- #[cfg_attr(target_os = "windows", ignore = "ONNX models not cached on Windows")]
510
- async fn test_chunking_with_embeddings() {
511
- use kreuzberg::core::config::EmbeddingConfig;
512
-
513
- let config = ExtractionConfig {
514
- chunking: Some(ChunkingConfig {
515
- max_chars: 100,
516
- max_overlap: 20,
517
- embedding: Some(EmbeddingConfig::default()),
518
- preset: None,
519
- }),
520
- ..Default::default()
521
- };
522
-
523
- let text = "This is a test document for embedding generation. ".repeat(10);
524
- let text_bytes = text.as_bytes();
525
-
526
- let result = extract_bytes(text_bytes, "text/plain", &config)
527
- .await
528
- .expect("Should extract successfully");
529
-
530
- assert!(result.chunks.is_some(), "Chunks should be present");
531
- let chunks = result.chunks.unwrap();
532
- assert!(chunks.len() > 1, "Should have multiple chunks");
533
-
534
- println!("Metadata: {:?}", result.metadata.additional);
535
-
536
- if let Some(error) = result.metadata.additional.get("embedding_error") {
537
- panic!("Embedding generation failed: {}", error);
538
- }
539
-
540
- assert!(
541
- result.metadata.additional.contains_key("embeddings_generated"),
542
- "Should have embeddings_generated metadata"
543
- );
544
- assert_eq!(
545
- result.metadata.additional.get("embeddings_generated").unwrap(),
546
- &serde_json::Value::Bool(true)
547
- );
548
-
549
- for chunk in &chunks {
550
- assert!(chunk.embedding.is_some(), "Each chunk should have an embedding");
551
- let embedding = chunk.embedding.as_ref().unwrap();
552
- assert_eq!(
553
- embedding.len(),
554
- 768,
555
- "Embedding should have 768 dimensions for balanced preset"
556
- );
557
-
558
- let magnitude: f32 = embedding.iter().map(|x| x * x).sum::<f32>().sqrt();
559
- assert!(
560
- (magnitude - 1.0).abs() < 0.01,
561
- "Embedding should be normalized (magnitude ~= 1.0)"
562
- );
563
- }
564
- }
565
-
566
- /// Test chunking with fast embedding preset.
567
- ///
568
- /// This test requires ONNX Runtime to be installed as a system dependency.
569
- /// On macOS with Homebrew: `brew install onnxruntime`
570
- /// On Linux: Install via your package manager or download from https://github.com/microsoft/onnxruntime/releases
571
- /// On Windows: Download from https://github.com/microsoft/onnxruntime/releases
572
- #[tokio::test]
573
- #[cfg(feature = "embeddings")]
574
- #[cfg_attr(target_os = "macos", ignore = "ONNX models not cached on macOS")]
575
- #[cfg_attr(target_os = "windows", ignore = "ONNX models not cached on Windows")]
576
- async fn test_chunking_with_fast_embeddings() {
577
- use kreuzberg::core::config::{EmbeddingConfig, EmbeddingModelType};
578
-
579
- let config = ExtractionConfig {
580
- chunking: Some(ChunkingConfig {
581
- max_chars: 100,
582
- max_overlap: 20,
583
- embedding: Some(EmbeddingConfig {
584
- model: EmbeddingModelType::Preset {
585
- name: "fast".to_string(),
586
- },
587
- ..Default::default()
588
- }),
589
- preset: None,
590
- }),
591
- ..Default::default()
592
- };
593
-
594
- let text = "Fast embedding test. ".repeat(10);
595
- let text_bytes = text.as_bytes();
596
-
597
- let result = extract_bytes(text_bytes, "text/plain", &config)
598
- .await
599
- .expect("Should extract successfully");
600
-
601
- let chunks = result.chunks.expect("Should have chunks");
602
- assert!(!chunks.is_empty(), "Should have at least one chunk");
603
-
604
- if let Some(error) = result.metadata.additional.get("embedding_error") {
605
- panic!("Embedding generation failed: {}", error);
606
- }
607
-
608
- for chunk in &chunks {
609
- let embedding = chunk.embedding.as_ref().expect("Should have embedding");
610
- assert_eq!(embedding.len(), 384, "Fast preset should produce 384-dim embeddings");
611
- }
612
- }
1
+ //! Configuration features integration tests.
2
+ //!
3
+ //! Tests for chunking, language detection, caching, token reduction, and quality processing.
4
+ //! Validates that configuration options work correctly end-to-end.
5
+
6
+ #[cfg(feature = "chunking")]
7
+ use kreuzberg::core::config::ChunkingConfig;
8
+ use kreuzberg::core::config::ExtractionConfig;
9
+ #[cfg(feature = "language-detection")]
10
+ use kreuzberg::core::config::LanguageDetectionConfig;
11
+ use kreuzberg::core::config::TokenReductionConfig;
12
+ use kreuzberg::core::extractor::extract_bytes;
13
+
14
+ mod helpers;
15
+
16
+ /// Test chunking enabled - text split into chunks.
17
+ #[tokio::test]
18
+ #[cfg(feature = "chunking")]
19
+ async fn test_chunking_enabled() {
20
+ let config = ExtractionConfig {
21
+ chunking: Some(ChunkingConfig {
22
+ max_chars: 50,
23
+ max_overlap: 10,
24
+ embedding: None,
25
+ preset: None,
26
+ }),
27
+ ..Default::default()
28
+ };
29
+
30
+ let text = "This is a long text that should be split into multiple chunks. ".repeat(10);
31
+ let text_bytes = text.as_bytes();
32
+
33
+ let result = extract_bytes(text_bytes, "text/plain", &config)
34
+ .await
35
+ .expect("Should extract successfully");
36
+
37
+ assert!(result.chunks.is_some(), "Chunks should be present");
38
+ let chunks = result.chunks.unwrap();
39
+ assert!(chunks.len() > 1, "Should have multiple chunks");
40
+
41
+ assert!(result.metadata.additional.contains_key("chunk_count"));
42
+ let chunk_count = result.metadata.additional.get("chunk_count").unwrap();
43
+ assert_eq!(
44
+ chunks.len(),
45
+ chunk_count.as_u64().unwrap() as usize,
46
+ "Chunks length should match chunk_count metadata"
47
+ );
48
+
49
+ for chunk in &chunks {
50
+ assert!(!chunk.content.is_empty(), "Chunk should not be empty");
51
+ assert!(
52
+ chunk.content.len() <= 50 + 10,
53
+ "Chunk length {} exceeds max_chars + overlap",
54
+ chunk.content.len()
55
+ );
56
+ }
57
+ }
58
+
59
+ /// Test chunking with overlap - overlap preserved between chunks.
60
+ #[tokio::test]
61
+ #[cfg(feature = "chunking")]
62
+ async fn test_chunking_with_overlap() {
63
+ let config = ExtractionConfig {
64
+ chunking: Some(ChunkingConfig {
65
+ max_chars: 100,
66
+ max_overlap: 20,
67
+ embedding: None,
68
+ preset: None,
69
+ }),
70
+ ..Default::default()
71
+ };
72
+
73
+ let text = "a".repeat(250);
74
+ let text_bytes = text.as_bytes();
75
+
76
+ let result = extract_bytes(text_bytes, "text/plain", &config)
77
+ .await
78
+ .expect("Should extract successfully");
79
+
80
+ assert!(result.chunks.is_some(), "Chunks should be present");
81
+ let chunks = result.chunks.unwrap();
82
+ assert!(chunks.len() >= 2, "Should have at least 2 chunks");
83
+
84
+ assert!(result.metadata.additional.contains_key("chunk_count"));
85
+
86
+ if chunks.len() >= 2 {
87
+ let chunk1 = &chunks[0];
88
+ let chunk2 = &chunks[1];
89
+
90
+ let chunk1_end = &chunk1.content[chunk1.content.len().saturating_sub(20)..];
91
+ assert!(
92
+ chunk2.content.starts_with(chunk1_end)
93
+ || chunk1_end.starts_with(&chunk2.content[..chunk1_end.len().min(chunk2.content.len())]),
94
+ "Chunks should have overlap"
95
+ );
96
+ }
97
+ }
98
+
99
+ /// Test chunking with custom sizes - custom chunk size and overlap.
100
+ #[tokio::test]
101
+ #[cfg(feature = "chunking")]
102
+ async fn test_chunking_custom_sizes() {
103
+ let config = ExtractionConfig {
104
+ chunking: Some(ChunkingConfig {
105
+ max_chars: 200,
106
+ max_overlap: 50,
107
+ embedding: None,
108
+ preset: None,
109
+ }),
110
+ ..Default::default()
111
+ };
112
+
113
+ let text = "Custom chunk test. ".repeat(50);
114
+ let text_bytes = text.as_bytes();
115
+
116
+ let result = extract_bytes(text_bytes, "text/plain", &config)
117
+ .await
118
+ .expect("Should extract successfully");
119
+
120
+ assert!(result.chunks.is_some(), "Chunks should be present");
121
+ let chunks = result.chunks.unwrap();
122
+ assert!(!chunks.is_empty(), "Should have at least 1 chunk");
123
+
124
+ assert!(result.metadata.additional.contains_key("chunk_count"));
125
+
126
+ for chunk in &chunks {
127
+ assert!(
128
+ chunk.content.len() <= 200 + 50,
129
+ "Chunk length {} exceeds custom max_chars + overlap",
130
+ chunk.content.len()
131
+ );
132
+ }
133
+ }
134
+
135
+ /// Test chunking disabled - no chunking when disabled.
136
+ #[tokio::test]
137
+ async fn test_chunking_disabled() {
138
+ let config = ExtractionConfig {
139
+ chunking: None,
140
+ ..Default::default()
141
+ };
142
+
143
+ let text = "This is a long text that should NOT be split into chunks. ".repeat(10);
144
+ let text_bytes = text.as_bytes();
145
+
146
+ let result = extract_bytes(text_bytes, "text/plain", &config)
147
+ .await
148
+ .expect("Should extract successfully");
149
+
150
+ assert!(result.chunks.is_none(), "Should not have chunks when chunking disabled");
151
+ assert!(
152
+ !result.metadata.additional.contains_key("chunk_count"),
153
+ "Should not have chunk_count when chunking disabled"
154
+ );
155
+
156
+ assert!(!result.content.is_empty(), "Content should be extracted");
157
+ assert!(result.content.contains("long text"), "Should contain original text");
158
+ }
159
+
160
+ /// Test language detection for single language document.
161
+ #[tokio::test]
162
+ #[cfg(feature = "language-detection")]
163
+ async fn test_language_detection_single() {
164
+ let config = ExtractionConfig {
165
+ language_detection: Some(LanguageDetectionConfig {
166
+ enabled: true,
167
+ min_confidence: 0.8,
168
+ detect_multiple: false,
169
+ }),
170
+ ..Default::default()
171
+ };
172
+
173
+ let text = "Hello world! This is English text. It should be detected as English language.";
174
+ let text_bytes = text.as_bytes();
175
+
176
+ let result = extract_bytes(text_bytes, "text/plain", &config)
177
+ .await
178
+ .expect("Should extract successfully");
179
+
180
+ assert!(result.detected_languages.is_some(), "Should detect language");
181
+ let languages = result.detected_languages.unwrap();
182
+ assert!(!languages.is_empty(), "Should detect at least one language");
183
+ assert_eq!(languages[0], "eng", "Should detect English");
184
+ }
185
+
186
+ /// Test language detection for multi-language document.
187
+ #[cfg_attr(coverage, ignore = "coverage instrumentation affects multi-language heuristics")]
188
+ #[tokio::test]
189
+ #[cfg(feature = "language-detection")]
190
+ async fn test_language_detection_multiple() {
191
+ let config = ExtractionConfig {
192
+ language_detection: Some(LanguageDetectionConfig {
193
+ enabled: true,
194
+ min_confidence: 0.7,
195
+ detect_multiple: true,
196
+ }),
197
+ ..Default::default()
198
+ };
199
+
200
+ let text = "Hello world! This is English. ".repeat(10) + "Hola mundo! Este es español. ".repeat(10).as_str();
201
+ let text_bytes = text.as_bytes();
202
+
203
+ let result = extract_bytes(text_bytes, "text/plain", &config)
204
+ .await
205
+ .expect("Should extract successfully");
206
+
207
+ assert!(result.detected_languages.is_some(), "Should detect languages");
208
+ let languages = result.detected_languages.unwrap();
209
+ assert!(!languages.is_empty(), "Should detect at least one language");
210
+ }
211
+
212
+ /// Test language detection with confidence threshold.
213
+ #[tokio::test]
214
+ #[cfg(feature = "language-detection")]
215
+ async fn test_language_detection_confidence() {
216
+ let config = ExtractionConfig {
217
+ language_detection: Some(LanguageDetectionConfig {
218
+ enabled: true,
219
+ min_confidence: 0.9,
220
+ detect_multiple: false,
221
+ }),
222
+ ..Default::default()
223
+ };
224
+
225
+ let text = "This is clear English text that should have high confidence.";
226
+ let text_bytes = text.as_bytes();
227
+
228
+ let result = extract_bytes(text_bytes, "text/plain", &config)
229
+ .await
230
+ .expect("Should extract successfully");
231
+
232
+ if let Some(languages) = result.detected_languages {
233
+ assert!(!languages.is_empty());
234
+ }
235
+ }
236
+
237
+ /// Test language detection disabled.
238
+ #[tokio::test]
239
+ #[cfg(feature = "language-detection")]
240
+ async fn test_language_detection_disabled() {
241
+ let config = ExtractionConfig {
242
+ language_detection: Some(LanguageDetectionConfig {
243
+ enabled: false,
244
+ min_confidence: 0.8,
245
+ detect_multiple: false,
246
+ }),
247
+ ..Default::default()
248
+ };
249
+
250
+ let text = "Hello world! This is English text.";
251
+ let text_bytes = text.as_bytes();
252
+
253
+ let result = extract_bytes(text_bytes, "text/plain", &config)
254
+ .await
255
+ .expect("Should extract successfully");
256
+
257
+ assert!(
258
+ result.detected_languages.is_none(),
259
+ "Should not detect language when disabled"
260
+ );
261
+ }
262
+
263
+ /// Test cache hit behavior - second extraction from cache.
264
+ #[tokio::test]
265
+ async fn test_cache_hit_behavior() {
266
+ let config = ExtractionConfig {
267
+ use_cache: true,
268
+ ..Default::default()
269
+ };
270
+
271
+ let text = "Test text for caching behavior.";
272
+ let text_bytes = text.as_bytes();
273
+
274
+ let result1 = extract_bytes(text_bytes, "text/plain", &config)
275
+ .await
276
+ .expect("First extraction should succeed");
277
+
278
+ let result2 = extract_bytes(text_bytes, "text/plain", &config)
279
+ .await
280
+ .expect("Second extraction should succeed");
281
+
282
+ assert_eq!(result1.content, result2.content);
283
+ }
284
+
285
+ /// Test cache miss and invalidation.
286
+ #[tokio::test]
287
+ async fn test_cache_miss_invalidation() {
288
+ let config = ExtractionConfig {
289
+ use_cache: true,
290
+ ..Default::default()
291
+ };
292
+
293
+ let text1 = "First text for cache test.";
294
+ let text2 = "Second different text.";
295
+
296
+ let result1 = extract_bytes(text1.as_bytes(), "text/plain", &config)
297
+ .await
298
+ .expect("First extraction should succeed");
299
+
300
+ let result2 = extract_bytes(text2.as_bytes(), "text/plain", &config)
301
+ .await
302
+ .expect("Second extraction should succeed");
303
+
304
+ assert_ne!(result1.content, result2.content);
305
+ }
306
+
307
+ /// Test custom cache directory (Note: OCR cache uses hardcoded directory).
308
+ #[tokio::test]
309
+ async fn test_custom_cache_directory() {
310
+ let config = ExtractionConfig {
311
+ use_cache: true,
312
+ ..Default::default()
313
+ };
314
+
315
+ let text = "Test text for cache directory test.";
316
+ let text_bytes = text.as_bytes();
317
+
318
+ let result = extract_bytes(text_bytes, "text/plain", &config)
319
+ .await
320
+ .expect("Should extract successfully");
321
+
322
+ assert!(!result.content.is_empty());
323
+ }
324
+
325
+ /// Test cache disabled - bypass cache.
326
+ #[tokio::test]
327
+ async fn test_cache_disabled() {
328
+ let config = ExtractionConfig {
329
+ use_cache: false,
330
+ ..Default::default()
331
+ };
332
+
333
+ let text = "Test text without caching.";
334
+ let text_bytes = text.as_bytes();
335
+
336
+ let result1 = extract_bytes(text_bytes, "text/plain", &config)
337
+ .await
338
+ .expect("First extraction should succeed");
339
+
340
+ let result2 = extract_bytes(text_bytes, "text/plain", &config)
341
+ .await
342
+ .expect("Second extraction should succeed");
343
+
344
+ assert_eq!(result1.content, result2.content);
345
+ }
346
+
347
+ /// Test token reduction in aggressive mode.
348
+ #[tokio::test]
349
+ async fn test_token_reduction_aggressive() {
350
+ let config = ExtractionConfig {
351
+ token_reduction: Some(TokenReductionConfig {
352
+ mode: "aggressive".to_string(),
353
+ preserve_important_words: true,
354
+ }),
355
+ ..Default::default()
356
+ };
357
+
358
+ let text = "This is a very long sentence with many unnecessary words that could be reduced. ".repeat(5);
359
+ let text_bytes = text.as_bytes();
360
+
361
+ let result = extract_bytes(text_bytes, "text/plain", &config)
362
+ .await
363
+ .expect("Should extract successfully");
364
+
365
+ assert!(!result.content.is_empty());
366
+ }
367
+
368
+ /// Test token reduction in conservative mode.
369
+ #[tokio::test]
370
+ async fn test_token_reduction_conservative() {
371
+ let config = ExtractionConfig {
372
+ token_reduction: Some(TokenReductionConfig {
373
+ mode: "light".to_string(),
374
+ preserve_important_words: true,
375
+ }),
376
+ ..Default::default()
377
+ };
378
+
379
+ let text = "Conservative token reduction test with moderate text length.";
380
+ let text_bytes = text.as_bytes();
381
+
382
+ let result = extract_bytes(text_bytes, "text/plain", &config)
383
+ .await
384
+ .expect("Should extract successfully");
385
+
386
+ assert!(!result.content.is_empty());
387
+ }
388
+
389
+ /// Test token reduction disabled.
390
+ #[tokio::test]
391
+ async fn test_token_reduction_disabled() {
392
+ let config = ExtractionConfig {
393
+ token_reduction: Some(TokenReductionConfig {
394
+ mode: "off".to_string(),
395
+ preserve_important_words: false,
396
+ }),
397
+ ..Default::default()
398
+ };
399
+
400
+ let text = "Text without token reduction applied.";
401
+ let text_bytes = text.as_bytes();
402
+
403
+ let result = extract_bytes(text_bytes, "text/plain", &config)
404
+ .await
405
+ .expect("Should extract successfully");
406
+
407
+ assert!(result.content.contains("without token reduction"));
408
+ }
409
+
410
+ /// Test quality processing enabled - quality scoring applied.
411
+ #[tokio::test]
412
+ #[cfg(feature = "quality")]
413
+ async fn test_quality_processing_enabled() {
414
+ let config = ExtractionConfig {
415
+ enable_quality_processing: true,
416
+ ..Default::default()
417
+ };
418
+
419
+ let text = "This is well-structured text. It has multiple sentences. And proper punctuation.";
420
+ let text_bytes = text.as_bytes();
421
+
422
+ let result = extract_bytes(text_bytes, "text/plain", &config)
423
+ .await
424
+ .expect("Should extract successfully");
425
+
426
+ if let Some(score) = result.metadata.additional.get("quality_score") {
427
+ let score_value = score.as_f64().unwrap();
428
+ assert!((0.0..=1.0).contains(&score_value));
429
+ }
430
+
431
+ assert!(!result.content.is_empty());
432
+ }
433
+
434
+ /// Test quality processing calculates score for different text quality.
435
+ #[tokio::test]
436
+ #[cfg(feature = "quality")]
437
+ async fn test_quality_threshold_filtering() {
438
+ let config = ExtractionConfig {
439
+ enable_quality_processing: true,
440
+ ..Default::default()
441
+ };
442
+
443
+ let high_quality = "This is a well-structured document. It has proper sentences. And good formatting.";
444
+ let result_high = extract_bytes(high_quality.as_bytes(), "text/plain", &config)
445
+ .await
446
+ .expect("Should extract successfully");
447
+
448
+ let low_quality = "a b c d ....... word123mixed . . ";
449
+ let result_low = extract_bytes(low_quality.as_bytes(), "text/plain", &config)
450
+ .await
451
+ .expect("Should extract successfully");
452
+
453
+ assert!(
454
+ result_high.metadata.additional.contains_key("quality_score"),
455
+ "High quality should have score"
456
+ );
457
+ assert!(
458
+ result_low.metadata.additional.contains_key("quality_score"),
459
+ "Low quality should have score"
460
+ );
461
+
462
+ let score_high = result_high
463
+ .metadata
464
+ .additional
465
+ .get("quality_score")
466
+ .unwrap()
467
+ .as_f64()
468
+ .unwrap();
469
+ let score_low = result_low
470
+ .metadata
471
+ .additional
472
+ .get("quality_score")
473
+ .unwrap()
474
+ .as_f64()
475
+ .unwrap();
476
+
477
+ assert!((0.0..=1.0).contains(&score_high));
478
+ assert!((0.0..=1.0).contains(&score_low));
479
+ }
480
+
481
+ /// Test quality processing disabled.
482
+ #[tokio::test]
483
+ async fn test_quality_processing_disabled() {
484
+ let config = ExtractionConfig {
485
+ enable_quality_processing: false,
486
+ ..Default::default()
487
+ };
488
+
489
+ let text = "Text without quality processing.";
490
+ let text_bytes = text.as_bytes();
491
+
492
+ let result = extract_bytes(text_bytes, "text/plain", &config)
493
+ .await
494
+ .expect("Should extract successfully");
495
+
496
+ assert!(!result.metadata.additional.contains_key("quality_score"));
497
+ assert!(!result.content.is_empty());
498
+ }
499
+
500
+ /// Test chunking with embeddings using balanced preset.
501
+ ///
502
+ /// This test requires ONNX Runtime to be installed as a system dependency.
503
+ /// On macOS with Homebrew: `brew install onnxruntime`
504
+ /// On Linux: Install via your package manager or download from https://github.com/microsoft/onnxruntime/releases
505
+ /// On Windows: Download from https://github.com/microsoft/onnxruntime/releases
506
+ #[tokio::test]
507
+ #[cfg(feature = "embeddings")]
508
+ #[cfg_attr(target_os = "macos", ignore = "ONNX models not cached on macOS")]
509
+ #[cfg_attr(target_os = "windows", ignore = "ONNX models not cached on Windows")]
510
+ async fn test_chunking_with_embeddings() {
511
+ use kreuzberg::core::config::EmbeddingConfig;
512
+
513
+ let config = ExtractionConfig {
514
+ chunking: Some(ChunkingConfig {
515
+ max_chars: 100,
516
+ max_overlap: 20,
517
+ embedding: Some(EmbeddingConfig::default()),
518
+ preset: None,
519
+ }),
520
+ ..Default::default()
521
+ };
522
+
523
+ let text = "This is a test document for embedding generation. ".repeat(10);
524
+ let text_bytes = text.as_bytes();
525
+
526
+ let result = extract_bytes(text_bytes, "text/plain", &config)
527
+ .await
528
+ .expect("Should extract successfully");
529
+
530
+ assert!(result.chunks.is_some(), "Chunks should be present");
531
+ let chunks = result.chunks.unwrap();
532
+ assert!(chunks.len() > 1, "Should have multiple chunks");
533
+
534
+ println!("Metadata: {:?}", result.metadata.additional);
535
+
536
+ if let Some(error) = result.metadata.additional.get("embedding_error") {
537
+ panic!("Embedding generation failed: {}", error);
538
+ }
539
+
540
+ assert!(
541
+ result.metadata.additional.contains_key("embeddings_generated"),
542
+ "Should have embeddings_generated metadata"
543
+ );
544
+ assert_eq!(
545
+ result.metadata.additional.get("embeddings_generated").unwrap(),
546
+ &serde_json::Value::Bool(true)
547
+ );
548
+
549
+ for chunk in &chunks {
550
+ assert!(chunk.embedding.is_some(), "Each chunk should have an embedding");
551
+ let embedding = chunk.embedding.as_ref().unwrap();
552
+ assert_eq!(
553
+ embedding.len(),
554
+ 768,
555
+ "Embedding should have 768 dimensions for balanced preset"
556
+ );
557
+
558
+ let magnitude: f32 = embedding.iter().map(|x| x * x).sum::<f32>().sqrt();
559
+ assert!(
560
+ (magnitude - 1.0).abs() < 0.01,
561
+ "Embedding should be normalized (magnitude ~= 1.0)"
562
+ );
563
+ }
564
+ }
565
+
566
+ /// Test chunking with fast embedding preset.
567
+ ///
568
+ /// This test requires ONNX Runtime to be installed as a system dependency.
569
+ /// On macOS with Homebrew: `brew install onnxruntime`
570
+ /// On Linux: Install via your package manager or download from https://github.com/microsoft/onnxruntime/releases
571
+ /// On Windows: Download from https://github.com/microsoft/onnxruntime/releases
572
+ #[tokio::test]
573
+ #[cfg(feature = "embeddings")]
574
+ #[cfg_attr(target_os = "macos", ignore = "ONNX models not cached on macOS")]
575
+ #[cfg_attr(target_os = "windows", ignore = "ONNX models not cached on Windows")]
576
+ async fn test_chunking_with_fast_embeddings() {
577
+ use kreuzberg::core::config::{EmbeddingConfig, EmbeddingModelType};
578
+
579
+ let config = ExtractionConfig {
580
+ chunking: Some(ChunkingConfig {
581
+ max_chars: 100,
582
+ max_overlap: 20,
583
+ embedding: Some(EmbeddingConfig {
584
+ model: EmbeddingModelType::Preset {
585
+ name: "fast".to_string(),
586
+ },
587
+ ..Default::default()
588
+ }),
589
+ preset: None,
590
+ }),
591
+ ..Default::default()
592
+ };
593
+
594
+ let text = "Fast embedding test. ".repeat(10);
595
+ let text_bytes = text.as_bytes();
596
+
597
+ let result = extract_bytes(text_bytes, "text/plain", &config)
598
+ .await
599
+ .expect("Should extract successfully");
600
+
601
+ let chunks = result.chunks.expect("Should have chunks");
602
+ assert!(!chunks.is_empty(), "Should have at least one chunk");
603
+
604
+ if let Some(error) = result.metadata.additional.get("embedding_error") {
605
+ panic!("Embedding generation failed: {}", error);
606
+ }
607
+
608
+ for chunk in &chunks {
609
+ let embedding = chunk.embedding.as_ref().expect("Should have embedding");
610
+ assert_eq!(embedding.len(), 384, "Fast preset should produce 384-dim embeddings");
611
+ }
612
+ }