kreuzberg 4.0.0.pre.rc.11 → 4.0.0.pre.rc.13
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/.gitignore +14 -14
- data/.rspec +3 -3
- data/.rubocop.yaml +1 -1
- data/.rubocop.yml +538 -538
- data/Gemfile +8 -8
- data/Gemfile.lock +2 -105
- data/README.md +454 -454
- data/Rakefile +25 -25
- data/Steepfile +47 -47
- data/examples/async_patterns.rb +341 -341
- data/ext/kreuzberg_rb/extconf.rb +45 -45
- data/ext/kreuzberg_rb/native/.cargo/config.toml +2 -2
- data/ext/kreuzberg_rb/native/Cargo.lock +6941 -6941
- data/ext/kreuzberg_rb/native/Cargo.toml +54 -54
- data/ext/kreuzberg_rb/native/README.md +425 -425
- data/ext/kreuzberg_rb/native/build.rs +15 -15
- data/ext/kreuzberg_rb/native/include/ieeefp.h +11 -11
- data/ext/kreuzberg_rb/native/include/msvc_compat/strings.h +14 -14
- data/ext/kreuzberg_rb/native/include/strings.h +20 -20
- data/ext/kreuzberg_rb/native/include/unistd.h +47 -47
- data/ext/kreuzberg_rb/native/src/lib.rs +3158 -3158
- data/extconf.rb +28 -28
- data/kreuzberg.gemspec +214 -214
- data/lib/kreuzberg/api_proxy.rb +142 -142
- data/lib/kreuzberg/cache_api.rb +81 -81
- data/lib/kreuzberg/cli.rb +55 -55
- data/lib/kreuzberg/cli_proxy.rb +127 -127
- data/lib/kreuzberg/config.rb +724 -724
- data/lib/kreuzberg/error_context.rb +80 -80
- data/lib/kreuzberg/errors.rb +118 -118
- data/lib/kreuzberg/extraction_api.rb +340 -340
- data/lib/kreuzberg/mcp_proxy.rb +186 -186
- data/lib/kreuzberg/ocr_backend_protocol.rb +113 -113
- data/lib/kreuzberg/post_processor_protocol.rb +86 -86
- data/lib/kreuzberg/result.rb +279 -279
- data/lib/kreuzberg/setup_lib_path.rb +80 -80
- data/lib/kreuzberg/validator_protocol.rb +89 -89
- data/lib/kreuzberg/version.rb +5 -5
- data/lib/kreuzberg.rb +109 -109
- data/lib/{libpdfium.dylib → pdfium.dll} +0 -0
- data/sig/kreuzberg/internal.rbs +184 -184
- data/sig/kreuzberg.rbs +546 -546
- data/spec/binding/cache_spec.rb +227 -227
- data/spec/binding/cli_proxy_spec.rb +85 -85
- data/spec/binding/cli_spec.rb +55 -55
- data/spec/binding/config_spec.rb +345 -345
- data/spec/binding/config_validation_spec.rb +283 -283
- data/spec/binding/error_handling_spec.rb +213 -213
- data/spec/binding/errors_spec.rb +66 -66
- data/spec/binding/plugins/ocr_backend_spec.rb +307 -307
- data/spec/binding/plugins/postprocessor_spec.rb +269 -269
- data/spec/binding/plugins/validator_spec.rb +274 -274
- data/spec/fixtures/config.toml +39 -39
- data/spec/fixtures/config.yaml +41 -41
- data/spec/fixtures/invalid_config.toml +4 -4
- data/spec/smoke/package_spec.rb +178 -178
- data/spec/spec_helper.rb +42 -42
- data/vendor/Cargo.toml +2 -1
- data/vendor/kreuzberg/Cargo.toml +2 -2
- data/vendor/kreuzberg/README.md +230 -230
- data/vendor/kreuzberg/benches/otel_overhead.rs +48 -48
- data/vendor/kreuzberg/build.rs +843 -843
- data/vendor/kreuzberg/src/api/error.rs +81 -81
- data/vendor/kreuzberg/src/api/handlers.rs +199 -199
- data/vendor/kreuzberg/src/api/mod.rs +79 -79
- data/vendor/kreuzberg/src/api/server.rs +353 -353
- data/vendor/kreuzberg/src/api/types.rs +170 -170
- data/vendor/kreuzberg/src/cache/mod.rs +1167 -1167
- data/vendor/kreuzberg/src/chunking/mod.rs +1877 -1877
- data/vendor/kreuzberg/src/chunking/processor.rs +220 -220
- data/vendor/kreuzberg/src/core/batch_mode.rs +95 -95
- data/vendor/kreuzberg/src/core/config.rs +1080 -1080
- data/vendor/kreuzberg/src/core/extractor.rs +1156 -1156
- data/vendor/kreuzberg/src/core/io.rs +329 -329
- data/vendor/kreuzberg/src/core/mime.rs +605 -605
- data/vendor/kreuzberg/src/core/mod.rs +47 -47
- data/vendor/kreuzberg/src/core/pipeline.rs +1184 -1184
- data/vendor/kreuzberg/src/embeddings.rs +500 -500
- data/vendor/kreuzberg/src/error.rs +431 -431
- data/vendor/kreuzberg/src/extraction/archive.rs +954 -954
- data/vendor/kreuzberg/src/extraction/docx.rs +398 -398
- data/vendor/kreuzberg/src/extraction/email.rs +854 -854
- data/vendor/kreuzberg/src/extraction/excel.rs +688 -688
- data/vendor/kreuzberg/src/extraction/html.rs +601 -601
- data/vendor/kreuzberg/src/extraction/image.rs +491 -491
- data/vendor/kreuzberg/src/extraction/libreoffice.rs +574 -562
- data/vendor/kreuzberg/src/extraction/markdown.rs +213 -213
- data/vendor/kreuzberg/src/extraction/mod.rs +81 -81
- data/vendor/kreuzberg/src/extraction/office_metadata/app_properties.rs +398 -398
- data/vendor/kreuzberg/src/extraction/office_metadata/core_properties.rs +247 -247
- data/vendor/kreuzberg/src/extraction/office_metadata/custom_properties.rs +240 -240
- data/vendor/kreuzberg/src/extraction/office_metadata/mod.rs +130 -130
- data/vendor/kreuzberg/src/extraction/office_metadata/odt_properties.rs +284 -284
- data/vendor/kreuzberg/src/extraction/pptx.rs +3100 -3100
- data/vendor/kreuzberg/src/extraction/structured.rs +490 -490
- data/vendor/kreuzberg/src/extraction/table.rs +328 -328
- data/vendor/kreuzberg/src/extraction/text.rs +269 -269
- data/vendor/kreuzberg/src/extraction/xml.rs +333 -333
- data/vendor/kreuzberg/src/extractors/archive.rs +447 -447
- data/vendor/kreuzberg/src/extractors/bibtex.rs +470 -470
- data/vendor/kreuzberg/src/extractors/docbook.rs +504 -504
- data/vendor/kreuzberg/src/extractors/docx.rs +400 -400
- data/vendor/kreuzberg/src/extractors/email.rs +157 -157
- data/vendor/kreuzberg/src/extractors/epub.rs +708 -708
- data/vendor/kreuzberg/src/extractors/excel.rs +345 -345
- data/vendor/kreuzberg/src/extractors/fictionbook.rs +492 -492
- data/vendor/kreuzberg/src/extractors/html.rs +407 -407
- data/vendor/kreuzberg/src/extractors/image.rs +219 -219
- data/vendor/kreuzberg/src/extractors/jats.rs +1054 -1054
- data/vendor/kreuzberg/src/extractors/jupyter.rs +368 -368
- data/vendor/kreuzberg/src/extractors/latex.rs +653 -653
- data/vendor/kreuzberg/src/extractors/markdown.rs +701 -701
- data/vendor/kreuzberg/src/extractors/mod.rs +429 -429
- data/vendor/kreuzberg/src/extractors/odt.rs +628 -628
- data/vendor/kreuzberg/src/extractors/opml.rs +635 -635
- data/vendor/kreuzberg/src/extractors/orgmode.rs +529 -529
- data/vendor/kreuzberg/src/extractors/pdf.rs +749 -722
- data/vendor/kreuzberg/src/extractors/pptx.rs +267 -267
- data/vendor/kreuzberg/src/extractors/rst.rs +577 -577
- data/vendor/kreuzberg/src/extractors/rtf.rs +809 -809
- data/vendor/kreuzberg/src/extractors/security.rs +484 -484
- data/vendor/kreuzberg/src/extractors/security_tests.rs +367 -367
- data/vendor/kreuzberg/src/extractors/structured.rs +142 -142
- data/vendor/kreuzberg/src/extractors/text.rs +265 -265
- data/vendor/kreuzberg/src/extractors/typst.rs +651 -651
- data/vendor/kreuzberg/src/extractors/xml.rs +147 -147
- data/vendor/kreuzberg/src/image/dpi.rs +164 -164
- data/vendor/kreuzberg/src/image/mod.rs +6 -6
- data/vendor/kreuzberg/src/image/preprocessing.rs +417 -417
- data/vendor/kreuzberg/src/image/resize.rs +89 -89
- data/vendor/kreuzberg/src/keywords/config.rs +154 -154
- data/vendor/kreuzberg/src/keywords/mod.rs +237 -237
- data/vendor/kreuzberg/src/keywords/processor.rs +275 -275
- data/vendor/kreuzberg/src/keywords/rake.rs +293 -293
- data/vendor/kreuzberg/src/keywords/types.rs +68 -68
- data/vendor/kreuzberg/src/keywords/yake.rs +163 -163
- data/vendor/kreuzberg/src/language_detection/mod.rs +985 -985
- data/vendor/kreuzberg/src/language_detection/processor.rs +219 -219
- data/vendor/kreuzberg/src/lib.rs +113 -113
- data/vendor/kreuzberg/src/mcp/mod.rs +35 -35
- data/vendor/kreuzberg/src/mcp/server.rs +2076 -2076
- data/vendor/kreuzberg/src/ocr/cache.rs +469 -469
- data/vendor/kreuzberg/src/ocr/error.rs +37 -37
- data/vendor/kreuzberg/src/ocr/hocr.rs +216 -216
- data/vendor/kreuzberg/src/ocr/mod.rs +58 -58
- data/vendor/kreuzberg/src/ocr/processor.rs +863 -863
- data/vendor/kreuzberg/src/ocr/table/mod.rs +4 -4
- data/vendor/kreuzberg/src/ocr/table/tsv_parser.rs +144 -144
- data/vendor/kreuzberg/src/ocr/tesseract_backend.rs +452 -452
- data/vendor/kreuzberg/src/ocr/types.rs +393 -393
- data/vendor/kreuzberg/src/ocr/utils.rs +47 -47
- data/vendor/kreuzberg/src/ocr/validation.rs +206 -206
- data/vendor/kreuzberg/src/panic_context.rs +154 -154
- data/vendor/kreuzberg/src/pdf/bindings.rs +44 -44
- data/vendor/kreuzberg/src/pdf/bundled.rs +346 -346
- data/vendor/kreuzberg/src/pdf/error.rs +130 -130
- data/vendor/kreuzberg/src/pdf/images.rs +139 -139
- data/vendor/kreuzberg/src/pdf/metadata.rs +489 -489
- data/vendor/kreuzberg/src/pdf/mod.rs +68 -68
- data/vendor/kreuzberg/src/pdf/rendering.rs +368 -368
- data/vendor/kreuzberg/src/pdf/table.rs +420 -420
- data/vendor/kreuzberg/src/pdf/text.rs +240 -240
- data/vendor/kreuzberg/src/plugins/extractor.rs +1044 -1044
- data/vendor/kreuzberg/src/plugins/mod.rs +212 -212
- data/vendor/kreuzberg/src/plugins/ocr.rs +639 -639
- data/vendor/kreuzberg/src/plugins/processor.rs +650 -650
- data/vendor/kreuzberg/src/plugins/registry.rs +1339 -1339
- data/vendor/kreuzberg/src/plugins/traits.rs +258 -258
- data/vendor/kreuzberg/src/plugins/validator.rs +967 -967
- data/vendor/kreuzberg/src/stopwords/mod.rs +1470 -1470
- data/vendor/kreuzberg/src/text/mod.rs +25 -25
- data/vendor/kreuzberg/src/text/quality.rs +697 -697
- data/vendor/kreuzberg/src/text/quality_processor.rs +219 -219
- data/vendor/kreuzberg/src/text/string_utils.rs +217 -217
- data/vendor/kreuzberg/src/text/token_reduction/cjk_utils.rs +164 -164
- data/vendor/kreuzberg/src/text/token_reduction/config.rs +100 -100
- data/vendor/kreuzberg/src/text/token_reduction/core.rs +796 -796
- data/vendor/kreuzberg/src/text/token_reduction/filters.rs +902 -902
- data/vendor/kreuzberg/src/text/token_reduction/mod.rs +160 -160
- data/vendor/kreuzberg/src/text/token_reduction/semantic.rs +619 -619
- data/vendor/kreuzberg/src/text/token_reduction/simd_text.rs +147 -147
- data/vendor/kreuzberg/src/types.rs +1055 -1055
- data/vendor/kreuzberg/src/utils/mod.rs +17 -17
- data/vendor/kreuzberg/src/utils/quality.rs +959 -959
- data/vendor/kreuzberg/src/utils/string_utils.rs +381 -381
- data/vendor/kreuzberg/stopwords/af_stopwords.json +53 -53
- data/vendor/kreuzberg/stopwords/ar_stopwords.json +482 -482
- data/vendor/kreuzberg/stopwords/bg_stopwords.json +261 -261
- data/vendor/kreuzberg/stopwords/bn_stopwords.json +400 -400
- data/vendor/kreuzberg/stopwords/br_stopwords.json +1205 -1205
- data/vendor/kreuzberg/stopwords/ca_stopwords.json +280 -280
- data/vendor/kreuzberg/stopwords/cs_stopwords.json +425 -425
- data/vendor/kreuzberg/stopwords/da_stopwords.json +172 -172
- data/vendor/kreuzberg/stopwords/de_stopwords.json +622 -622
- data/vendor/kreuzberg/stopwords/el_stopwords.json +849 -849
- data/vendor/kreuzberg/stopwords/en_stopwords.json +1300 -1300
- data/vendor/kreuzberg/stopwords/eo_stopwords.json +175 -175
- data/vendor/kreuzberg/stopwords/es_stopwords.json +734 -734
- data/vendor/kreuzberg/stopwords/et_stopwords.json +37 -37
- data/vendor/kreuzberg/stopwords/eu_stopwords.json +100 -100
- data/vendor/kreuzberg/stopwords/fa_stopwords.json +801 -801
- data/vendor/kreuzberg/stopwords/fi_stopwords.json +849 -849
- data/vendor/kreuzberg/stopwords/fr_stopwords.json +693 -693
- data/vendor/kreuzberg/stopwords/ga_stopwords.json +111 -111
- data/vendor/kreuzberg/stopwords/gl_stopwords.json +162 -162
- data/vendor/kreuzberg/stopwords/gu_stopwords.json +226 -226
- data/vendor/kreuzberg/stopwords/ha_stopwords.json +41 -41
- data/vendor/kreuzberg/stopwords/he_stopwords.json +196 -196
- data/vendor/kreuzberg/stopwords/hi_stopwords.json +227 -227
- data/vendor/kreuzberg/stopwords/hr_stopwords.json +181 -181
- data/vendor/kreuzberg/stopwords/hu_stopwords.json +791 -791
- data/vendor/kreuzberg/stopwords/hy_stopwords.json +47 -47
- data/vendor/kreuzberg/stopwords/id_stopwords.json +760 -760
- data/vendor/kreuzberg/stopwords/it_stopwords.json +634 -634
- data/vendor/kreuzberg/stopwords/ja_stopwords.json +136 -136
- data/vendor/kreuzberg/stopwords/kn_stopwords.json +84 -84
- data/vendor/kreuzberg/stopwords/ko_stopwords.json +681 -681
- data/vendor/kreuzberg/stopwords/ku_stopwords.json +64 -64
- data/vendor/kreuzberg/stopwords/la_stopwords.json +51 -51
- data/vendor/kreuzberg/stopwords/lt_stopwords.json +476 -476
- data/vendor/kreuzberg/stopwords/lv_stopwords.json +163 -163
- data/vendor/kreuzberg/stopwords/ml_stopwords.json +1 -1
- data/vendor/kreuzberg/stopwords/mr_stopwords.json +101 -101
- data/vendor/kreuzberg/stopwords/ms_stopwords.json +477 -477
- data/vendor/kreuzberg/stopwords/ne_stopwords.json +490 -490
- data/vendor/kreuzberg/stopwords/nl_stopwords.json +415 -415
- data/vendor/kreuzberg/stopwords/no_stopwords.json +223 -223
- data/vendor/kreuzberg/stopwords/pl_stopwords.json +331 -331
- data/vendor/kreuzberg/stopwords/pt_stopwords.json +562 -562
- data/vendor/kreuzberg/stopwords/ro_stopwords.json +436 -436
- data/vendor/kreuzberg/stopwords/ru_stopwords.json +561 -561
- data/vendor/kreuzberg/stopwords/si_stopwords.json +193 -193
- data/vendor/kreuzberg/stopwords/sk_stopwords.json +420 -420
- data/vendor/kreuzberg/stopwords/sl_stopwords.json +448 -448
- data/vendor/kreuzberg/stopwords/so_stopwords.json +32 -32
- data/vendor/kreuzberg/stopwords/st_stopwords.json +33 -33
- data/vendor/kreuzberg/stopwords/sv_stopwords.json +420 -420
- data/vendor/kreuzberg/stopwords/sw_stopwords.json +76 -76
- data/vendor/kreuzberg/stopwords/ta_stopwords.json +129 -129
- data/vendor/kreuzberg/stopwords/te_stopwords.json +54 -54
- data/vendor/kreuzberg/stopwords/th_stopwords.json +118 -118
- data/vendor/kreuzberg/stopwords/tl_stopwords.json +149 -149
- data/vendor/kreuzberg/stopwords/tr_stopwords.json +506 -506
- data/vendor/kreuzberg/stopwords/uk_stopwords.json +75 -75
- data/vendor/kreuzberg/stopwords/ur_stopwords.json +519 -519
- data/vendor/kreuzberg/stopwords/vi_stopwords.json +647 -647
- data/vendor/kreuzberg/stopwords/yo_stopwords.json +62 -62
- data/vendor/kreuzberg/stopwords/zh_stopwords.json +796 -796
- data/vendor/kreuzberg/stopwords/zu_stopwords.json +31 -31
- data/vendor/kreuzberg/tests/api_extract_multipart.rs +52 -52
- data/vendor/kreuzberg/tests/api_tests.rs +966 -966
- data/vendor/kreuzberg/tests/archive_integration.rs +545 -545
- data/vendor/kreuzberg/tests/batch_orchestration.rs +556 -556
- data/vendor/kreuzberg/tests/batch_processing.rs +318 -318
- data/vendor/kreuzberg/tests/bibtex_parity_test.rs +421 -421
- data/vendor/kreuzberg/tests/concurrency_stress.rs +533 -533
- data/vendor/kreuzberg/tests/config_features.rs +612 -612
- data/vendor/kreuzberg/tests/config_loading_tests.rs +416 -416
- data/vendor/kreuzberg/tests/core_integration.rs +510 -510
- data/vendor/kreuzberg/tests/csv_integration.rs +414 -414
- data/vendor/kreuzberg/tests/docbook_extractor_tests.rs +500 -500
- data/vendor/kreuzberg/tests/docx_metadata_extraction_test.rs +122 -122
- data/vendor/kreuzberg/tests/docx_vs_pandoc_comparison.rs +370 -370
- data/vendor/kreuzberg/tests/email_integration.rs +327 -327
- data/vendor/kreuzberg/tests/epub_native_extractor_tests.rs +275 -275
- data/vendor/kreuzberg/tests/error_handling.rs +402 -402
- data/vendor/kreuzberg/tests/fictionbook_extractor_tests.rs +228 -228
- data/vendor/kreuzberg/tests/format_integration.rs +164 -164
- data/vendor/kreuzberg/tests/helpers/mod.rs +142 -142
- data/vendor/kreuzberg/tests/html_table_test.rs +551 -551
- data/vendor/kreuzberg/tests/image_integration.rs +255 -255
- data/vendor/kreuzberg/tests/instrumentation_test.rs +139 -139
- data/vendor/kreuzberg/tests/jats_extractor_tests.rs +639 -639
- data/vendor/kreuzberg/tests/jupyter_extractor_tests.rs +704 -704
- data/vendor/kreuzberg/tests/keywords_integration.rs +479 -479
- data/vendor/kreuzberg/tests/keywords_quality.rs +509 -509
- data/vendor/kreuzberg/tests/latex_extractor_tests.rs +496 -496
- data/vendor/kreuzberg/tests/markdown_extractor_tests.rs +490 -490
- data/vendor/kreuzberg/tests/mime_detection.rs +429 -429
- data/vendor/kreuzberg/tests/ocr_configuration.rs +514 -514
- data/vendor/kreuzberg/tests/ocr_errors.rs +698 -698
- data/vendor/kreuzberg/tests/ocr_quality.rs +629 -629
- data/vendor/kreuzberg/tests/ocr_stress.rs +469 -469
- data/vendor/kreuzberg/tests/odt_extractor_tests.rs +674 -674
- data/vendor/kreuzberg/tests/opml_extractor_tests.rs +616 -616
- data/vendor/kreuzberg/tests/orgmode_extractor_tests.rs +822 -822
- data/vendor/kreuzberg/tests/pdf_integration.rs +45 -45
- data/vendor/kreuzberg/tests/pdfium_linking.rs +374 -374
- data/vendor/kreuzberg/tests/pipeline_integration.rs +1436 -1436
- data/vendor/kreuzberg/tests/plugin_ocr_backend_test.rs +776 -776
- data/vendor/kreuzberg/tests/plugin_postprocessor_test.rs +560 -560
- data/vendor/kreuzberg/tests/plugin_system.rs +927 -927
- data/vendor/kreuzberg/tests/plugin_validator_test.rs +783 -783
- data/vendor/kreuzberg/tests/registry_integration_tests.rs +587 -587
- data/vendor/kreuzberg/tests/rst_extractor_tests.rs +694 -694
- data/vendor/kreuzberg/tests/rtf_extractor_tests.rs +775 -775
- data/vendor/kreuzberg/tests/security_validation.rs +416 -416
- data/vendor/kreuzberg/tests/stopwords_integration_test.rs +888 -888
- data/vendor/kreuzberg/tests/test_fastembed.rs +631 -631
- data/vendor/kreuzberg/tests/typst_behavioral_tests.rs +1260 -1260
- data/vendor/kreuzberg/tests/typst_extractor_tests.rs +648 -648
- data/vendor/kreuzberg/tests/xlsx_metadata_extraction_test.rs +87 -87
- data/vendor/kreuzberg-ffi/Cargo.toml +3 -3
- data/vendor/kreuzberg-ffi/README.md +851 -851
- data/vendor/kreuzberg-ffi/build.rs +176 -176
- data/vendor/kreuzberg-ffi/cbindgen.toml +27 -27
- data/vendor/kreuzberg-ffi/kreuzberg-ffi-install.pc +12 -12
- data/vendor/kreuzberg-ffi/kreuzberg-ffi.pc.in +12 -12
- data/vendor/kreuzberg-ffi/kreuzberg.h +1087 -1087
- data/vendor/kreuzberg-ffi/src/lib.rs +3616 -3616
- data/vendor/kreuzberg-ffi/src/panic_shield.rs +247 -247
- data/vendor/kreuzberg-ffi/tests.disabled/README.md +48 -48
- data/vendor/kreuzberg-ffi/tests.disabled/config_loading_tests.rs +299 -299
- data/vendor/kreuzberg-ffi/tests.disabled/config_tests.rs +346 -346
- data/vendor/kreuzberg-ffi/tests.disabled/extractor_tests.rs +232 -232
- data/vendor/kreuzberg-ffi/tests.disabled/plugin_registration_tests.rs +470 -470
- data/vendor/kreuzberg-tesseract/.commitlintrc.json +13 -13
- data/vendor/kreuzberg-tesseract/.crate-ignore +2 -2
- data/vendor/kreuzberg-tesseract/Cargo.lock +2933 -2933
- data/vendor/kreuzberg-tesseract/Cargo.toml +2 -2
- data/vendor/kreuzberg-tesseract/LICENSE +22 -22
- data/vendor/kreuzberg-tesseract/README.md +399 -399
- data/vendor/kreuzberg-tesseract/build.rs +1354 -1354
- data/vendor/kreuzberg-tesseract/patches/README.md +71 -71
- data/vendor/kreuzberg-tesseract/patches/tesseract.diff +199 -199
- data/vendor/kreuzberg-tesseract/src/api.rs +1371 -1371
- data/vendor/kreuzberg-tesseract/src/choice_iterator.rs +77 -77
- data/vendor/kreuzberg-tesseract/src/enums.rs +297 -297
- data/vendor/kreuzberg-tesseract/src/error.rs +81 -81
- data/vendor/kreuzberg-tesseract/src/lib.rs +145 -145
- data/vendor/kreuzberg-tesseract/src/monitor.rs +57 -57
- data/vendor/kreuzberg-tesseract/src/mutable_iterator.rs +197 -197
- data/vendor/kreuzberg-tesseract/src/page_iterator.rs +253 -253
- data/vendor/kreuzberg-tesseract/src/result_iterator.rs +286 -286
- data/vendor/kreuzberg-tesseract/src/result_renderer.rs +183 -183
- data/vendor/kreuzberg-tesseract/tests/integration_test.rs +211 -211
- data/vendor/rb-sys/.cargo_vcs_info.json +5 -5
- data/vendor/rb-sys/Cargo.lock +393 -393
- data/vendor/rb-sys/Cargo.toml +70 -70
- data/vendor/rb-sys/Cargo.toml.orig +57 -57
- data/vendor/rb-sys/LICENSE-APACHE +190 -190
- data/vendor/rb-sys/LICENSE-MIT +21 -21
- data/vendor/rb-sys/build/features.rs +111 -111
- data/vendor/rb-sys/build/main.rs +286 -286
- data/vendor/rb-sys/build/stable_api_config.rs +155 -155
- data/vendor/rb-sys/build/version.rs +50 -50
- data/vendor/rb-sys/readme.md +36 -36
- data/vendor/rb-sys/src/bindings.rs +21 -21
- data/vendor/rb-sys/src/hidden.rs +11 -11
- data/vendor/rb-sys/src/lib.rs +35 -35
- data/vendor/rb-sys/src/macros.rs +371 -371
- data/vendor/rb-sys/src/memory.rs +53 -53
- data/vendor/rb-sys/src/ruby_abi_version.rs +38 -38
- data/vendor/rb-sys/src/special_consts.rs +31 -31
- data/vendor/rb-sys/src/stable_api/compiled.c +179 -179
- data/vendor/rb-sys/src/stable_api/compiled.rs +257 -257
- data/vendor/rb-sys/src/stable_api/ruby_2_7.rs +324 -324
- data/vendor/rb-sys/src/stable_api/ruby_3_0.rs +332 -332
- data/vendor/rb-sys/src/stable_api/ruby_3_1.rs +325 -325
- data/vendor/rb-sys/src/stable_api/ruby_3_2.rs +323 -323
- data/vendor/rb-sys/src/stable_api/ruby_3_3.rs +339 -339
- data/vendor/rb-sys/src/stable_api/ruby_3_4.rs +339 -339
- data/vendor/rb-sys/src/stable_api.rs +260 -260
- data/vendor/rb-sys/src/symbol.rs +31 -31
- data/vendor/rb-sys/src/tracking_allocator.rs +330 -330
- data/vendor/rb-sys/src/utils.rs +89 -89
- data/vendor/rb-sys/src/value_type.rs +7 -7
- metadata +7 -80
|
@@ -1,490 +1,490 @@
|
|
|
1
|
-
//! Comprehensive Markdown Extractor Tests
|
|
2
|
-
//!
|
|
3
|
-
//! This test suite uses Pandoc as a baseline for validating markdown extraction capabilities.
|
|
4
|
-
//! It tests:
|
|
5
|
-
//! - YAML frontmatter metadata extraction (both standard and extended fields)
|
|
6
|
-
//! - Table extraction from various markdown table formats
|
|
7
|
-
//! - Complex formatting and structure preservation
|
|
8
|
-
//! - Comparison with Pandoc's metadata extraction capabilities
|
|
9
|
-
|
|
10
|
-
#![cfg(feature = "office")]
|
|
11
|
-
|
|
12
|
-
use std::path::PathBuf;
|
|
13
|
-
|
|
14
|
-
use kreuzberg::core::config::ExtractionConfig;
|
|
15
|
-
use kreuzberg::extractors::markdown::MarkdownExtractor;
|
|
16
|
-
use kreuzberg::plugins::DocumentExtractor;
|
|
17
|
-
|
|
18
|
-
fn markdown_fixture_path(relative: &str) -> PathBuf {
|
|
19
|
-
PathBuf::from(env!("CARGO_MANIFEST_DIR"))
|
|
20
|
-
.join("../../test_documents/markdown")
|
|
21
|
-
.join(relative)
|
|
22
|
-
}
|
|
23
|
-
|
|
24
|
-
fn read_markdown_fixture(relative: &str) -> Vec<u8> {
|
|
25
|
-
let path = markdown_fixture_path(relative);
|
|
26
|
-
std::fs::read(&path).unwrap_or_else(|err| panic!("Failed to read markdown fixture {}: {}", path.display(), err))
|
|
27
|
-
}
|
|
28
|
-
|
|
29
|
-
/// Test comprehensive YAML frontmatter with all Pandoc-recognized fields
|
|
30
|
-
#[tokio::test]
|
|
31
|
-
async fn test_pandoc_baseline_yaml_fields() {
|
|
32
|
-
let markdown_with_yaml = b"---\ntitle: Test Document\nauthor: John Doe\ndate: 2024-01-15\nkeywords:\n - markdown\n - testing\n - rust\ndescription: A comprehensive test document\nabstract: This is an abstract\nsubject: Testing Subject\ncategory: Documentation\ntags:\n - important\n - draft\nlanguage: en\nversion: 1.0.0\n---\n\n# Content\n\nThis is the main content.";
|
|
33
|
-
|
|
34
|
-
let extractor = MarkdownExtractor::new();
|
|
35
|
-
let result = extractor
|
|
36
|
-
.extract_bytes(markdown_with_yaml, "text/markdown", &ExtractionConfig::default())
|
|
37
|
-
.await
|
|
38
|
-
.expect("Should extract markdown with frontmatter");
|
|
39
|
-
|
|
40
|
-
assert_eq!(
|
|
41
|
-
result.metadata.additional.get("title").and_then(|v| v.as_str()),
|
|
42
|
-
Some("Test Document")
|
|
43
|
-
);
|
|
44
|
-
assert_eq!(
|
|
45
|
-
result.metadata.additional.get("author").and_then(|v| v.as_str()),
|
|
46
|
-
Some("John Doe")
|
|
47
|
-
);
|
|
48
|
-
assert_eq!(result.metadata.date, Some("2024-01-15".to_string()));
|
|
49
|
-
|
|
50
|
-
assert!(result.metadata.additional.contains_key("keywords"));
|
|
51
|
-
let keywords = result
|
|
52
|
-
.metadata
|
|
53
|
-
.additional
|
|
54
|
-
.get("keywords")
|
|
55
|
-
.and_then(|v| v.as_str())
|
|
56
|
-
.unwrap_or("");
|
|
57
|
-
assert!(keywords.contains("markdown"));
|
|
58
|
-
assert!(keywords.contains("testing"));
|
|
59
|
-
assert!(keywords.contains("rust"));
|
|
60
|
-
|
|
61
|
-
assert_eq!(
|
|
62
|
-
result.metadata.additional.get("abstract").and_then(|v| v.as_str()),
|
|
63
|
-
Some("This is an abstract")
|
|
64
|
-
);
|
|
65
|
-
|
|
66
|
-
assert_eq!(result.metadata.subject, Some("Testing Subject".to_string()));
|
|
67
|
-
|
|
68
|
-
assert_eq!(
|
|
69
|
-
result.metadata.additional.get("category").and_then(|v| v.as_str()),
|
|
70
|
-
Some("Documentation")
|
|
71
|
-
);
|
|
72
|
-
|
|
73
|
-
assert!(result.metadata.additional.contains_key("tags"));
|
|
74
|
-
let tags = result
|
|
75
|
-
.metadata
|
|
76
|
-
.additional
|
|
77
|
-
.get("tags")
|
|
78
|
-
.and_then(|v| v.as_str())
|
|
79
|
-
.unwrap_or("");
|
|
80
|
-
assert!(tags.contains("important"));
|
|
81
|
-
assert!(tags.contains("draft"));
|
|
82
|
-
|
|
83
|
-
assert_eq!(
|
|
84
|
-
result.metadata.additional.get("language").and_then(|v| v.as_str()),
|
|
85
|
-
Some("en")
|
|
86
|
-
);
|
|
87
|
-
|
|
88
|
-
assert_eq!(
|
|
89
|
-
result.metadata.additional.get("version").and_then(|v| v.as_str()),
|
|
90
|
-
Some("1.0.0")
|
|
91
|
-
);
|
|
92
|
-
}
|
|
93
|
-
|
|
94
|
-
/// Test table extraction from pipe-format markdown tables
|
|
95
|
-
#[tokio::test]
|
|
96
|
-
async fn test_extract_simple_pipe_tables() {
|
|
97
|
-
let markdown = b"# Tables Example\n\n| Header 1 | Header 2 | Header 3 |\n|----------|----------|----------|\n| Row1Col1 | Row1Col2 | Row1Col3 |\n| Row2Col1 | Row2Col2 | Row2Col3 |";
|
|
98
|
-
|
|
99
|
-
let extractor = MarkdownExtractor::new();
|
|
100
|
-
let result = extractor
|
|
101
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
102
|
-
.await
|
|
103
|
-
.expect("Should extract tables");
|
|
104
|
-
|
|
105
|
-
assert!(!result.tables.is_empty(), "Should extract at least one table");
|
|
106
|
-
let table = &result.tables[0];
|
|
107
|
-
|
|
108
|
-
assert_eq!(table.cells.len(), 3, "Should have 3 rows (header + 2 data rows)");
|
|
109
|
-
assert_eq!(table.cells[0].len(), 3, "Should have 3 columns");
|
|
110
|
-
|
|
111
|
-
assert_eq!(table.cells[0][0], "Header 1");
|
|
112
|
-
assert_eq!(table.cells[0][1], "Header 2");
|
|
113
|
-
assert_eq!(table.cells[0][2], "Header 3");
|
|
114
|
-
|
|
115
|
-
assert_eq!(table.cells[1][0], "Row1Col1");
|
|
116
|
-
assert_eq!(table.cells[2][0], "Row2Col1");
|
|
117
|
-
|
|
118
|
-
assert!(table.markdown.contains("Header 1"));
|
|
119
|
-
assert!(table.markdown.contains("Row1Col1"));
|
|
120
|
-
assert!(table.markdown.contains("---"));
|
|
121
|
-
}
|
|
122
|
-
|
|
123
|
-
/// Test extraction of grid tables (as found in comprehensive.md)
|
|
124
|
-
#[tokio::test]
|
|
125
|
-
async fn test_extract_grid_tables() {
|
|
126
|
-
let markdown = b"# Grid Table Example\n\n+--------+--------+\n| Cell 1 | Cell 2 |\n+========+========+\n| Cell 3 | Cell 4 |\n+--------+--------+\n| Cell 5 | Cell 6 |\n+--------+--------+";
|
|
127
|
-
|
|
128
|
-
let extractor = MarkdownExtractor::new();
|
|
129
|
-
let result = extractor
|
|
130
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
131
|
-
.await
|
|
132
|
-
.expect("Should extract grid tables");
|
|
133
|
-
|
|
134
|
-
let _ = result.tables;
|
|
135
|
-
}
|
|
136
|
-
|
|
137
|
-
/// Test extraction of tables with multiple blocks in cells
|
|
138
|
-
#[tokio::test]
|
|
139
|
-
async fn test_extract_complex_table_cells() {
|
|
140
|
-
let markdown = b"# Complex Table\n\n| Header 1 | Header 2 |\n|----------|----------|\n| - bullet 1<br/>- bullet 2 | Simple text |\n| **Bold** *italic* | `code` |";
|
|
141
|
-
|
|
142
|
-
let extractor = MarkdownExtractor::new();
|
|
143
|
-
let result = extractor
|
|
144
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
145
|
-
.await
|
|
146
|
-
.expect("Should extract tables with complex formatting");
|
|
147
|
-
|
|
148
|
-
assert!(!result.tables.is_empty());
|
|
149
|
-
assert!(!result.content.is_empty());
|
|
150
|
-
}
|
|
151
|
-
|
|
152
|
-
/// Test multiline table from tables.markdown
|
|
153
|
-
#[tokio::test]
|
|
154
|
-
async fn test_pandoc_style_multiline_table() {
|
|
155
|
-
let markdown = b"Simple table with caption:\n\n Right Left Center Default\n ------- ------ -------- ---------\n 12 12 12 12\n 123 123 123 123\n 1 1 1 1\n\n : Demonstration of simple table syntax.";
|
|
156
|
-
|
|
157
|
-
let extractor = MarkdownExtractor::new();
|
|
158
|
-
let result = extractor
|
|
159
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
160
|
-
.await
|
|
161
|
-
.expect("Should extract pandoc-style tables");
|
|
162
|
-
|
|
163
|
-
assert!(result.content.contains("12") || result.content.contains("Demonstration"));
|
|
164
|
-
}
|
|
165
|
-
|
|
166
|
-
/// Test YAML frontmatter with author as list (Pandoc style)
|
|
167
|
-
#[tokio::test]
|
|
168
|
-
async fn test_pandoc_author_list() {
|
|
169
|
-
let markdown = b"% Title\n% Author One; Author Two; Author Three\n\n# Content\n\nBody text.";
|
|
170
|
-
|
|
171
|
-
let extractor = MarkdownExtractor::new();
|
|
172
|
-
let result = extractor
|
|
173
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
174
|
-
.await
|
|
175
|
-
.expect("Should extract markdown");
|
|
176
|
-
|
|
177
|
-
assert!(!result.content.is_empty());
|
|
178
|
-
}
|
|
179
|
-
|
|
180
|
-
/// Test YAML with array keywords field (Pandoc format)
|
|
181
|
-
#[tokio::test]
|
|
182
|
-
async fn test_keywords_array_extraction() {
|
|
183
|
-
let markdown =
|
|
184
|
-
b"---\ntitle: Document\nkeywords:\n - rust\n - markdown\n - pandoc\n---\n\n# Main Content\n\nText here.";
|
|
185
|
-
|
|
186
|
-
let extractor = MarkdownExtractor::new();
|
|
187
|
-
let result = extractor
|
|
188
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
189
|
-
.await
|
|
190
|
-
.expect("Should extract keywords array");
|
|
191
|
-
|
|
192
|
-
assert!(result.metadata.additional.contains_key("keywords"));
|
|
193
|
-
let keywords = result
|
|
194
|
-
.metadata
|
|
195
|
-
.additional
|
|
196
|
-
.get("keywords")
|
|
197
|
-
.and_then(|v| v.as_str())
|
|
198
|
-
.unwrap_or("");
|
|
199
|
-
assert!(keywords.contains("rust"));
|
|
200
|
-
assert!(keywords.contains("markdown"));
|
|
201
|
-
assert!(keywords.contains("pandoc"));
|
|
202
|
-
}
|
|
203
|
-
|
|
204
|
-
/// Test complex formatting in content (links, code, emphasis)
|
|
205
|
-
#[tokio::test]
|
|
206
|
-
async fn test_complex_markdown_formatting() {
|
|
207
|
-
let markdown = b"# Document\n\nThis is a paragraph with [links](http://example.com) and `code blocks`.\n\n## Subsection\n\n- **Bold text**\n- *Italic text*\n- ***Bold italic***\n\n```rust\nfn main() {\n println!(\"Hello, world!\");\n}\n```";
|
|
208
|
-
|
|
209
|
-
let extractor = MarkdownExtractor::new();
|
|
210
|
-
let result = extractor
|
|
211
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
212
|
-
.await
|
|
213
|
-
.expect("Should extract complex markdown");
|
|
214
|
-
|
|
215
|
-
assert!(result.content.contains("links"));
|
|
216
|
-
assert!(result.content.contains("code blocks"));
|
|
217
|
-
assert!(result.content.contains("Bold text"));
|
|
218
|
-
assert!(result.content.contains("println"));
|
|
219
|
-
}
|
|
220
|
-
|
|
221
|
-
/// Test extraction of raw HTML and LaTeX in markdown
|
|
222
|
-
#[tokio::test]
|
|
223
|
-
async fn test_raw_content_extraction() {
|
|
224
|
-
let markdown = b"# Document\n\nSome text.\n\n<div>Raw HTML</div>\n\nMore text.\n\n\\\\begin{equation}\nx = y\n\\\\end{equation}";
|
|
225
|
-
|
|
226
|
-
let extractor = MarkdownExtractor::new();
|
|
227
|
-
let result = extractor
|
|
228
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
229
|
-
.await
|
|
230
|
-
.expect("Should extract raw content");
|
|
231
|
-
|
|
232
|
-
assert!(!result.content.is_empty());
|
|
233
|
-
}
|
|
234
|
-
|
|
235
|
-
/// Test comprehensive.md from test_documents
|
|
236
|
-
#[tokio::test]
|
|
237
|
-
async fn test_comprehensive_md_extraction() {
|
|
238
|
-
let markdown = read_markdown_fixture("comprehensive.md");
|
|
239
|
-
|
|
240
|
-
let extractor = MarkdownExtractor::new();
|
|
241
|
-
let result = extractor
|
|
242
|
-
.extract_bytes(&markdown, "text/markdown", &ExtractionConfig::default())
|
|
243
|
-
.await
|
|
244
|
-
.expect("Should extract comprehensive.md");
|
|
245
|
-
|
|
246
|
-
assert!(!result.content.is_empty());
|
|
247
|
-
|
|
248
|
-
let _has_title_or_author =
|
|
249
|
-
result.metadata.additional.contains_key("title") || result.metadata.additional.contains_key("author");
|
|
250
|
-
|
|
251
|
-
assert!(result.content.contains("Additional markdown reader tests") || result.content.contains("markdown"));
|
|
252
|
-
|
|
253
|
-
let _ = result.tables;
|
|
254
|
-
}
|
|
255
|
-
|
|
256
|
-
/// Test tables.markdown from test_documents
|
|
257
|
-
#[tokio::test]
|
|
258
|
-
async fn test_tables_markdown_extraction() {
|
|
259
|
-
let markdown = read_markdown_fixture("tables.markdown");
|
|
260
|
-
|
|
261
|
-
let extractor = MarkdownExtractor::new();
|
|
262
|
-
let result = extractor
|
|
263
|
-
.extract_bytes(&markdown, "text/markdown", &ExtractionConfig::default())
|
|
264
|
-
.await
|
|
265
|
-
.expect("Should extract tables.markdown");
|
|
266
|
-
|
|
267
|
-
assert!(!result.content.is_empty());
|
|
268
|
-
|
|
269
|
-
assert!(result.content.contains("Right") || result.content.contains("Left") || result.content.contains("table"));
|
|
270
|
-
}
|
|
271
|
-
|
|
272
|
-
/// Test empty YAML frontmatter handling
|
|
273
|
-
#[tokio::test]
|
|
274
|
-
async fn test_empty_frontmatter() {
|
|
275
|
-
let markdown = b"---\n---\n\n# Main Title\n\nContent here.";
|
|
276
|
-
|
|
277
|
-
let extractor = MarkdownExtractor::new();
|
|
278
|
-
let result = extractor
|
|
279
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
280
|
-
.await
|
|
281
|
-
.expect("Should handle empty frontmatter");
|
|
282
|
-
|
|
283
|
-
assert!(result.content.contains("Main Title"));
|
|
284
|
-
assert!(result.content.contains("Content here"));
|
|
285
|
-
}
|
|
286
|
-
|
|
287
|
-
/// Test malformed YAML frontmatter fallback
|
|
288
|
-
#[tokio::test]
|
|
289
|
-
async fn test_malformed_frontmatter_graceful_fallback() {
|
|
290
|
-
let markdown = b"---\ninvalid: yaml: syntax: here:\n---\n\nContent here.";
|
|
291
|
-
|
|
292
|
-
let extractor = MarkdownExtractor::new();
|
|
293
|
-
let result = extractor
|
|
294
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
295
|
-
.await
|
|
296
|
-
.expect("Should handle malformed YAML gracefully");
|
|
297
|
-
|
|
298
|
-
assert!(!result.content.is_empty());
|
|
299
|
-
}
|
|
300
|
-
|
|
301
|
-
/// Test metadata field extraction for standard YAML fields
|
|
302
|
-
#[tokio::test]
|
|
303
|
-
async fn test_standard_yaml_metadata_fields() {
|
|
304
|
-
let markdown =
|
|
305
|
-
b"---\ntitle: Standard Fields Test\nauthor: Test Author\ndate: 2024-12-06\n---\n\n# Content\n\nTest body.";
|
|
306
|
-
|
|
307
|
-
let extractor = MarkdownExtractor::new();
|
|
308
|
-
let result = extractor
|
|
309
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
310
|
-
.await
|
|
311
|
-
.expect("Should extract standard fields");
|
|
312
|
-
|
|
313
|
-
assert_eq!(
|
|
314
|
-
result.metadata.additional.get("title").and_then(|v| v.as_str()),
|
|
315
|
-
Some("Standard Fields Test")
|
|
316
|
-
);
|
|
317
|
-
assert_eq!(
|
|
318
|
-
result.metadata.additional.get("author").and_then(|v| v.as_str()),
|
|
319
|
-
Some("Test Author")
|
|
320
|
-
);
|
|
321
|
-
assert_eq!(result.metadata.date, Some("2024-12-06".to_string()));
|
|
322
|
-
}
|
|
323
|
-
|
|
324
|
-
/// Test extraction of description field (maps to subject)
|
|
325
|
-
#[tokio::test]
|
|
326
|
-
async fn test_description_to_subject_mapping() {
|
|
327
|
-
let markdown = b"---\ntitle: Test\ndescription: This is the document description\n---\n\nContent.";
|
|
328
|
-
|
|
329
|
-
let extractor = MarkdownExtractor::new();
|
|
330
|
-
let result = extractor
|
|
331
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
332
|
-
.await
|
|
333
|
-
.expect("Should extract description");
|
|
334
|
-
|
|
335
|
-
assert_eq!(
|
|
336
|
-
result.metadata.subject,
|
|
337
|
-
Some("This is the document description".to_string())
|
|
338
|
-
);
|
|
339
|
-
}
|
|
340
|
-
|
|
341
|
-
/// Test multi-line title extraction from YAML
|
|
342
|
-
#[tokio::test]
|
|
343
|
-
async fn test_multiline_title_in_yaml() {
|
|
344
|
-
let markdown = b"---\ntitle: |\n This is a\n multi-line title\nauthor: Test\n---\n\n# Content\n\nBody.";
|
|
345
|
-
|
|
346
|
-
let extractor = MarkdownExtractor::new();
|
|
347
|
-
let result = extractor
|
|
348
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
349
|
-
.await
|
|
350
|
-
.expect("Should extract multiline title");
|
|
351
|
-
|
|
352
|
-
let title = result.metadata.additional.get("title").and_then(|v| v.as_str());
|
|
353
|
-
assert!(title.is_some());
|
|
354
|
-
}
|
|
355
|
-
|
|
356
|
-
/// Test table page numbering
|
|
357
|
-
#[tokio::test]
|
|
358
|
-
async fn test_table_page_numbering() {
|
|
359
|
-
let markdown = b"# Document\n\n| A | B |\n|---|---|\n| 1 | 2 |\n\nSome text between tables.\n\n| X | Y |\n|---|---|\n| 3 | 4 |";
|
|
360
|
-
|
|
361
|
-
let extractor = MarkdownExtractor::new();
|
|
362
|
-
let result = extractor
|
|
363
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
364
|
-
.await
|
|
365
|
-
.expect("Should extract multiple tables");
|
|
366
|
-
|
|
367
|
-
assert_eq!(result.tables.len(), 2);
|
|
368
|
-
assert_eq!(result.tables[0].page_number, 1);
|
|
369
|
-
assert_eq!(result.tables[1].page_number, 2);
|
|
370
|
-
}
|
|
371
|
-
|
|
372
|
-
/// Test unicode content extraction
|
|
373
|
-
#[tokio::test]
|
|
374
|
-
async fn test_unicode_markdown_extraction() {
|
|
375
|
-
let markdown = "---\ntitle: Unicode Test\nauthor: 日本人\n---\n\n# こんにちは\n\nThis document has:\n- 中文 (Chinese)\n- 日本語 (Japanese)\n- Русский (Russian)\n- العربية (Arabic)".as_bytes();
|
|
376
|
-
|
|
377
|
-
let extractor = MarkdownExtractor::new();
|
|
378
|
-
let result = extractor
|
|
379
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
380
|
-
.await
|
|
381
|
-
.expect("Should extract unicode content");
|
|
382
|
-
|
|
383
|
-
assert!(result.content.contains("こんにちは") || result.content.contains("Chinese"));
|
|
384
|
-
}
|
|
385
|
-
|
|
386
|
-
/// Test YAML list to comma-separated conversion for keywords
|
|
387
|
-
#[tokio::test]
|
|
388
|
-
async fn test_keywords_list_comma_separation() {
|
|
389
|
-
let markdown = b"---\nkeywords:\n - first\n - second\n - third\n---\n\nContent.";
|
|
390
|
-
|
|
391
|
-
let extractor = MarkdownExtractor::new();
|
|
392
|
-
let result = extractor
|
|
393
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
394
|
-
.await
|
|
395
|
-
.expect("Should extract keywords list");
|
|
396
|
-
|
|
397
|
-
let keywords = result
|
|
398
|
-
.metadata
|
|
399
|
-
.additional
|
|
400
|
-
.get("keywords")
|
|
401
|
-
.and_then(|v| v.as_str())
|
|
402
|
-
.unwrap_or("");
|
|
403
|
-
|
|
404
|
-
assert!(keywords.contains(","));
|
|
405
|
-
assert!(keywords.contains("first"));
|
|
406
|
-
assert!(keywords.contains("second"));
|
|
407
|
-
assert!(keywords.contains("third"));
|
|
408
|
-
}
|
|
409
|
-
|
|
410
|
-
/// Test extraction without any frontmatter
|
|
411
|
-
#[tokio::test]
|
|
412
|
-
async fn test_no_frontmatter_extraction() {
|
|
413
|
-
let markdown = b"# Document Title\n\nJust a document without frontmatter.\n\n## Section\n\nWith content.";
|
|
414
|
-
|
|
415
|
-
let extractor = MarkdownExtractor::new();
|
|
416
|
-
let result = extractor
|
|
417
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
418
|
-
.await
|
|
419
|
-
.expect("Should extract markdown without frontmatter");
|
|
420
|
-
|
|
421
|
-
assert!(result.content.contains("Document Title"));
|
|
422
|
-
assert!(result.content.contains("document") || result.content.contains("Section"));
|
|
423
|
-
|
|
424
|
-
let title = result.metadata.additional.get("title").and_then(|v| v.as_str());
|
|
425
|
-
assert_eq!(title, Some("Document Title"));
|
|
426
|
-
}
|
|
427
|
-
|
|
428
|
-
/// Test code block extraction
|
|
429
|
-
#[tokio::test]
|
|
430
|
-
async fn test_code_block_extraction() {
|
|
431
|
-
let markdown = b"# Code Examples\n\n```rust\nfn add(a: i32, b: i32) -> i32 {\n a + b\n}\n```\n\n```python\ndef add(a, b):\n return a + b\n```";
|
|
432
|
-
|
|
433
|
-
let extractor = MarkdownExtractor::new();
|
|
434
|
-
let result = extractor
|
|
435
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
436
|
-
.await
|
|
437
|
-
.expect("Should extract code blocks");
|
|
438
|
-
|
|
439
|
-
assert!(result.content.contains("add"));
|
|
440
|
-
assert!(result.content.contains("return"));
|
|
441
|
-
}
|
|
442
|
-
|
|
443
|
-
/// Test extraction with various mime types
|
|
444
|
-
#[tokio::test]
|
|
445
|
-
async fn test_supported_mime_types() {
|
|
446
|
-
let markdown = b"# Test\n\nContent.";
|
|
447
|
-
let extractor = MarkdownExtractor::new();
|
|
448
|
-
|
|
449
|
-
for mime_type in &["text/markdown", "text/x-markdown", "text/x-gfm", "text/x-commonmark"] {
|
|
450
|
-
let result = extractor
|
|
451
|
-
.extract_bytes(markdown, mime_type, &ExtractionConfig::default())
|
|
452
|
-
.await
|
|
453
|
-
.unwrap_or_else(|_| panic!("Should support {}", mime_type));
|
|
454
|
-
|
|
455
|
-
assert_eq!(result.mime_type, *mime_type);
|
|
456
|
-
assert!(result.content.contains("Test"));
|
|
457
|
-
}
|
|
458
|
-
}
|
|
459
|
-
|
|
460
|
-
/// Test that metadata extraction handles nested YAML structures
|
|
461
|
-
/// (Currently not fully supported - documents what's missing)
|
|
462
|
-
#[tokio::test]
|
|
463
|
-
async fn test_nested_yaml_awareness() {
|
|
464
|
-
let markdown = b"---\ntitle: Test\nmetadata:\n organization: Test Corp\n location:\n city: San Francisco\n state: CA\n---\n\nContent.";
|
|
465
|
-
|
|
466
|
-
let extractor = MarkdownExtractor::new();
|
|
467
|
-
let result = extractor
|
|
468
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
469
|
-
.await
|
|
470
|
-
.expect("Should extract document");
|
|
471
|
-
|
|
472
|
-
let title = result.metadata.additional.get("title").and_then(|v| v.as_str());
|
|
473
|
-
assert_eq!(title, Some("Test"));
|
|
474
|
-
}
|
|
475
|
-
|
|
476
|
-
/// Test extraction with special characters in metadata
|
|
477
|
-
#[tokio::test]
|
|
478
|
-
async fn test_special_characters_in_metadata() {
|
|
479
|
-
let markdown = b"---\ntitle: \"Document: Part 1 & 2\"\nauthor: O'Brien\nkeywords: \"C++, C#, F#\"\n---\n\nContent.";
|
|
480
|
-
|
|
481
|
-
let extractor = MarkdownExtractor::new();
|
|
482
|
-
let result = extractor
|
|
483
|
-
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
484
|
-
.await
|
|
485
|
-
.expect("Should extract with special characters");
|
|
486
|
-
|
|
487
|
-
let title = result.metadata.additional.get("title").and_then(|v| v.as_str());
|
|
488
|
-
assert!(title.is_some());
|
|
489
|
-
assert!(title.unwrap().contains("&") || title.unwrap().contains("Part"));
|
|
490
|
-
}
|
|
1
|
+
//! Comprehensive Markdown Extractor Tests
|
|
2
|
+
//!
|
|
3
|
+
//! This test suite uses Pandoc as a baseline for validating markdown extraction capabilities.
|
|
4
|
+
//! It tests:
|
|
5
|
+
//! - YAML frontmatter metadata extraction (both standard and extended fields)
|
|
6
|
+
//! - Table extraction from various markdown table formats
|
|
7
|
+
//! - Complex formatting and structure preservation
|
|
8
|
+
//! - Comparison with Pandoc's metadata extraction capabilities
|
|
9
|
+
|
|
10
|
+
#![cfg(feature = "office")]
|
|
11
|
+
|
|
12
|
+
use std::path::PathBuf;
|
|
13
|
+
|
|
14
|
+
use kreuzberg::core::config::ExtractionConfig;
|
|
15
|
+
use kreuzberg::extractors::markdown::MarkdownExtractor;
|
|
16
|
+
use kreuzberg::plugins::DocumentExtractor;
|
|
17
|
+
|
|
18
|
+
fn markdown_fixture_path(relative: &str) -> PathBuf {
|
|
19
|
+
PathBuf::from(env!("CARGO_MANIFEST_DIR"))
|
|
20
|
+
.join("../../test_documents/markdown")
|
|
21
|
+
.join(relative)
|
|
22
|
+
}
|
|
23
|
+
|
|
24
|
+
fn read_markdown_fixture(relative: &str) -> Vec<u8> {
|
|
25
|
+
let path = markdown_fixture_path(relative);
|
|
26
|
+
std::fs::read(&path).unwrap_or_else(|err| panic!("Failed to read markdown fixture {}: {}", path.display(), err))
|
|
27
|
+
}
|
|
28
|
+
|
|
29
|
+
/// Test comprehensive YAML frontmatter with all Pandoc-recognized fields
|
|
30
|
+
#[tokio::test]
|
|
31
|
+
async fn test_pandoc_baseline_yaml_fields() {
|
|
32
|
+
let markdown_with_yaml = b"---\ntitle: Test Document\nauthor: John Doe\ndate: 2024-01-15\nkeywords:\n - markdown\n - testing\n - rust\ndescription: A comprehensive test document\nabstract: This is an abstract\nsubject: Testing Subject\ncategory: Documentation\ntags:\n - important\n - draft\nlanguage: en\nversion: 1.0.0\n---\n\n# Content\n\nThis is the main content.";
|
|
33
|
+
|
|
34
|
+
let extractor = MarkdownExtractor::new();
|
|
35
|
+
let result = extractor
|
|
36
|
+
.extract_bytes(markdown_with_yaml, "text/markdown", &ExtractionConfig::default())
|
|
37
|
+
.await
|
|
38
|
+
.expect("Should extract markdown with frontmatter");
|
|
39
|
+
|
|
40
|
+
assert_eq!(
|
|
41
|
+
result.metadata.additional.get("title").and_then(|v| v.as_str()),
|
|
42
|
+
Some("Test Document")
|
|
43
|
+
);
|
|
44
|
+
assert_eq!(
|
|
45
|
+
result.metadata.additional.get("author").and_then(|v| v.as_str()),
|
|
46
|
+
Some("John Doe")
|
|
47
|
+
);
|
|
48
|
+
assert_eq!(result.metadata.date, Some("2024-01-15".to_string()));
|
|
49
|
+
|
|
50
|
+
assert!(result.metadata.additional.contains_key("keywords"));
|
|
51
|
+
let keywords = result
|
|
52
|
+
.metadata
|
|
53
|
+
.additional
|
|
54
|
+
.get("keywords")
|
|
55
|
+
.and_then(|v| v.as_str())
|
|
56
|
+
.unwrap_or("");
|
|
57
|
+
assert!(keywords.contains("markdown"));
|
|
58
|
+
assert!(keywords.contains("testing"));
|
|
59
|
+
assert!(keywords.contains("rust"));
|
|
60
|
+
|
|
61
|
+
assert_eq!(
|
|
62
|
+
result.metadata.additional.get("abstract").and_then(|v| v.as_str()),
|
|
63
|
+
Some("This is an abstract")
|
|
64
|
+
);
|
|
65
|
+
|
|
66
|
+
assert_eq!(result.metadata.subject, Some("Testing Subject".to_string()));
|
|
67
|
+
|
|
68
|
+
assert_eq!(
|
|
69
|
+
result.metadata.additional.get("category").and_then(|v| v.as_str()),
|
|
70
|
+
Some("Documentation")
|
|
71
|
+
);
|
|
72
|
+
|
|
73
|
+
assert!(result.metadata.additional.contains_key("tags"));
|
|
74
|
+
let tags = result
|
|
75
|
+
.metadata
|
|
76
|
+
.additional
|
|
77
|
+
.get("tags")
|
|
78
|
+
.and_then(|v| v.as_str())
|
|
79
|
+
.unwrap_or("");
|
|
80
|
+
assert!(tags.contains("important"));
|
|
81
|
+
assert!(tags.contains("draft"));
|
|
82
|
+
|
|
83
|
+
assert_eq!(
|
|
84
|
+
result.metadata.additional.get("language").and_then(|v| v.as_str()),
|
|
85
|
+
Some("en")
|
|
86
|
+
);
|
|
87
|
+
|
|
88
|
+
assert_eq!(
|
|
89
|
+
result.metadata.additional.get("version").and_then(|v| v.as_str()),
|
|
90
|
+
Some("1.0.0")
|
|
91
|
+
);
|
|
92
|
+
}
|
|
93
|
+
|
|
94
|
+
/// Test table extraction from pipe-format markdown tables
|
|
95
|
+
#[tokio::test]
|
|
96
|
+
async fn test_extract_simple_pipe_tables() {
|
|
97
|
+
let markdown = b"# Tables Example\n\n| Header 1 | Header 2 | Header 3 |\n|----------|----------|----------|\n| Row1Col1 | Row1Col2 | Row1Col3 |\n| Row2Col1 | Row2Col2 | Row2Col3 |";
|
|
98
|
+
|
|
99
|
+
let extractor = MarkdownExtractor::new();
|
|
100
|
+
let result = extractor
|
|
101
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
102
|
+
.await
|
|
103
|
+
.expect("Should extract tables");
|
|
104
|
+
|
|
105
|
+
assert!(!result.tables.is_empty(), "Should extract at least one table");
|
|
106
|
+
let table = &result.tables[0];
|
|
107
|
+
|
|
108
|
+
assert_eq!(table.cells.len(), 3, "Should have 3 rows (header + 2 data rows)");
|
|
109
|
+
assert_eq!(table.cells[0].len(), 3, "Should have 3 columns");
|
|
110
|
+
|
|
111
|
+
assert_eq!(table.cells[0][0], "Header 1");
|
|
112
|
+
assert_eq!(table.cells[0][1], "Header 2");
|
|
113
|
+
assert_eq!(table.cells[0][2], "Header 3");
|
|
114
|
+
|
|
115
|
+
assert_eq!(table.cells[1][0], "Row1Col1");
|
|
116
|
+
assert_eq!(table.cells[2][0], "Row2Col1");
|
|
117
|
+
|
|
118
|
+
assert!(table.markdown.contains("Header 1"));
|
|
119
|
+
assert!(table.markdown.contains("Row1Col1"));
|
|
120
|
+
assert!(table.markdown.contains("---"));
|
|
121
|
+
}
|
|
122
|
+
|
|
123
|
+
/// Test extraction of grid tables (as found in comprehensive.md)
|
|
124
|
+
#[tokio::test]
|
|
125
|
+
async fn test_extract_grid_tables() {
|
|
126
|
+
let markdown = b"# Grid Table Example\n\n+--------+--------+\n| Cell 1 | Cell 2 |\n+========+========+\n| Cell 3 | Cell 4 |\n+--------+--------+\n| Cell 5 | Cell 6 |\n+--------+--------+";
|
|
127
|
+
|
|
128
|
+
let extractor = MarkdownExtractor::new();
|
|
129
|
+
let result = extractor
|
|
130
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
131
|
+
.await
|
|
132
|
+
.expect("Should extract grid tables");
|
|
133
|
+
|
|
134
|
+
let _ = result.tables;
|
|
135
|
+
}
|
|
136
|
+
|
|
137
|
+
/// Test extraction of tables with multiple blocks in cells
|
|
138
|
+
#[tokio::test]
|
|
139
|
+
async fn test_extract_complex_table_cells() {
|
|
140
|
+
let markdown = b"# Complex Table\n\n| Header 1 | Header 2 |\n|----------|----------|\n| - bullet 1<br/>- bullet 2 | Simple text |\n| **Bold** *italic* | `code` |";
|
|
141
|
+
|
|
142
|
+
let extractor = MarkdownExtractor::new();
|
|
143
|
+
let result = extractor
|
|
144
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
145
|
+
.await
|
|
146
|
+
.expect("Should extract tables with complex formatting");
|
|
147
|
+
|
|
148
|
+
assert!(!result.tables.is_empty());
|
|
149
|
+
assert!(!result.content.is_empty());
|
|
150
|
+
}
|
|
151
|
+
|
|
152
|
+
/// Test multiline table from tables.markdown
|
|
153
|
+
#[tokio::test]
|
|
154
|
+
async fn test_pandoc_style_multiline_table() {
|
|
155
|
+
let markdown = b"Simple table with caption:\n\n Right Left Center Default\n ------- ------ -------- ---------\n 12 12 12 12\n 123 123 123 123\n 1 1 1 1\n\n : Demonstration of simple table syntax.";
|
|
156
|
+
|
|
157
|
+
let extractor = MarkdownExtractor::new();
|
|
158
|
+
let result = extractor
|
|
159
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
160
|
+
.await
|
|
161
|
+
.expect("Should extract pandoc-style tables");
|
|
162
|
+
|
|
163
|
+
assert!(result.content.contains("12") || result.content.contains("Demonstration"));
|
|
164
|
+
}
|
|
165
|
+
|
|
166
|
+
/// Test YAML frontmatter with author as list (Pandoc style)
|
|
167
|
+
#[tokio::test]
|
|
168
|
+
async fn test_pandoc_author_list() {
|
|
169
|
+
let markdown = b"% Title\n% Author One; Author Two; Author Three\n\n# Content\n\nBody text.";
|
|
170
|
+
|
|
171
|
+
let extractor = MarkdownExtractor::new();
|
|
172
|
+
let result = extractor
|
|
173
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
174
|
+
.await
|
|
175
|
+
.expect("Should extract markdown");
|
|
176
|
+
|
|
177
|
+
assert!(!result.content.is_empty());
|
|
178
|
+
}
|
|
179
|
+
|
|
180
|
+
/// Test YAML with array keywords field (Pandoc format)
|
|
181
|
+
#[tokio::test]
|
|
182
|
+
async fn test_keywords_array_extraction() {
|
|
183
|
+
let markdown =
|
|
184
|
+
b"---\ntitle: Document\nkeywords:\n - rust\n - markdown\n - pandoc\n---\n\n# Main Content\n\nText here.";
|
|
185
|
+
|
|
186
|
+
let extractor = MarkdownExtractor::new();
|
|
187
|
+
let result = extractor
|
|
188
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
189
|
+
.await
|
|
190
|
+
.expect("Should extract keywords array");
|
|
191
|
+
|
|
192
|
+
assert!(result.metadata.additional.contains_key("keywords"));
|
|
193
|
+
let keywords = result
|
|
194
|
+
.metadata
|
|
195
|
+
.additional
|
|
196
|
+
.get("keywords")
|
|
197
|
+
.and_then(|v| v.as_str())
|
|
198
|
+
.unwrap_or("");
|
|
199
|
+
assert!(keywords.contains("rust"));
|
|
200
|
+
assert!(keywords.contains("markdown"));
|
|
201
|
+
assert!(keywords.contains("pandoc"));
|
|
202
|
+
}
|
|
203
|
+
|
|
204
|
+
/// Test complex formatting in content (links, code, emphasis)
|
|
205
|
+
#[tokio::test]
|
|
206
|
+
async fn test_complex_markdown_formatting() {
|
|
207
|
+
let markdown = b"# Document\n\nThis is a paragraph with [links](http://example.com) and `code blocks`.\n\n## Subsection\n\n- **Bold text**\n- *Italic text*\n- ***Bold italic***\n\n```rust\nfn main() {\n println!(\"Hello, world!\");\n}\n```";
|
|
208
|
+
|
|
209
|
+
let extractor = MarkdownExtractor::new();
|
|
210
|
+
let result = extractor
|
|
211
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
212
|
+
.await
|
|
213
|
+
.expect("Should extract complex markdown");
|
|
214
|
+
|
|
215
|
+
assert!(result.content.contains("links"));
|
|
216
|
+
assert!(result.content.contains("code blocks"));
|
|
217
|
+
assert!(result.content.contains("Bold text"));
|
|
218
|
+
assert!(result.content.contains("println"));
|
|
219
|
+
}
|
|
220
|
+
|
|
221
|
+
/// Test extraction of raw HTML and LaTeX in markdown
|
|
222
|
+
#[tokio::test]
|
|
223
|
+
async fn test_raw_content_extraction() {
|
|
224
|
+
let markdown = b"# Document\n\nSome text.\n\n<div>Raw HTML</div>\n\nMore text.\n\n\\\\begin{equation}\nx = y\n\\\\end{equation}";
|
|
225
|
+
|
|
226
|
+
let extractor = MarkdownExtractor::new();
|
|
227
|
+
let result = extractor
|
|
228
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
229
|
+
.await
|
|
230
|
+
.expect("Should extract raw content");
|
|
231
|
+
|
|
232
|
+
assert!(!result.content.is_empty());
|
|
233
|
+
}
|
|
234
|
+
|
|
235
|
+
/// Test comprehensive.md from test_documents
|
|
236
|
+
#[tokio::test]
|
|
237
|
+
async fn test_comprehensive_md_extraction() {
|
|
238
|
+
let markdown = read_markdown_fixture("comprehensive.md");
|
|
239
|
+
|
|
240
|
+
let extractor = MarkdownExtractor::new();
|
|
241
|
+
let result = extractor
|
|
242
|
+
.extract_bytes(&markdown, "text/markdown", &ExtractionConfig::default())
|
|
243
|
+
.await
|
|
244
|
+
.expect("Should extract comprehensive.md");
|
|
245
|
+
|
|
246
|
+
assert!(!result.content.is_empty());
|
|
247
|
+
|
|
248
|
+
let _has_title_or_author =
|
|
249
|
+
result.metadata.additional.contains_key("title") || result.metadata.additional.contains_key("author");
|
|
250
|
+
|
|
251
|
+
assert!(result.content.contains("Additional markdown reader tests") || result.content.contains("markdown"));
|
|
252
|
+
|
|
253
|
+
let _ = result.tables;
|
|
254
|
+
}
|
|
255
|
+
|
|
256
|
+
/// Test tables.markdown from test_documents
|
|
257
|
+
#[tokio::test]
|
|
258
|
+
async fn test_tables_markdown_extraction() {
|
|
259
|
+
let markdown = read_markdown_fixture("tables.markdown");
|
|
260
|
+
|
|
261
|
+
let extractor = MarkdownExtractor::new();
|
|
262
|
+
let result = extractor
|
|
263
|
+
.extract_bytes(&markdown, "text/markdown", &ExtractionConfig::default())
|
|
264
|
+
.await
|
|
265
|
+
.expect("Should extract tables.markdown");
|
|
266
|
+
|
|
267
|
+
assert!(!result.content.is_empty());
|
|
268
|
+
|
|
269
|
+
assert!(result.content.contains("Right") || result.content.contains("Left") || result.content.contains("table"));
|
|
270
|
+
}
|
|
271
|
+
|
|
272
|
+
/// Test empty YAML frontmatter handling
|
|
273
|
+
#[tokio::test]
|
|
274
|
+
async fn test_empty_frontmatter() {
|
|
275
|
+
let markdown = b"---\n---\n\n# Main Title\n\nContent here.";
|
|
276
|
+
|
|
277
|
+
let extractor = MarkdownExtractor::new();
|
|
278
|
+
let result = extractor
|
|
279
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
280
|
+
.await
|
|
281
|
+
.expect("Should handle empty frontmatter");
|
|
282
|
+
|
|
283
|
+
assert!(result.content.contains("Main Title"));
|
|
284
|
+
assert!(result.content.contains("Content here"));
|
|
285
|
+
}
|
|
286
|
+
|
|
287
|
+
/// Test malformed YAML frontmatter fallback
|
|
288
|
+
#[tokio::test]
|
|
289
|
+
async fn test_malformed_frontmatter_graceful_fallback() {
|
|
290
|
+
let markdown = b"---\ninvalid: yaml: syntax: here:\n---\n\nContent here.";
|
|
291
|
+
|
|
292
|
+
let extractor = MarkdownExtractor::new();
|
|
293
|
+
let result = extractor
|
|
294
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
295
|
+
.await
|
|
296
|
+
.expect("Should handle malformed YAML gracefully");
|
|
297
|
+
|
|
298
|
+
assert!(!result.content.is_empty());
|
|
299
|
+
}
|
|
300
|
+
|
|
301
|
+
/// Test metadata field extraction for standard YAML fields
|
|
302
|
+
#[tokio::test]
|
|
303
|
+
async fn test_standard_yaml_metadata_fields() {
|
|
304
|
+
let markdown =
|
|
305
|
+
b"---\ntitle: Standard Fields Test\nauthor: Test Author\ndate: 2024-12-06\n---\n\n# Content\n\nTest body.";
|
|
306
|
+
|
|
307
|
+
let extractor = MarkdownExtractor::new();
|
|
308
|
+
let result = extractor
|
|
309
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
310
|
+
.await
|
|
311
|
+
.expect("Should extract standard fields");
|
|
312
|
+
|
|
313
|
+
assert_eq!(
|
|
314
|
+
result.metadata.additional.get("title").and_then(|v| v.as_str()),
|
|
315
|
+
Some("Standard Fields Test")
|
|
316
|
+
);
|
|
317
|
+
assert_eq!(
|
|
318
|
+
result.metadata.additional.get("author").and_then(|v| v.as_str()),
|
|
319
|
+
Some("Test Author")
|
|
320
|
+
);
|
|
321
|
+
assert_eq!(result.metadata.date, Some("2024-12-06".to_string()));
|
|
322
|
+
}
|
|
323
|
+
|
|
324
|
+
/// Test extraction of description field (maps to subject)
|
|
325
|
+
#[tokio::test]
|
|
326
|
+
async fn test_description_to_subject_mapping() {
|
|
327
|
+
let markdown = b"---\ntitle: Test\ndescription: This is the document description\n---\n\nContent.";
|
|
328
|
+
|
|
329
|
+
let extractor = MarkdownExtractor::new();
|
|
330
|
+
let result = extractor
|
|
331
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
332
|
+
.await
|
|
333
|
+
.expect("Should extract description");
|
|
334
|
+
|
|
335
|
+
assert_eq!(
|
|
336
|
+
result.metadata.subject,
|
|
337
|
+
Some("This is the document description".to_string())
|
|
338
|
+
);
|
|
339
|
+
}
|
|
340
|
+
|
|
341
|
+
/// Test multi-line title extraction from YAML
|
|
342
|
+
#[tokio::test]
|
|
343
|
+
async fn test_multiline_title_in_yaml() {
|
|
344
|
+
let markdown = b"---\ntitle: |\n This is a\n multi-line title\nauthor: Test\n---\n\n# Content\n\nBody.";
|
|
345
|
+
|
|
346
|
+
let extractor = MarkdownExtractor::new();
|
|
347
|
+
let result = extractor
|
|
348
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
349
|
+
.await
|
|
350
|
+
.expect("Should extract multiline title");
|
|
351
|
+
|
|
352
|
+
let title = result.metadata.additional.get("title").and_then(|v| v.as_str());
|
|
353
|
+
assert!(title.is_some());
|
|
354
|
+
}
|
|
355
|
+
|
|
356
|
+
/// Test table page numbering
|
|
357
|
+
#[tokio::test]
|
|
358
|
+
async fn test_table_page_numbering() {
|
|
359
|
+
let markdown = b"# Document\n\n| A | B |\n|---|---|\n| 1 | 2 |\n\nSome text between tables.\n\n| X | Y |\n|---|---|\n| 3 | 4 |";
|
|
360
|
+
|
|
361
|
+
let extractor = MarkdownExtractor::new();
|
|
362
|
+
let result = extractor
|
|
363
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
364
|
+
.await
|
|
365
|
+
.expect("Should extract multiple tables");
|
|
366
|
+
|
|
367
|
+
assert_eq!(result.tables.len(), 2);
|
|
368
|
+
assert_eq!(result.tables[0].page_number, 1);
|
|
369
|
+
assert_eq!(result.tables[1].page_number, 2);
|
|
370
|
+
}
|
|
371
|
+
|
|
372
|
+
/// Test unicode content extraction
|
|
373
|
+
#[tokio::test]
|
|
374
|
+
async fn test_unicode_markdown_extraction() {
|
|
375
|
+
let markdown = "---\ntitle: Unicode Test\nauthor: 日本人\n---\n\n# こんにちは\n\nThis document has:\n- 中文 (Chinese)\n- 日本語 (Japanese)\n- Русский (Russian)\n- العربية (Arabic)".as_bytes();
|
|
376
|
+
|
|
377
|
+
let extractor = MarkdownExtractor::new();
|
|
378
|
+
let result = extractor
|
|
379
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
380
|
+
.await
|
|
381
|
+
.expect("Should extract unicode content");
|
|
382
|
+
|
|
383
|
+
assert!(result.content.contains("こんにちは") || result.content.contains("Chinese"));
|
|
384
|
+
}
|
|
385
|
+
|
|
386
|
+
/// Test YAML list to comma-separated conversion for keywords
|
|
387
|
+
#[tokio::test]
|
|
388
|
+
async fn test_keywords_list_comma_separation() {
|
|
389
|
+
let markdown = b"---\nkeywords:\n - first\n - second\n - third\n---\n\nContent.";
|
|
390
|
+
|
|
391
|
+
let extractor = MarkdownExtractor::new();
|
|
392
|
+
let result = extractor
|
|
393
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
394
|
+
.await
|
|
395
|
+
.expect("Should extract keywords list");
|
|
396
|
+
|
|
397
|
+
let keywords = result
|
|
398
|
+
.metadata
|
|
399
|
+
.additional
|
|
400
|
+
.get("keywords")
|
|
401
|
+
.and_then(|v| v.as_str())
|
|
402
|
+
.unwrap_or("");
|
|
403
|
+
|
|
404
|
+
assert!(keywords.contains(","));
|
|
405
|
+
assert!(keywords.contains("first"));
|
|
406
|
+
assert!(keywords.contains("second"));
|
|
407
|
+
assert!(keywords.contains("third"));
|
|
408
|
+
}
|
|
409
|
+
|
|
410
|
+
/// Test extraction without any frontmatter
|
|
411
|
+
#[tokio::test]
|
|
412
|
+
async fn test_no_frontmatter_extraction() {
|
|
413
|
+
let markdown = b"# Document Title\n\nJust a document without frontmatter.\n\n## Section\n\nWith content.";
|
|
414
|
+
|
|
415
|
+
let extractor = MarkdownExtractor::new();
|
|
416
|
+
let result = extractor
|
|
417
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
418
|
+
.await
|
|
419
|
+
.expect("Should extract markdown without frontmatter");
|
|
420
|
+
|
|
421
|
+
assert!(result.content.contains("Document Title"));
|
|
422
|
+
assert!(result.content.contains("document") || result.content.contains("Section"));
|
|
423
|
+
|
|
424
|
+
let title = result.metadata.additional.get("title").and_then(|v| v.as_str());
|
|
425
|
+
assert_eq!(title, Some("Document Title"));
|
|
426
|
+
}
|
|
427
|
+
|
|
428
|
+
/// Test code block extraction
|
|
429
|
+
#[tokio::test]
|
|
430
|
+
async fn test_code_block_extraction() {
|
|
431
|
+
let markdown = b"# Code Examples\n\n```rust\nfn add(a: i32, b: i32) -> i32 {\n a + b\n}\n```\n\n```python\ndef add(a, b):\n return a + b\n```";
|
|
432
|
+
|
|
433
|
+
let extractor = MarkdownExtractor::new();
|
|
434
|
+
let result = extractor
|
|
435
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
436
|
+
.await
|
|
437
|
+
.expect("Should extract code blocks");
|
|
438
|
+
|
|
439
|
+
assert!(result.content.contains("add"));
|
|
440
|
+
assert!(result.content.contains("return"));
|
|
441
|
+
}
|
|
442
|
+
|
|
443
|
+
/// Test extraction with various mime types
|
|
444
|
+
#[tokio::test]
|
|
445
|
+
async fn test_supported_mime_types() {
|
|
446
|
+
let markdown = b"# Test\n\nContent.";
|
|
447
|
+
let extractor = MarkdownExtractor::new();
|
|
448
|
+
|
|
449
|
+
for mime_type in &["text/markdown", "text/x-markdown", "text/x-gfm", "text/x-commonmark"] {
|
|
450
|
+
let result = extractor
|
|
451
|
+
.extract_bytes(markdown, mime_type, &ExtractionConfig::default())
|
|
452
|
+
.await
|
|
453
|
+
.unwrap_or_else(|_| panic!("Should support {}", mime_type));
|
|
454
|
+
|
|
455
|
+
assert_eq!(result.mime_type, *mime_type);
|
|
456
|
+
assert!(result.content.contains("Test"));
|
|
457
|
+
}
|
|
458
|
+
}
|
|
459
|
+
|
|
460
|
+
/// Test that metadata extraction handles nested YAML structures
|
|
461
|
+
/// (Currently not fully supported - documents what's missing)
|
|
462
|
+
#[tokio::test]
|
|
463
|
+
async fn test_nested_yaml_awareness() {
|
|
464
|
+
let markdown = b"---\ntitle: Test\nmetadata:\n organization: Test Corp\n location:\n city: San Francisco\n state: CA\n---\n\nContent.";
|
|
465
|
+
|
|
466
|
+
let extractor = MarkdownExtractor::new();
|
|
467
|
+
let result = extractor
|
|
468
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
469
|
+
.await
|
|
470
|
+
.expect("Should extract document");
|
|
471
|
+
|
|
472
|
+
let title = result.metadata.additional.get("title").and_then(|v| v.as_str());
|
|
473
|
+
assert_eq!(title, Some("Test"));
|
|
474
|
+
}
|
|
475
|
+
|
|
476
|
+
/// Test extraction with special characters in metadata
|
|
477
|
+
#[tokio::test]
|
|
478
|
+
async fn test_special_characters_in_metadata() {
|
|
479
|
+
let markdown = b"---\ntitle: \"Document: Part 1 & 2\"\nauthor: O'Brien\nkeywords: \"C++, C#, F#\"\n---\n\nContent.";
|
|
480
|
+
|
|
481
|
+
let extractor = MarkdownExtractor::new();
|
|
482
|
+
let result = extractor
|
|
483
|
+
.extract_bytes(markdown, "text/markdown", &ExtractionConfig::default())
|
|
484
|
+
.await
|
|
485
|
+
.expect("Should extract with special characters");
|
|
486
|
+
|
|
487
|
+
let title = result.metadata.additional.get("title").and_then(|v| v.as_str());
|
|
488
|
+
assert!(title.is_some());
|
|
489
|
+
assert!(title.unwrap().contains("&") || title.unwrap().contains("Part"));
|
|
490
|
+
}
|