llguidance 0.7.11__tar.gz → 0.7.13__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (174) hide show
  1. {llguidance-0.7.11 → llguidance-0.7.13}/CHANGELOG.md +13 -3
  2. {llguidance-0.7.11 → llguidance-0.7.13}/Cargo.lock +7 -7
  3. {llguidance-0.7.11 → llguidance-0.7.13}/PKG-INFO +1 -1
  4. {llguidance-0.7.11 → llguidance-0.7.13}/docs/syntax.md +40 -3
  5. {llguidance-0.7.11 → llguidance-0.7.13}/json_stats/expected_maskbench.json +3 -7
  6. {llguidance-0.7.11 → llguidance-0.7.13}/json_stats/jstats.sh +5 -0
  7. {llguidance-0.7.11 → llguidance-0.7.13}/json_stats/src/json_stats.rs +33 -16
  8. {llguidance-0.7.11 → llguidance-0.7.13}/parser/Cargo.toml +2 -2
  9. {llguidance-0.7.11 → llguidance-0.7.13}/parser/llguidance.h +14 -0
  10. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/api.rs +9 -0
  11. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/earley/from_guidance.rs +60 -61
  12. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/earley/grammar.rs +86 -36
  13. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/earley/lexer.rs +33 -1
  14. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/earley/lexerspec.rs +39 -11
  15. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/earley/mod.rs +1 -2
  16. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/earley/parser.rs +46 -43
  17. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/earley/perf.rs +3 -0
  18. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/earley/regexvec.rs +19 -2
  19. llguidance-0.7.13/parser/src/earley/slicer.rs +394 -0
  20. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/factory.rs +37 -39
  21. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/ffi.rs +124 -63
  22. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/grammar_builder.rs +21 -24
  23. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/json/compiler.rs +19 -1
  24. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/lark/lexer.rs +1 -1
  25. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/lib.rs +1 -0
  26. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/matcher.rs +20 -1
  27. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/tokenparser.rs +18 -37
  28. {llguidance-0.7.11 → llguidance-0.7.13}/pyproject.toml +1 -1
  29. {llguidance-0.7.11 → llguidance-0.7.13}/python/llguidance/__init__.py +2 -0
  30. {llguidance-0.7.11 → llguidance-0.7.13}/python/llguidance/_lib.pyi +86 -5
  31. {llguidance-0.7.11 → llguidance-0.7.13}/python/torch_tests/test_matcher.py +31 -7
  32. {llguidance-0.7.11 → llguidance-0.7.13}/python_ext/Cargo.toml +1 -1
  33. {llguidance-0.7.11 → llguidance-0.7.13}/python_ext/src/lib.rs +2 -0
  34. {llguidance-0.7.11 → llguidance-0.7.13}/python_ext/src/llinterpreter.rs +14 -13
  35. {llguidance-0.7.11 → llguidance-0.7.13}/python_ext/src/llmatcher.rs +19 -22
  36. llguidance-0.7.13/python_ext/src/parserlimits.rs +106 -0
  37. {llguidance-0.7.11 → llguidance-0.7.13}/python_ext/src/py.rs +1 -2
  38. llguidance-0.7.13/sample_parser/src/minimal.rs +83 -0
  39. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/src/sample_parser.rs +70 -109
  40. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/tests/test_lark.rs +6 -3
  41. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/tests/test_ll.rs +1 -2
  42. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie/Cargo.toml +1 -1
  43. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie/src/tokenv.rs +4 -0
  44. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie/src/toktree.rs +13 -0
  45. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie_hf_downloader/Cargo.toml +1 -1
  46. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie_hf_tokenizers/Cargo.toml +1 -1
  47. llguidance-0.7.11/parser/src/earley/slicer.rs +0 -249
  48. llguidance-0.7.11/sample_parser/src/minimal.rs +0 -179
  49. {llguidance-0.7.11 → llguidance-0.7.13}/.github/workflows/rust.yml +0 -0
  50. {llguidance-0.7.11 → llguidance-0.7.13}/.github/workflows/wheels.yml +0 -0
  51. {llguidance-0.7.11 → llguidance-0.7.13}/.gitignore +0 -0
  52. {llguidance-0.7.11 → llguidance-0.7.13}/CODE_OF_CONDUCT.md +0 -0
  53. {llguidance-0.7.11 → llguidance-0.7.13}/Cargo.toml +0 -0
  54. {llguidance-0.7.11 → llguidance-0.7.13}/LICENSE +0 -0
  55. {llguidance-0.7.11 → llguidance-0.7.13}/README.md +0 -0
  56. {llguidance-0.7.11 → llguidance-0.7.13}/SECURITY.md +0 -0
  57. {llguidance-0.7.11 → llguidance-0.7.13}/SUPPORT.md +0 -0
  58. {llguidance-0.7.11 → llguidance-0.7.13}/c_sample/Makefile +0 -0
  59. {llguidance-0.7.11 → llguidance-0.7.13}/c_sample/README.md +0 -0
  60. {llguidance-0.7.11 → llguidance-0.7.13}/c_sample/c_sample.cpp +0 -0
  61. {llguidance-0.7.11 → llguidance-0.7.13}/docs/fast_forward.md +0 -0
  62. {llguidance-0.7.11 → llguidance-0.7.13}/docs/json_schema.md +0 -0
  63. {llguidance-0.7.11 → llguidance-0.7.13}/docs/mask_plot.png +0 -0
  64. {llguidance-0.7.11 → llguidance-0.7.13}/docs/optimizations.md +0 -0
  65. {llguidance-0.7.11 → llguidance-0.7.13}/docs/special_tokens.md +0 -0
  66. {llguidance-0.7.11 → llguidance-0.7.13}/docs/toktrie.md +0 -0
  67. {llguidance-0.7.11 → llguidance-0.7.13}/json_stats/Cargo.toml +0 -0
  68. {llguidance-0.7.11 → llguidance-0.7.13}/json_stats/scripts/split-stats.sh +0 -0
  69. {llguidance-0.7.11 → llguidance-0.7.13}/json_stats/scripts/split_plot.py +0 -0
  70. {llguidance-0.7.11 → llguidance-0.7.13}/json_stats/src/lib.rs +0 -0
  71. {llguidance-0.7.11 → llguidance-0.7.13}/json_stats/src/stats.rs +0 -0
  72. {llguidance-0.7.11 → llguidance-0.7.13}/parser/LICENSE +0 -0
  73. {llguidance-0.7.11 → llguidance-0.7.13}/parser/README.md +0 -0
  74. {llguidance-0.7.11 → llguidance-0.7.13}/parser/build.rs +0 -0
  75. {llguidance-0.7.11 → llguidance-0.7.13}/parser/cbindgen.toml +0 -0
  76. {llguidance-0.7.11 → llguidance-0.7.13}/parser/grammars/character.json +0 -0
  77. {llguidance-0.7.11 → llguidance-0.7.13}/parser/grammars/json.json +0 -0
  78. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/constraint.rs +0 -0
  79. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/ffi_par.rs +0 -0
  80. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/json/README.md +0 -0
  81. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/json/context_ref.rs +0 -0
  82. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/json/context_simple/context.rs +0 -0
  83. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/json/context_simple/draft.rs +0 -0
  84. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/json/context_simple/mod.rs +0 -0
  85. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/json/formats.rs +0 -0
  86. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/json/mod.rs +0 -0
  87. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/json/numeric.rs +0 -0
  88. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/json/schema.rs +0 -0
  89. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/json/shared_context.rs +0 -0
  90. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/json_validation.rs +0 -0
  91. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/lark/README.md +0 -0
  92. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/lark/ast.rs +0 -0
  93. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/lark/common.rs +0 -0
  94. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/lark/compiler.rs +0 -0
  95. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/lark/mod.rs +0 -0
  96. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/lark/parser.rs +0 -0
  97. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/logging.rs +0 -0
  98. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/output.rs +0 -0
  99. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/panic_utils.rs +0 -0
  100. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/stop_controller.rs +0 -0
  101. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/substring.rs +0 -0
  102. {llguidance-0.7.11 → llguidance-0.7.13}/parser/src/tokenizer_json.rs +0 -0
  103. {llguidance-0.7.11 → llguidance-0.7.13}/plan.md +0 -0
  104. {llguidance-0.7.11 → llguidance-0.7.13}/python/llguidance/_grammar_from.py +0 -0
  105. {llguidance-0.7.11 → llguidance-0.7.13}/python/llguidance/_struct_tag.py +0 -0
  106. {llguidance-0.7.11 → llguidance-0.7.13}/python/llguidance/_tokenizer.py +0 -0
  107. {llguidance-0.7.11 → llguidance-0.7.13}/python/llguidance/_util.py +0 -0
  108. {llguidance-0.7.11 → llguidance-0.7.13}/python/llguidance/cli.py +0 -0
  109. {llguidance-0.7.11 → llguidance-0.7.13}/python/llguidance/gbnf_to_lark.py +0 -0
  110. {llguidance-0.7.11 → llguidance-0.7.13}/python/llguidance/hf.py +0 -0
  111. {llguidance-0.7.11 → llguidance-0.7.13}/python/llguidance/mlx.py +0 -0
  112. {llguidance-0.7.11 → llguidance-0.7.13}/python/llguidance/numpy.py +0 -0
  113. {llguidance-0.7.11 → llguidance-0.7.13}/python/llguidance/py.typed +0 -0
  114. {llguidance-0.7.11 → llguidance-0.7.13}/python/llguidance/torch.py +0 -0
  115. {llguidance-0.7.11 → llguidance-0.7.13}/python/mypy.ini +0 -0
  116. {llguidance-0.7.11 → llguidance-0.7.13}/python/torch_tests/__init__.py +0 -0
  117. {llguidance-0.7.11 → llguidance-0.7.13}/python/torch_tests/test_bitmask.py +0 -0
  118. {llguidance-0.7.11 → llguidance-0.7.13}/python/torch_tests/test_hf.py +0 -0
  119. {llguidance-0.7.11 → llguidance-0.7.13}/python_ext/src/pyjson.rs +0 -0
  120. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/Cargo.toml +0 -0
  121. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/README.md +0 -0
  122. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/cli.sh +0 -0
  123. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/blog.sample.json +0 -0
  124. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/blog.schema.json +0 -0
  125. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/blog.schema.ll.json +0 -0
  126. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/from-llama.cpp/README.md +0 -0
  127. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/from-llama.cpp/arithmetic.gbnf +0 -0
  128. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/from-llama.cpp/c.gbnf +0 -0
  129. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/from-llama.cpp/chess.gbnf +0 -0
  130. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/from-llama.cpp/english.gbnf +0 -0
  131. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/from-llama.cpp/japanese.gbnf +0 -0
  132. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/from-llama.cpp/json.gbnf +0 -0
  133. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/from-llama.cpp/json_arr.gbnf +0 -0
  134. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/from-llama.cpp/list.gbnf +0 -0
  135. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/from-llama.cpp/vllm-sql.gbnf +0 -0
  136. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/lark.lark +0 -0
  137. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/rfc.lark +0 -0
  138. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/data/rfc.xml +0 -0
  139. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/gtest.sh +0 -0
  140. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/lark.sh +0 -0
  141. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/run.sh +0 -0
  142. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/src/lib.rs +0 -0
  143. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/tests/test_raw_parser.rs +0 -0
  144. {llguidance-0.7.11 → llguidance-0.7.13}/sample_parser/tests/test_stop.rs +0 -0
  145. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/annotate_asm.js +0 -0
  146. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/bump.py +0 -0
  147. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/cbindgen.sh +0 -0
  148. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/checklinks.py +0 -0
  149. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/checklinks.sh +0 -0
  150. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/ci-publish.py +0 -0
  151. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/disasm.sh +0 -0
  152. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/gbnf_to_lark.py +0 -0
  153. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/gen-testcase.py +0 -0
  154. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/git-version.sh +0 -0
  155. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/install-deps.sh +0 -0
  156. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/jsonschema-stats.js +0 -0
  157. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/remote-guidance-test.sh +0 -0
  158. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/rust-size.js +0 -0
  159. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/rust_size.py +0 -0
  160. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/test-guidance.sh +0 -0
  161. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/tokenizer_test.py +0 -0
  162. {llguidance-0.7.11 → llguidance-0.7.13}/scripts/update-git.py +0 -0
  163. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie/LICENSE +0 -0
  164. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie/README.md +0 -0
  165. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie/src/bytes.rs +0 -0
  166. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie/src/lib.rs +0 -0
  167. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie/src/recognizer.rs +0 -0
  168. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie/src/rng.rs +0 -0
  169. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie/src/svob.rs +0 -0
  170. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie/tests/test_svob.rs +0 -0
  171. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie_hf_downloader/LICENSE +0 -0
  172. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie_hf_downloader/src/lib.rs +0 -0
  173. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie_hf_tokenizers/LICENSE +0 -0
  174. {llguidance-0.7.11 → llguidance-0.7.13}/toktrie_hf_tokenizers/src/lib.rs +0 -0
@@ -4,6 +4,18 @@ All notable changes to this project will be documented in this file. Dates are d
4
4
 
5
5
  If a release doesn't introduce any interesting changes (build fixes etc.), it's skipped.
6
6
 
7
+ #### [0.7.13](https://github.com/guidance-ai/llguidance/compare/v0.7.12...0.7.13) 2025-04-05
8
+
9
+ - expose LLParserLimits in Python API [`598dc8f`](https://github.com/guidance-ai/llguidance/commit/598dc8f37f69f51244e54d9885445abf02a515a7)
10
+ - pre-compute lexer states for particularly large regexes (can be disabled in ParserLimits)
11
+
12
+ #### [0.7.12](https://github.com/guidance-ai/llguidance/compare/v0.7.11...0.7.12) 2025-04-04
13
+
14
+ - performance optimizations
15
+ - use factory in C FFI (otherwise slicer was not used)
16
+ - add some null checks and safety comments in C FFI
17
+ - implement subgrammar lexeme class merging; fixes [`#113`](https://github.com/guidance-ai/llguidance/issues/113)
18
+
7
19
  #### [0.7.11](https://github.com/guidance-ai/llguidance/compare/v0.7.10...0.7.11) 2025-03-27
8
20
 
9
21
  - add StructTag python API; fixes [`#146`](https://github.com/guidance-ai/llguidance/issues/146)
@@ -46,7 +58,7 @@ If a release doesn't introduce any interesting changes (build fixes etc.), it's
46
58
  #### [v0.7.1](https://github.com/guidance-ai/llguidance/compare/v0.7.0...v0.7.1) 2025-03-18
47
59
 
48
60
  - add `LLMatcher` interface in python
49
- - add whitespace_pattern to JsonCompileOptions [`04a5491`](https://github.com/guidance-ai/llguidance/commit/04a54912cf6d082669674340833f06385f7b66f8)
61
+ - add whitespace_pattern to JsonCompileOptions [`04a5491`](https://github.com/guidance-ai/llguidance/commit/04a54912cf6d082669674340833f06385f7b66f8)
50
62
  - enable mypy in CI [`#140`](https://github.com/guidance-ai/llguidance/pull/140)
51
63
  - add py.typed for annotations information [`#139`](https://github.com/guidance-ai/llguidance/pull/139)
52
64
  - fix clippy warnings
@@ -60,7 +72,6 @@ If a release doesn't introduce any interesting changes (build fixes etc.), it's
60
72
  - fix https://github.com/guidance-ai/guidance/issues/1131 - backtracking+prompt healing [`#1131`](https://github.com/guidance-ai/guidance/issues/1131)
61
73
  - optimize substring [`9950600`](https://github.com/guidance-ai/llguidance/commit/9950600f46e433b4c42506f8816f61cee331774f)
62
74
 
63
-
64
75
  #### [v0.6.29](https://github.com/guidance-ai/llguidance/compare/v0.6.28...v0.6.29) 2025-02-25
65
76
 
66
77
  - [JSON] "x-guidance" JsonCompileOptions [`#130`](https://github.com/guidance-ai/llguidance/pull/130)
@@ -110,4 +121,3 @@ Plus a few releases messing with, deps, unsafe code cleanup.
110
121
 
111
122
  - fixes for numeric tokens [`b7c9970`](https://github.com/guidance-ai/llguidance/commit/b7c99709a9cb7f7a8a3c4716092e4d94fae2ff2c)
112
123
  - make capture explicit in lark syntax [`2a57678`](https://github.com/guidance-ai/llguidance/commit/2a57678d9397e8be54cb0c9f14c4270604f8e1a5)
113
-
@@ -401,9 +401,9 @@ dependencies = [
401
401
 
402
402
  [[package]]
403
403
  name = "derivre"
404
- version = "0.3.1"
404
+ version = "0.3.5"
405
405
  source = "registry+https://github.com/rust-lang/crates.io-index"
406
- checksum = "3a3c2606b3ffc46f91fd62d954d55659ba9fb391bb673311b70f50daf9c15e49"
406
+ checksum = "2c15da97393af596fc344dc746f804c6742f1ddfd2e88fd7b75232a989c85c4e"
407
407
  dependencies = [
408
408
  "ahash",
409
409
  "anyhow",
@@ -1177,7 +1177,7 @@ checksum = "4ee93343901ab17bd981295f2cf0026d4ad018c7c31ba84549a4ddbb47a45104"
1177
1177
 
1178
1178
  [[package]]
1179
1179
  name = "llguidance"
1180
- version = "0.7.11"
1180
+ version = "0.7.13"
1181
1181
  dependencies = [
1182
1182
  "anyhow",
1183
1183
  "derivre",
@@ -1196,7 +1196,7 @@ dependencies = [
1196
1196
 
1197
1197
  [[package]]
1198
1198
  name = "llguidance_py"
1199
- version = "0.7.11"
1199
+ version = "0.7.13"
1200
1200
  dependencies = [
1201
1201
  "anyhow",
1202
1202
  "bytemuck",
@@ -2356,7 +2356,7 @@ dependencies = [
2356
2356
 
2357
2357
  [[package]]
2358
2358
  name = "toktrie"
2359
- version = "0.7.11"
2359
+ version = "0.7.13"
2360
2360
  dependencies = [
2361
2361
  "anyhow",
2362
2362
  "bytemuck",
@@ -2367,7 +2367,7 @@ dependencies = [
2367
2367
 
2368
2368
  [[package]]
2369
2369
  name = "toktrie_hf_downloader"
2370
- version = "0.7.11"
2370
+ version = "0.7.13"
2371
2371
  dependencies = [
2372
2372
  "anyhow",
2373
2373
  "hf-hub",
@@ -2378,7 +2378,7 @@ dependencies = [
2378
2378
 
2379
2379
  [[package]]
2380
2380
  name = "toktrie_hf_tokenizers"
2381
- version = "0.7.11"
2381
+ version = "0.7.13"
2382
2382
  dependencies = [
2383
2383
  "anyhow",
2384
2384
  "log",
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: llguidance
3
- Version: 0.7.11
3
+ Version: 0.7.13
4
4
  License-File: LICENSE
5
5
  Summary: Bindings for the Low-level Guidance (llguidance) Rust library for use within Guidance
6
6
  Author: Michal Moskal
@@ -293,12 +293,13 @@ that llguidance should be used to process the grammar.
293
293
  ### Multiple grammars
294
294
 
295
295
  The input to LLGuidance consists of a list of grammars. This can be accessed via
296
- [LLGuidance API](../parser/src/api.rs). Each of these can be a Lark grammar, a JSON schema,
297
- or a grammar in the API format. With the introduction of `%json` in Lark syntax
296
+ [LLGuidance API](../parser/src/api.rs). Each of these can be a Lark grammar or a JSON schema.
297
+ With the introduction of `%json` in Lark syntax
298
298
  there is less need now for using multiple grammars, but it is still supported.
299
+ We may add nested lark grammars in future.
299
300
 
300
301
  Inside of Lark grammar, you can reference other grammars using syntax like `@my_grammar`,
301
- refering to grammar with `"name": "my_grammar"` (numeric reference like `@17` are no longer supported).
302
+ refering to grammar with `"name": "my_grammar"` (numeric reference like `@17` are **no longer supported**).
302
303
  The top-level grammar is at index 0.
303
304
 
304
305
  You can specify temperature for subgrammar by referencing it via
@@ -317,6 +318,42 @@ Example:
317
318
  }
318
319
  ```
319
320
 
321
+ #### Subgrammar details
322
+
323
+ Generally, subgrammars share the same context-free grammar but have a separate
324
+ set of lexemes (lexeme class).
325
+ The parser keeps track of a stack of lexeme classes, and considers the top one
326
+ to be the current lexeme class.
327
+ The `%ignore` is applied based on the top lexeme class.
328
+
329
+ Temperature and `max_tokens` can be applied to the current lexeme class as well.
330
+
331
+ There may be issues with subgrammars spanning LLM token boundaries.
332
+
333
+ If two lexeme classes share the `%ignore` regex, and `max_tokens=` and `temperature=`
334
+ are not used, the lexeme classes are merged, which generally allows for parsing
335
+ of more grammars.
336
+ For example, consider:
337
+
338
+ ```lark
339
+ start: a | b
340
+ a: %json { A }
341
+ b: %json { B }
342
+ ```
343
+
344
+ Normally, the parser would have to pick between lexeme class for either A or B
345
+ at the first `{` (it would always pick A since it comes first in the grammar).
346
+ However, if the classes for A and B are merged, the grammar will be equivalent to
347
+ `start: %json { "anyOf": [A, B] }` which is generally what the
348
+ [users expect](https://github.com/guidance-ai/llguidance/issues/113).
349
+
350
+
351
+ ### Features to avoid
352
+
353
+ - `stop=...` - use `suffix=...` or just `lazy`
354
+ - `max_tokens=...` - any use of `max_tokens` will disable rollback, which is needed for spec-decoding; it also makes the parser slower and prevents subgrammar merging
355
+ - `temperature=...` - this is not supported in most server side integrations and prevents subgrammar merging
356
+
320
357
  ### Unsupported Lark features
321
358
 
322
359
  Following features of Lark syntax are currently not supported:
@@ -1291,9 +1291,7 @@
1291
1291
  "Github_easy---o21209.json": {},
1292
1292
  "Github_easy---o21393.json": {},
1293
1293
  "Github_easy---o21455.json": {},
1294
- "Github_easy---o21456.json": {
1295
- "json_error": "Unable to determine if regex is empty: (And (Regex \"([^@^\\\\s]+@[^@^\\\\.^\\\\s]+(\\\\.[^@^\\\\.^\\\\s]*)*.gov.uk).*\") (Regex \"(?s:.{5,254})\"))"
1296
- },
1294
+ "Github_easy---o21456.json": {},
1297
1295
  "Github_easy---o21458.json": {},
1298
1296
  "Github_easy---o21459.json": {},
1299
1297
  "Github_easy---o21460.json": {},
@@ -5037,7 +5035,7 @@
5037
5035
  "Github_hard---o69969.json": {},
5038
5036
  "Github_hard---o69970.json": {},
5039
5037
  "Github_hard---o69972.json": {
5040
- "json_error": "Unable to determine if regex is empty: (And (And (Regex \"(\\\\w+([\\\\.-]?\\\\w+)*@\\\\w+([\\\\.-]?\\\\w+)*(\\\\.\\\\w{2,})+)\") (Regex \"((?P<local_part>(?P<dot_string>[^\\\\s@\\\\.]+(\\\\.[^\\\\s@\\\\.]+)*))@((?P<domain>(?P<sub_domain>[a-zA-Z0-9]([a-zA-Z0-9-]*[a-zA-Z0-9])?)(\\\\.(?P<sub_domain2>[a-zA-Z0-9]([a-zA-Z0-9-]*[a-zA-Z0-9])?))*)|\\\\[(?P<ipv4>((([0-9])|(([1-9])[0-9]|(25[0-5]|(2[0-4]|(1)[0-9])[0-9])))\\\\.){3}(([0-9])|(([1-9])[0-9]|(25[0-5]|(2[0-4]|(1)[0-9])[0-9]))))\\\\]))\")) (Regex \"(?s:.{6,})\"))"
5038
+ "validation_error": "test #0: token not accepted at ⟦loc‧-‧1‧\",\"‧code‧\":\"‧LOC‧-‧1‧\",\"‧name‧\":\"‧Main‧ Library‧\",\"‧library‧\":{\"‧$‧ref‧\":\"‧https‧://‧ils‧.r‧ero‧.ch‧/api‧/lib‧raries‧/lib‧-‧1‧\"},\"‧allow‧_request‧\":‧true‧,\"‧send‧_notification‧\":‧true‧,\"‧notification‧_email‧\":\"‧library‧@example‧.com‧\",\"⟧ * ⟦is⟧ * ⟦_online‧\":‧false‧,\"⟧ forced tokens \"⟦restrict‧_pick‧up‧_to⟧\" != \"⟦is‧_online‧\\\":‧false⟧\""
5041
5039
  },
5042
5040
  "Github_hard---o69976.json": {},
5043
5041
  "Github_hard---o70037.json": {},
@@ -7356,9 +7354,7 @@
7356
7354
  "Github_medium---o6378.json": {},
7357
7355
  "Github_medium---o63935.json": {},
7358
7356
  "Github_medium---o63937.json": {},
7359
- "Github_medium---o63939.json": {
7360
- "json_error": "Unable to determine if regex is empty: (And (Regex \"([\\\\w\\\\-\\\\./]+\\\\.php+)\") (Regex \"(?s:.{16,1024})\"))"
7361
- },
7357
+ "Github_medium---o63939.json": {},
7362
7358
  "Github_medium---o63941.json": {},
7363
7359
  "Github_medium---o63945.json": {},
7364
7360
  "Github_medium---o63998.json": {},
@@ -18,6 +18,11 @@ if [ "$1" == "--bench" ] ; then
18
18
  done
19
19
  fi
20
20
 
21
+ if [ "$1" == "--exp" ] ; then
22
+ shift
23
+ DEFAULT_ARGS="--expected expected_maskbench.json"
24
+ fi
25
+
21
26
  if [ -z "$PERF" ]; then
22
27
  cargo build --release
23
28
  ../target/release/json_stats $DEFAULT_ARGS "$@"
@@ -5,11 +5,7 @@ use json_stats::SchemaStats;
5
5
  use jsonschema::Validator;
6
6
  use llguidance::{
7
7
  api::{GrammarInit, StopReason, TopLevelGrammar},
8
- earley::{
9
- perf::{num_with_commas, ParserPerfCounters},
10
- regexvec::LexerStats,
11
- XorShift,
12
- },
8
+ earley::{perf::num_with_commas, regexvec::LexerStats, XorShift},
13
9
  toktrie::{InferenceCapabilities, SimpleVob, TokEnv},
14
10
  Constraint, HashMap, JsonCompileOptions, ParserFactory, TokenParser,
15
11
  };
@@ -60,6 +56,10 @@ pub struct CliOptions {
60
56
  #[arg(long)]
61
57
  llg_no_forcing: bool,
62
58
 
59
+ /// Set stderr log level; implies --num-threads 1
60
+ #[arg(long, default_value = "0")]
61
+ llg_log_level: u32,
62
+
63
63
  /// Test the slicer optimization against un-sliced parser
64
64
  #[arg(long)]
65
65
  llg_test_slicer: bool,
@@ -80,6 +80,10 @@ pub struct CliOptions {
80
80
  #[arg(long)]
81
81
  csv: bool,
82
82
 
83
+ /// Don't print JSON output and perf counters
84
+ #[arg(long)]
85
+ quiet: bool,
86
+
83
87
  /// Test rollback mechanism for speculative decoding
84
88
  #[arg(long)]
85
89
  rollback: bool,
@@ -282,7 +286,6 @@ struct TestEnv {
282
286
  factory: Arc<ParserFactory>,
283
287
  ref_factory: Arc<ParserFactory>,
284
288
  file_name: String,
285
- perf_counters: Arc<ParserPerfCounters>,
286
289
  hash_rnd: Arc<ahash::RandomState>,
287
290
  }
288
291
 
@@ -702,8 +705,10 @@ impl TestEnv {
702
705
  Ok(schema) => schema,
703
706
  Err(e) => {
704
707
  res.json_error = Some(format!("{e}"));
708
+ if self.cli.llg_log_level > 0 {
709
+ eprintln!("{} Error JSON: {}", self.file_name, e);
710
+ }
705
711
  limit_string(&mut res.json_error);
706
- // eprintln!("{} Error Compile: {}", file, e);
707
712
  return res;
708
713
  }
709
714
  };
@@ -729,8 +734,7 @@ impl TestEnv {
729
734
 
730
735
  let t2 = std::time::Instant::now();
731
736
  let parser = match parser {
732
- Ok(mut parser) => {
733
- parser.parser.set_perf_counters(self.perf_counters.clone());
737
+ Ok(parser) => {
734
738
  let mut constraint = Constraint::new(parser.clone());
735
739
  constraint.compute_mask().unwrap();
736
740
  res.first_mask_us = t2.elapsed().as_micros() as usize;
@@ -743,6 +747,9 @@ impl TestEnv {
743
747
  Err(e) => {
744
748
  // eprintln!("{} Error Parser: {}", self.file_name, e);
745
749
  res.parser_error = Some(format!("{e}"));
750
+ if self.cli.llg_log_level > 0 {
751
+ eprintln!("{} Error JSON: {}", self.file_name, e);
752
+ }
746
753
  limit_string(&mut res.parser_error);
747
754
  return res;
748
755
  }
@@ -757,6 +764,9 @@ impl TestEnv {
757
764
  if let Err(e) = self.run_llg_test(&mut res, &parser, ref_parser.as_ref(), t) {
758
765
  if res.validation_error.is_none() {
759
766
  res.validation_error = Some(format!("test #{idx}: {e}"));
767
+ if self.cli.llg_log_level > 0 {
768
+ eprintln!("{} Error Validating: {}", self.file_name, e);
769
+ }
760
770
  limit_string(&mut res.validation_error);
761
771
  }
762
772
  } else if t.valid {
@@ -905,6 +915,9 @@ fn main() {
905
915
  if options.llg_validate_tokens {
906
916
  options.llg_compile = true;
907
917
  }
918
+ if options.llg_log_level > 0 {
919
+ options.num_threads = Some(1);
920
+ }
908
921
 
909
922
  // set max thread numbers
910
923
  let num_cores = std::thread::available_parallelism().unwrap().get();
@@ -961,8 +974,9 @@ fn main() {
961
974
  };
962
975
 
963
976
  let mut factory = ParserFactory::new(&tok_env, caps.clone(), &slices).unwrap();
964
- factory.quiet();
965
- // factory.set_stderr_log_level(2);
977
+ factory.set_buffer_log_level(0);
978
+ factory.set_stderr_log_level(options.llg_log_level);
979
+
966
980
  // factory.limits_mut().step_lexer_fuel = 10_000_000;
967
981
 
968
982
  let mut ref_factory = ParserFactory::new(&tok_env, caps.clone(), &[]).unwrap();
@@ -976,7 +990,6 @@ fn main() {
976
990
 
977
991
  let t0 = std::time::Instant::now();
978
992
  let par = num_threads > 1;
979
- let perf_counters = Arc::new(ParserPerfCounters::new());
980
993
  let hash_rnd = Arc::new(ahash::RandomState::new());
981
994
  let do_file = |file: &String| {
982
995
  let env = TestEnv {
@@ -985,7 +998,6 @@ fn main() {
985
998
  ref_factory: ref_factory.clone(),
986
999
  file_name: file.to_string(),
987
1000
  cli: options.clone(),
988
- perf_counters: perf_counters.clone(),
989
1001
  hash_rnd: hash_rnd.clone(),
990
1002
  };
991
1003
  env.run_test()
@@ -1131,10 +1143,15 @@ fn main() {
1131
1143
  total.llg.mask_ms_total_a /= 1000;
1132
1144
 
1133
1145
  total.llg_json = llg_totals.clone();
1134
- eprintln!("{}", serde_json::to_string_pretty(&total).unwrap());
1146
+ if !options.quiet {
1147
+ eprintln!(
1148
+ "{}\n{}",
1149
+ serde_json::to_string_pretty(&total).unwrap(),
1150
+ &factory.perf_counters(),
1151
+ );
1152
+ }
1135
1153
  eprintln!(
1136
- "{}Total time: {}ms TTFM {}μs, mask {}μs, ff {}μs, mask+ff {}ms + compile {}ms",
1137
- perf_counters,
1154
+ "Total time: {}ms TTFM {}μs, mask {}μs, ff {}μs, mask+ff {}ms + compile {}ms",
1138
1155
  t0.elapsed().as_millis(),
1139
1156
  total.llg.ttfm_us,
1140
1157
  total.llg.mask_us,
@@ -1,6 +1,6 @@
1
1
  [package]
2
2
  name = "llguidance"
3
- version = "0.7.11"
3
+ version = "0.7.13"
4
4
  edition = "2021"
5
5
  license = "MIT"
6
6
  description = "Super-fast Structured Outputs"
@@ -8,7 +8,7 @@ repository = "https://github.com/guidance-ai/llguidance"
8
8
 
9
9
  [dependencies]
10
10
  toktrie = { workspace = true }
11
- derivre = { version = "=0.3.1", default-features = false, features = ["compress"] }
11
+ derivre = { version = "=0.3.5", default-features = false, features = ["compress"] }
12
12
  serde = { version = "1.0.217", features = ["derive"] }
13
13
  serde_json = { version = "1.0.138", features = ["preserve_order"] }
14
14
  anyhow = "1.0.95"
@@ -70,6 +70,14 @@ typedef struct LlgParserLimits {
70
70
  * Default: 500_000 (a few megabytes of JSON)
71
71
  */
72
72
  size_t max_grammar_size;
73
+ /**
74
+ * If true, we'll run any extremely large regexes against the whole
75
+ * trie of the tokenizer while constructing the lexer.
76
+ * This reduces future mask computation time, but increases
77
+ * the time it takes to construct the lexer.
78
+ * Default: true
79
+ */
80
+ bool precompute_large_lexemes;
73
81
  } LlgParserLimits;
74
82
 
75
83
  typedef struct LlgConstraintInit {
@@ -219,6 +227,12 @@ typedef struct LlgTokenizerInit {
219
227
  * User data to pass to the tokenize_fn
220
228
  */
221
229
  const void *tokenize_user_data;
230
+ /**
231
+ * Tokenizer partitions for the slicer optimization.
232
+ * This is array of pointers to strings, terminated with NULL (argv style).
233
+ * Pass NULL to use defaults. Pass empty array to disable.
234
+ */
235
+ const char *const *slices;
222
236
  } LlgTokenizerInit;
223
237
 
224
238
 
@@ -42,6 +42,7 @@ pub struct LLGuidanceOptions {
42
42
 
43
43
  /// If set, the grammar will allow invalid utf8 byte sequences.
44
44
  /// Any Unicode regex will cause an error.
45
+ /// This is very unlikely what you need.
45
46
  #[serde(default)]
46
47
  pub allow_invalid_utf8: bool,
47
48
  }
@@ -248,6 +249,13 @@ pub struct ParserLimits {
248
249
  /// Maximum size of the grammar (symbols in productions)
249
250
  /// Default: 500_000 (a few megabytes of JSON)
250
251
  pub max_grammar_size: usize,
252
+
253
+ /// If true, we'll run any extremely large regexes against the whole
254
+ /// trie of the tokenizer while constructing the lexer.
255
+ /// This reduces future mask computation time, but increases
256
+ /// the time it takes to construct the lexer.
257
+ /// Default: true
258
+ pub precompute_large_lexemes: bool,
251
259
  }
252
260
 
253
261
  impl Default for ParserLimits {
@@ -259,6 +267,7 @@ impl Default for ParserLimits {
259
267
  max_lexer_states: 250_000, //
260
268
  max_grammar_size: 500_000, // fhir schema => 200k
261
269
  step_max_items: 50_000, //
270
+ precompute_large_lexemes: true,
262
271
  }
263
272
  }
264
273
  }
@@ -13,78 +13,77 @@ use crate::{GrammarBuilder, HashMap};
13
13
  use anyhow::{bail, ensure, Result};
14
14
  use toktrie::TokEnv;
15
15
 
16
- fn process_grammar(ctx: &mut CompileCtx, input: GrammarWithLexer) -> Result<(SymIdx, LexemeClass)> {
17
- let builder = std::mem::take(&mut ctx.builder).unwrap();
18
-
19
- let res = if let Some(lark) = input.lark_grammar {
20
- ensure!(
21
- input.json_schema.is_none(),
22
- "cannot have both lark_grammar and json_schema"
23
- );
24
- lark_to_llguidance(builder, &lark)?
25
- } else if let Some(mut json_schema) = input.json_schema {
26
- let mut opts = JsonCompileOptions::default();
27
- if let Some(x_guidance) = json_schema.get("x-guidance") {
28
- opts = serde_json::from_value(x_guidance.clone())?;
29
- // TODO not removing it causes oneOf to be handled as anyOf in Github_medium---o61004.json
30
- json_schema.as_object_mut().unwrap().remove("x-guidance");
31
- }
32
- opts.json_to_llg(builder, json_schema)?
33
- } else {
34
- bail!("grammar must have either lark_grammar or json_schema");
35
- };
16
+ struct CompileCtx {
17
+ builder: Option<GrammarBuilder>,
18
+ grammar_by_idx: HashMap<GrammarId, usize>,
19
+ grammar_roots: Vec<(SymIdx, LexemeClass)>,
20
+ }
36
21
 
37
- res.builder.check_limits()?;
22
+ impl CompileCtx {
23
+ fn run_one(&mut self, input: GrammarWithLexer) -> Result<(SymIdx, LexemeClass)> {
24
+ let builder = std::mem::take(&mut self.builder).unwrap();
25
+
26
+ let res = if let Some(lark) = input.lark_grammar {
27
+ ensure!(
28
+ input.json_schema.is_none(),
29
+ "cannot have both lark_grammar and json_schema"
30
+ );
31
+ lark_to_llguidance(builder, &lark)?
32
+ } else if let Some(mut json_schema) = input.json_schema {
33
+ let mut opts = JsonCompileOptions::default();
34
+ if let Some(x_guidance) = json_schema.get("x-guidance") {
35
+ opts = serde_json::from_value(x_guidance.clone())?;
36
+ // TODO not removing it causes oneOf to be handled as anyOf in Github_medium---o61004.json
37
+ json_schema.as_object_mut().unwrap().remove("x-guidance");
38
+ }
39
+ opts.json_to_llg(builder, json_schema)?
40
+ } else {
41
+ bail!("grammar must have either lark_grammar or json_schema");
42
+ };
38
43
 
39
- let grammar_id = res.builder.grammar.sym_props(res.start_node).grammar_id;
44
+ res.builder.check_limits()?;
40
45
 
41
- // restore builder
42
- ctx.builder = Some(res.builder);
46
+ let grammar_id = res.builder.grammar.sym_props(res.start_node).grammar_id;
43
47
 
44
- Ok((res.start_node, grammar_id))
45
- }
48
+ // restore builder
49
+ self.builder = Some(res.builder);
46
50
 
47
- fn process_all_grammars(
48
- mut ctx: CompileCtx,
49
- input: TopLevelGrammar,
50
- ) -> Result<(Grammar, LexerSpec)> {
51
- for (idx, grm) in input.grammars.iter().enumerate() {
52
- if grm.lark_grammar.is_none() && grm.json_schema.is_none() {
53
- bail!("grammar must have either lark_grammar or json_schema");
54
- }
55
- if let Some(n) = &grm.name {
56
- let n = GrammarId::Name(n.to_string());
57
- if ctx.grammar_by_idx.contains_key(&n) {
58
- bail!("duplicate grammar name: {}", n);
59
- }
60
- ctx.grammar_by_idx.insert(n, idx);
61
- }
51
+ Ok((res.start_node, grammar_id))
62
52
  }
63
53
 
64
- for (idx, grm) in input.grammars.into_iter().enumerate() {
65
- let v = process_grammar(&mut ctx, grm)?;
66
- ctx.grammar_roots[idx] = v;
67
- }
54
+ fn run(mut self, input: TopLevelGrammar) -> Result<(Grammar, LexerSpec)> {
55
+ for (idx, grm) in input.grammars.iter().enumerate() {
56
+ if grm.lark_grammar.is_none() && grm.json_schema.is_none() {
57
+ bail!("grammar must have either lark_grammar or json_schema");
58
+ }
59
+ if let Some(n) = &grm.name {
60
+ let n = GrammarId::Name(n.to_string());
61
+ if self.grammar_by_idx.contains_key(&n) {
62
+ bail!("duplicate grammar name: {}", n);
63
+ }
64
+ self.grammar_by_idx.insert(n, idx);
65
+ }
66
+ }
68
67
 
69
- let grammar_by_idx: HashMap<GrammarId, (SymIdx, LexemeClass)> = ctx
70
- .grammar_by_idx
71
- .into_iter()
72
- .map(|(k, v)| (k, ctx.grammar_roots[v]))
73
- .collect();
68
+ for (idx, grm) in input.grammars.into_iter().enumerate() {
69
+ let v = self.run_one(grm)?;
70
+ self.grammar_roots[idx] = v;
71
+ }
74
72
 
75
- let builder = ctx.builder.unwrap();
76
- let mut grammar = builder.grammar;
77
- let mut lexer_spec = builder.regex.spec;
73
+ let grammar_by_idx: HashMap<GrammarId, (SymIdx, LexemeClass)> = self
74
+ .grammar_by_idx
75
+ .into_iter()
76
+ .map(|(k, v)| (k, self.grammar_roots[v]))
77
+ .collect();
78
78
 
79
- grammar.resolve_grammar_refs(&mut lexer_spec, &grammar_by_idx)?;
79
+ let builder = self.builder.unwrap();
80
+ let mut grammar = builder.grammar;
81
+ let mut lexer_spec = builder.regex.spec;
80
82
 
81
- Ok((grammar, lexer_spec))
82
- }
83
+ grammar.resolve_grammar_refs(&mut lexer_spec, &grammar_by_idx)?;
83
84
 
84
- struct CompileCtx {
85
- builder: Option<GrammarBuilder>,
86
- grammar_by_idx: HashMap<GrammarId, usize>,
87
- grammar_roots: Vec<(SymIdx, LexemeClass)>,
85
+ Ok((grammar, lexer_spec))
86
+ }
88
87
  }
89
88
 
90
89
  impl GrammarInit {
@@ -107,7 +106,7 @@ impl GrammarInit {
107
106
  grammar_roots: vec![(SymIdx::BOGUS, LexemeClass::ROOT); input.grammars.len()],
108
107
  };
109
108
 
110
- process_all_grammars(ctx, input)
109
+ ctx.run(input)
111
110
  }
112
111
  }
113
112
  }