aiqtoolkit 1.1.0rc6__py3-none-any.whl → 1.2.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of aiqtoolkit might be problematic. Click here for more details.

Files changed (319) hide show
  1. aiqtoolkit-1.2.0.dist-info/METADATA +29 -0
  2. aiqtoolkit-1.2.0.dist-info/RECORD +4 -0
  3. {aiqtoolkit-1.1.0rc6.dist-info → aiqtoolkit-1.2.0.dist-info}/WHEEL +1 -1
  4. aiqtoolkit-1.2.0.dist-info/top_level.txt +1 -0
  5. aiq/agent/__init__.py +0 -0
  6. aiq/agent/base.py +0 -76
  7. aiq/agent/dual_node.py +0 -67
  8. aiq/agent/react_agent/__init__.py +0 -0
  9. aiq/agent/react_agent/agent.py +0 -322
  10. aiq/agent/react_agent/output_parser.py +0 -104
  11. aiq/agent/react_agent/prompt.py +0 -46
  12. aiq/agent/react_agent/register.py +0 -148
  13. aiq/agent/reasoning_agent/__init__.py +0 -0
  14. aiq/agent/reasoning_agent/reasoning_agent.py +0 -224
  15. aiq/agent/register.py +0 -23
  16. aiq/agent/rewoo_agent/__init__.py +0 -0
  17. aiq/agent/rewoo_agent/agent.py +0 -410
  18. aiq/agent/rewoo_agent/prompt.py +0 -108
  19. aiq/agent/rewoo_agent/register.py +0 -158
  20. aiq/agent/tool_calling_agent/__init__.py +0 -0
  21. aiq/agent/tool_calling_agent/agent.py +0 -123
  22. aiq/agent/tool_calling_agent/register.py +0 -105
  23. aiq/builder/__init__.py +0 -0
  24. aiq/builder/builder.py +0 -223
  25. aiq/builder/component_utils.py +0 -303
  26. aiq/builder/context.py +0 -227
  27. aiq/builder/embedder.py +0 -24
  28. aiq/builder/eval_builder.py +0 -120
  29. aiq/builder/evaluator.py +0 -29
  30. aiq/builder/framework_enum.py +0 -24
  31. aiq/builder/front_end.py +0 -73
  32. aiq/builder/function.py +0 -297
  33. aiq/builder/function_base.py +0 -376
  34. aiq/builder/function_info.py +0 -627
  35. aiq/builder/intermediate_step_manager.py +0 -176
  36. aiq/builder/llm.py +0 -25
  37. aiq/builder/retriever.py +0 -25
  38. aiq/builder/user_interaction_manager.py +0 -71
  39. aiq/builder/workflow.py +0 -143
  40. aiq/builder/workflow_builder.py +0 -757
  41. aiq/cli/__init__.py +0 -14
  42. aiq/cli/cli_utils/__init__.py +0 -0
  43. aiq/cli/cli_utils/config_override.py +0 -231
  44. aiq/cli/cli_utils/validation.py +0 -37
  45. aiq/cli/commands/__init__.py +0 -0
  46. aiq/cli/commands/configure/__init__.py +0 -0
  47. aiq/cli/commands/configure/channel/__init__.py +0 -0
  48. aiq/cli/commands/configure/channel/add.py +0 -28
  49. aiq/cli/commands/configure/channel/channel.py +0 -36
  50. aiq/cli/commands/configure/channel/remove.py +0 -30
  51. aiq/cli/commands/configure/channel/update.py +0 -30
  52. aiq/cli/commands/configure/configure.py +0 -33
  53. aiq/cli/commands/evaluate.py +0 -139
  54. aiq/cli/commands/info/__init__.py +0 -14
  55. aiq/cli/commands/info/info.py +0 -39
  56. aiq/cli/commands/info/list_channels.py +0 -32
  57. aiq/cli/commands/info/list_components.py +0 -129
  58. aiq/cli/commands/info/list_mcp.py +0 -126
  59. aiq/cli/commands/registry/__init__.py +0 -14
  60. aiq/cli/commands/registry/publish.py +0 -88
  61. aiq/cli/commands/registry/pull.py +0 -118
  62. aiq/cli/commands/registry/registry.py +0 -38
  63. aiq/cli/commands/registry/remove.py +0 -108
  64. aiq/cli/commands/registry/search.py +0 -155
  65. aiq/cli/commands/start.py +0 -250
  66. aiq/cli/commands/uninstall.py +0 -83
  67. aiq/cli/commands/validate.py +0 -47
  68. aiq/cli/commands/workflow/__init__.py +0 -14
  69. aiq/cli/commands/workflow/templates/__init__.py.j2 +0 -0
  70. aiq/cli/commands/workflow/templates/config.yml.j2 +0 -16
  71. aiq/cli/commands/workflow/templates/pyproject.toml.j2 +0 -22
  72. aiq/cli/commands/workflow/templates/register.py.j2 +0 -5
  73. aiq/cli/commands/workflow/templates/workflow.py.j2 +0 -36
  74. aiq/cli/commands/workflow/workflow.py +0 -37
  75. aiq/cli/commands/workflow/workflow_commands.py +0 -313
  76. aiq/cli/entrypoint.py +0 -133
  77. aiq/cli/main.py +0 -44
  78. aiq/cli/register_workflow.py +0 -408
  79. aiq/cli/type_registry.py +0 -879
  80. aiq/data_models/__init__.py +0 -14
  81. aiq/data_models/api_server.py +0 -588
  82. aiq/data_models/common.py +0 -143
  83. aiq/data_models/component.py +0 -46
  84. aiq/data_models/component_ref.py +0 -135
  85. aiq/data_models/config.py +0 -349
  86. aiq/data_models/dataset_handler.py +0 -122
  87. aiq/data_models/discovery_metadata.py +0 -286
  88. aiq/data_models/embedder.py +0 -26
  89. aiq/data_models/evaluate.py +0 -104
  90. aiq/data_models/evaluator.py +0 -26
  91. aiq/data_models/front_end.py +0 -26
  92. aiq/data_models/function.py +0 -30
  93. aiq/data_models/function_dependencies.py +0 -64
  94. aiq/data_models/interactive.py +0 -237
  95. aiq/data_models/intermediate_step.py +0 -269
  96. aiq/data_models/invocation_node.py +0 -38
  97. aiq/data_models/llm.py +0 -26
  98. aiq/data_models/logging.py +0 -26
  99. aiq/data_models/memory.py +0 -26
  100. aiq/data_models/profiler.py +0 -53
  101. aiq/data_models/registry_handler.py +0 -26
  102. aiq/data_models/retriever.py +0 -30
  103. aiq/data_models/step_adaptor.py +0 -64
  104. aiq/data_models/streaming.py +0 -33
  105. aiq/data_models/swe_bench_model.py +0 -54
  106. aiq/data_models/telemetry_exporter.py +0 -26
  107. aiq/embedder/__init__.py +0 -0
  108. aiq/embedder/langchain_client.py +0 -41
  109. aiq/embedder/nim_embedder.py +0 -58
  110. aiq/embedder/openai_embedder.py +0 -42
  111. aiq/embedder/register.py +0 -24
  112. aiq/eval/__init__.py +0 -14
  113. aiq/eval/config.py +0 -42
  114. aiq/eval/dataset_handler/__init__.py +0 -0
  115. aiq/eval/dataset_handler/dataset_downloader.py +0 -106
  116. aiq/eval/dataset_handler/dataset_filter.py +0 -52
  117. aiq/eval/dataset_handler/dataset_handler.py +0 -169
  118. aiq/eval/evaluate.py +0 -325
  119. aiq/eval/evaluator/__init__.py +0 -14
  120. aiq/eval/evaluator/evaluator_model.py +0 -44
  121. aiq/eval/intermediate_step_adapter.py +0 -93
  122. aiq/eval/rag_evaluator/__init__.py +0 -0
  123. aiq/eval/rag_evaluator/evaluate.py +0 -138
  124. aiq/eval/rag_evaluator/register.py +0 -138
  125. aiq/eval/register.py +0 -23
  126. aiq/eval/remote_workflow.py +0 -128
  127. aiq/eval/runtime_event_subscriber.py +0 -52
  128. aiq/eval/swe_bench_evaluator/__init__.py +0 -0
  129. aiq/eval/swe_bench_evaluator/evaluate.py +0 -215
  130. aiq/eval/swe_bench_evaluator/register.py +0 -36
  131. aiq/eval/trajectory_evaluator/__init__.py +0 -0
  132. aiq/eval/trajectory_evaluator/evaluate.py +0 -118
  133. aiq/eval/trajectory_evaluator/register.py +0 -40
  134. aiq/eval/tunable_rag_evaluator/__init__.py +0 -0
  135. aiq/eval/tunable_rag_evaluator/evaluate.py +0 -263
  136. aiq/eval/tunable_rag_evaluator/register.py +0 -50
  137. aiq/eval/utils/__init__.py +0 -0
  138. aiq/eval/utils/output_uploader.py +0 -131
  139. aiq/eval/utils/tqdm_position_registry.py +0 -40
  140. aiq/front_ends/__init__.py +0 -14
  141. aiq/front_ends/console/__init__.py +0 -14
  142. aiq/front_ends/console/console_front_end_config.py +0 -32
  143. aiq/front_ends/console/console_front_end_plugin.py +0 -107
  144. aiq/front_ends/console/register.py +0 -25
  145. aiq/front_ends/cron/__init__.py +0 -14
  146. aiq/front_ends/fastapi/__init__.py +0 -14
  147. aiq/front_ends/fastapi/fastapi_front_end_config.py +0 -150
  148. aiq/front_ends/fastapi/fastapi_front_end_plugin.py +0 -103
  149. aiq/front_ends/fastapi/fastapi_front_end_plugin_worker.py +0 -607
  150. aiq/front_ends/fastapi/intermediate_steps_subscriber.py +0 -80
  151. aiq/front_ends/fastapi/job_store.py +0 -161
  152. aiq/front_ends/fastapi/main.py +0 -70
  153. aiq/front_ends/fastapi/message_handler.py +0 -279
  154. aiq/front_ends/fastapi/message_validator.py +0 -345
  155. aiq/front_ends/fastapi/register.py +0 -25
  156. aiq/front_ends/fastapi/response_helpers.py +0 -195
  157. aiq/front_ends/fastapi/step_adaptor.py +0 -320
  158. aiq/front_ends/fastapi/websocket.py +0 -148
  159. aiq/front_ends/mcp/__init__.py +0 -14
  160. aiq/front_ends/mcp/mcp_front_end_config.py +0 -32
  161. aiq/front_ends/mcp/mcp_front_end_plugin.py +0 -93
  162. aiq/front_ends/mcp/register.py +0 -27
  163. aiq/front_ends/mcp/tool_converter.py +0 -242
  164. aiq/front_ends/register.py +0 -22
  165. aiq/front_ends/simple_base/__init__.py +0 -14
  166. aiq/front_ends/simple_base/simple_front_end_plugin_base.py +0 -52
  167. aiq/llm/__init__.py +0 -0
  168. aiq/llm/nim_llm.py +0 -45
  169. aiq/llm/openai_llm.py +0 -45
  170. aiq/llm/register.py +0 -22
  171. aiq/llm/utils/__init__.py +0 -14
  172. aiq/llm/utils/env_config_value.py +0 -94
  173. aiq/llm/utils/error.py +0 -17
  174. aiq/memory/__init__.py +0 -20
  175. aiq/memory/interfaces.py +0 -183
  176. aiq/memory/models.py +0 -112
  177. aiq/meta/module_to_distro.json +0 -3
  178. aiq/meta/pypi.md +0 -58
  179. aiq/observability/__init__.py +0 -0
  180. aiq/observability/async_otel_listener.py +0 -429
  181. aiq/observability/register.py +0 -99
  182. aiq/plugins/.namespace +0 -1
  183. aiq/profiler/__init__.py +0 -0
  184. aiq/profiler/callbacks/__init__.py +0 -0
  185. aiq/profiler/callbacks/agno_callback_handler.py +0 -295
  186. aiq/profiler/callbacks/base_callback_class.py +0 -20
  187. aiq/profiler/callbacks/langchain_callback_handler.py +0 -278
  188. aiq/profiler/callbacks/llama_index_callback_handler.py +0 -205
  189. aiq/profiler/callbacks/semantic_kernel_callback_handler.py +0 -238
  190. aiq/profiler/callbacks/token_usage_base_model.py +0 -27
  191. aiq/profiler/data_frame_row.py +0 -51
  192. aiq/profiler/decorators/__init__.py +0 -0
  193. aiq/profiler/decorators/framework_wrapper.py +0 -131
  194. aiq/profiler/decorators/function_tracking.py +0 -254
  195. aiq/profiler/forecasting/__init__.py +0 -0
  196. aiq/profiler/forecasting/config.py +0 -18
  197. aiq/profiler/forecasting/model_trainer.py +0 -75
  198. aiq/profiler/forecasting/models/__init__.py +0 -22
  199. aiq/profiler/forecasting/models/forecasting_base_model.py +0 -40
  200. aiq/profiler/forecasting/models/linear_model.py +0 -196
  201. aiq/profiler/forecasting/models/random_forest_regressor.py +0 -268
  202. aiq/profiler/inference_metrics_model.py +0 -25
  203. aiq/profiler/inference_optimization/__init__.py +0 -0
  204. aiq/profiler/inference_optimization/bottleneck_analysis/__init__.py +0 -0
  205. aiq/profiler/inference_optimization/bottleneck_analysis/nested_stack_analysis.py +0 -452
  206. aiq/profiler/inference_optimization/bottleneck_analysis/simple_stack_analysis.py +0 -258
  207. aiq/profiler/inference_optimization/data_models.py +0 -386
  208. aiq/profiler/inference_optimization/experimental/__init__.py +0 -0
  209. aiq/profiler/inference_optimization/experimental/concurrency_spike_analysis.py +0 -468
  210. aiq/profiler/inference_optimization/experimental/prefix_span_analysis.py +0 -405
  211. aiq/profiler/inference_optimization/llm_metrics.py +0 -212
  212. aiq/profiler/inference_optimization/prompt_caching.py +0 -163
  213. aiq/profiler/inference_optimization/token_uniqueness.py +0 -107
  214. aiq/profiler/inference_optimization/workflow_runtimes.py +0 -72
  215. aiq/profiler/intermediate_property_adapter.py +0 -102
  216. aiq/profiler/profile_runner.py +0 -433
  217. aiq/profiler/utils.py +0 -184
  218. aiq/registry_handlers/__init__.py +0 -0
  219. aiq/registry_handlers/local/__init__.py +0 -0
  220. aiq/registry_handlers/local/local_handler.py +0 -176
  221. aiq/registry_handlers/local/register_local.py +0 -37
  222. aiq/registry_handlers/metadata_factory.py +0 -60
  223. aiq/registry_handlers/package_utils.py +0 -198
  224. aiq/registry_handlers/pypi/__init__.py +0 -0
  225. aiq/registry_handlers/pypi/pypi_handler.py +0 -251
  226. aiq/registry_handlers/pypi/register_pypi.py +0 -40
  227. aiq/registry_handlers/register.py +0 -21
  228. aiq/registry_handlers/registry_handler_base.py +0 -157
  229. aiq/registry_handlers/rest/__init__.py +0 -0
  230. aiq/registry_handlers/rest/register_rest.py +0 -56
  231. aiq/registry_handlers/rest/rest_handler.py +0 -237
  232. aiq/registry_handlers/schemas/__init__.py +0 -0
  233. aiq/registry_handlers/schemas/headers.py +0 -42
  234. aiq/registry_handlers/schemas/package.py +0 -68
  235. aiq/registry_handlers/schemas/publish.py +0 -63
  236. aiq/registry_handlers/schemas/pull.py +0 -82
  237. aiq/registry_handlers/schemas/remove.py +0 -36
  238. aiq/registry_handlers/schemas/search.py +0 -91
  239. aiq/registry_handlers/schemas/status.py +0 -47
  240. aiq/retriever/__init__.py +0 -0
  241. aiq/retriever/interface.py +0 -37
  242. aiq/retriever/milvus/__init__.py +0 -14
  243. aiq/retriever/milvus/register.py +0 -81
  244. aiq/retriever/milvus/retriever.py +0 -228
  245. aiq/retriever/models.py +0 -74
  246. aiq/retriever/nemo_retriever/__init__.py +0 -14
  247. aiq/retriever/nemo_retriever/register.py +0 -60
  248. aiq/retriever/nemo_retriever/retriever.py +0 -190
  249. aiq/retriever/register.py +0 -22
  250. aiq/runtime/__init__.py +0 -14
  251. aiq/runtime/loader.py +0 -188
  252. aiq/runtime/runner.py +0 -176
  253. aiq/runtime/session.py +0 -140
  254. aiq/runtime/user_metadata.py +0 -131
  255. aiq/settings/__init__.py +0 -0
  256. aiq/settings/global_settings.py +0 -318
  257. aiq/test/.namespace +0 -1
  258. aiq/tool/__init__.py +0 -0
  259. aiq/tool/code_execution/__init__.py +0 -0
  260. aiq/tool/code_execution/code_sandbox.py +0 -188
  261. aiq/tool/code_execution/local_sandbox/Dockerfile.sandbox +0 -60
  262. aiq/tool/code_execution/local_sandbox/__init__.py +0 -13
  263. aiq/tool/code_execution/local_sandbox/local_sandbox_server.py +0 -83
  264. aiq/tool/code_execution/local_sandbox/sandbox.requirements.txt +0 -4
  265. aiq/tool/code_execution/local_sandbox/start_local_sandbox.sh +0 -25
  266. aiq/tool/code_execution/register.py +0 -70
  267. aiq/tool/code_execution/utils.py +0 -100
  268. aiq/tool/datetime_tools.py +0 -42
  269. aiq/tool/document_search.py +0 -141
  270. aiq/tool/github_tools/__init__.py +0 -0
  271. aiq/tool/github_tools/create_github_commit.py +0 -133
  272. aiq/tool/github_tools/create_github_issue.py +0 -87
  273. aiq/tool/github_tools/create_github_pr.py +0 -106
  274. aiq/tool/github_tools/get_github_file.py +0 -106
  275. aiq/tool/github_tools/get_github_issue.py +0 -166
  276. aiq/tool/github_tools/get_github_pr.py +0 -256
  277. aiq/tool/github_tools/update_github_issue.py +0 -100
  278. aiq/tool/mcp/__init__.py +0 -14
  279. aiq/tool/mcp/mcp_client.py +0 -220
  280. aiq/tool/mcp/mcp_tool.py +0 -95
  281. aiq/tool/memory_tools/__init__.py +0 -0
  282. aiq/tool/memory_tools/add_memory_tool.py +0 -79
  283. aiq/tool/memory_tools/delete_memory_tool.py +0 -67
  284. aiq/tool/memory_tools/get_memory_tool.py +0 -72
  285. aiq/tool/nvidia_rag.py +0 -95
  286. aiq/tool/register.py +0 -37
  287. aiq/tool/retriever.py +0 -89
  288. aiq/tool/server_tools.py +0 -63
  289. aiq/utils/__init__.py +0 -0
  290. aiq/utils/data_models/__init__.py +0 -0
  291. aiq/utils/data_models/schema_validator.py +0 -58
  292. aiq/utils/debugging_utils.py +0 -43
  293. aiq/utils/exception_handlers/__init__.py +0 -0
  294. aiq/utils/exception_handlers/schemas.py +0 -114
  295. aiq/utils/io/__init__.py +0 -0
  296. aiq/utils/io/yaml_tools.py +0 -119
  297. aiq/utils/metadata_utils.py +0 -74
  298. aiq/utils/optional_imports.py +0 -142
  299. aiq/utils/producer_consumer_queue.py +0 -178
  300. aiq/utils/reactive/__init__.py +0 -0
  301. aiq/utils/reactive/base/__init__.py +0 -0
  302. aiq/utils/reactive/base/observable_base.py +0 -65
  303. aiq/utils/reactive/base/observer_base.py +0 -55
  304. aiq/utils/reactive/base/subject_base.py +0 -79
  305. aiq/utils/reactive/observable.py +0 -59
  306. aiq/utils/reactive/observer.py +0 -76
  307. aiq/utils/reactive/subject.py +0 -131
  308. aiq/utils/reactive/subscription.py +0 -49
  309. aiq/utils/settings/__init__.py +0 -0
  310. aiq/utils/settings/global_settings.py +0 -197
  311. aiq/utils/type_converter.py +0 -232
  312. aiq/utils/type_utils.py +0 -397
  313. aiq/utils/url_utils.py +0 -27
  314. aiqtoolkit-1.1.0rc6.dist-info/METADATA +0 -331
  315. aiqtoolkit-1.1.0rc6.dist-info/RECORD +0 -316
  316. aiqtoolkit-1.1.0rc6.dist-info/entry_points.txt +0 -17
  317. aiqtoolkit-1.1.0rc6.dist-info/licenses/LICENSE-3rd-party.txt +0 -3686
  318. aiqtoolkit-1.1.0rc6.dist-info/licenses/LICENSE.md +0 -201
  319. aiqtoolkit-1.1.0rc6.dist-info/top_level.txt +0 -1
@@ -1,163 +0,0 @@
1
- # SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
2
- # SPDX-License-Identifier: Apache-2.0
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
-
16
- from aiq.data_models.intermediate_step import IntermediateStep
17
- from aiq.profiler.inference_optimization.data_models import CommonPrefixesOutput
18
- from aiq.profiler.inference_optimization.data_models import FrameworkLLMPrefixData
19
- from aiq.profiler.inference_optimization.data_models import PrefixInfo
20
- from aiq.profiler.utils import create_standardized_dataframe
21
-
22
-
23
- # -----------------------------------------------------------
24
- # 1. Helper: Build a prefix trie
25
- # -----------------------------------------------------------
26
- def build_prefix_trie(strings: list[str]) -> dict:
27
- """
28
- Build a trie from a list of strings.
29
-
30
- Returns a nested dictionary with::
31
-
32
- {
33
- 'count': int, # number of strings passing through this node
34
- 'children': dict[str, TrieNode]
35
- }
36
-
37
- """
38
- root = {'count': 0, 'children': {}}
39
- for s in strings:
40
- node = root
41
- node['count'] += 1 # every string passes through the root
42
- for ch in s:
43
- if ch not in node['children']:
44
- node['children'][ch] = {'count': 0, 'children': {}}
45
- node = node['children'][ch]
46
- node['count'] += 1
47
- return root
48
-
49
-
50
- # -----------------------------------------------------------
51
- # 2. Helper: Iterative traversal of the trie
52
- # -----------------------------------------------------------
53
- def collect_prefixes_iterative(root: dict, total_calls: int) -> list[dict]:
54
- """
55
- Iteratively traverse the trie to collect prefix statistics,
56
- avoiding recursion depth limits.
57
-
58
- :param root: Trie node with 'count' and 'children'
59
- :param total_calls: Number of total calls in this group (denominator for percentages)
60
- :return: A list of dicts, each dict containing prefix info
61
- """
62
- results = []
63
- # stack holds (node, prefix_so_far)
64
- stack = [(root, "")]
65
-
66
- while stack:
67
- node, prefix = stack.pop()
68
-
69
- # Skip storing the empty root prefix
70
- if prefix:
71
- calls_count = node['count']
72
- calls_percentage = calls_count / total_calls
73
- results.append({
74
- 'prefix': prefix,
75
- 'prefix_length': len(prefix),
76
- 'calls_count': calls_count,
77
- 'calls_percentage': calls_percentage
78
- })
79
-
80
- # Add children to the stack
81
- for ch, child_node in node['children'].items():
82
- stack.append((child_node, prefix + ch))
83
-
84
- return results
85
-
86
-
87
- # -----------------------------------------------------------
88
- # 3. Main Function
89
- # -----------------------------------------------------------
90
- def get_common_prefixes(all_steps: list[list[IntermediateStep]],
91
- min_call_percentage: float = 0.0) -> CommonPrefixesOutput:
92
- """
93
- Given a pandas DataFrame with columns 'framework', 'llm_name',
94
- and 'llm_text_input', return a Pydantic-validated RootModel
95
- keyed by "<llm_name>" with a sorted list of
96
- common prefix statistics.
97
-
98
- 1) Only includes prefixes with calls_percentage >= `min_call_percentage`.
99
- 2) Excludes any prefix that is a substring of another (longer) prefix
100
- that already meets the threshold and is retained.
101
- 3) Optionally writes the resulting dictionary to JSON if `output_path` is provided.
102
-
103
- :param all_steps: Intermediate Steps
104
- :param min_call_percentage: Exclude prefixes that appear in fewer than this fraction
105
- of total calls. (Default 0.0 = no filtering)
106
-
107
- Sorting: primarily by prefix length (descending),
108
- secondarily by frequency (descending).
109
- """
110
- # Validate necessary columns
111
- df = create_standardized_dataframe(all_steps)
112
-
113
- required_cols = {'framework', 'llm_name', 'llm_text_input'}
114
- if not required_cols.issubset(df.columns):
115
- missing = required_cols - set(df.columns)
116
- raise ValueError(f"DataFrame missing required columns: {missing}")
117
-
118
- output_data: dict[str, FrameworkLLMPrefixData] = {}
119
-
120
- # Group DataFrame by (framework, llm_name)
121
- grouped = df.groupby(['llm_name'])
122
- for llm_name, group_df in grouped:
123
- # Unpack llm_name Tuple
124
- llm_name = llm_name[0]
125
-
126
- text_inputs = group_df['llm_text_input'].astype(str).tolist()
127
- total_calls = len(text_inputs)
128
-
129
- # Build trie for all text inputs
130
- trie = build_prefix_trie(text_inputs)
131
-
132
- # Collect prefix info using iterative traversal
133
- results = collect_prefixes_iterative(trie, total_calls=total_calls)
134
-
135
- # 1) Filter out prefixes below min_call_percentage
136
- results_filtered = [r for r in results if r['calls_percentage'] >= min_call_percentage]
137
-
138
- # 2) Sort results: prefix_length desc, then calls_count desc
139
- results_sorted = sorted(results_filtered, key=lambda x: (x['prefix_length'], x['calls_count']), reverse=True)
140
-
141
- # 3) Substring filtering:
142
- # Because results_sorted is in descending length order,
143
- # if we keep a prefix, we exclude any shorter prefix that
144
- # is a substring of that already-kept prefix.
145
- final_results = []
146
- for r in results_sorted:
147
- pfx = r['prefix']
148
- # Check if this prefix is contained in any longer prefix we have kept
149
- if not any(pfx in kept['prefix'] for kept in final_results):
150
- final_results.append(r)
151
-
152
- # Convert each dict to a PrefixInfo model
153
- prefix_info_list = [PrefixInfo(**res) for res in final_results]
154
-
155
- # Construct the dictionary key
156
- framework_llm_key = f"{llm_name}"
157
-
158
- # Save the data for this group
159
- output_data[framework_llm_key] = FrameworkLLMPrefixData(total_calls=total_calls, prefix_info=prefix_info_list)
160
-
161
- # Package the final result in a validated RootModel
162
- result_model = CommonPrefixesOutput(root=output_data)
163
- return result_model
@@ -1,107 +0,0 @@
1
- # SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
2
- # SPDX-License-Identifier: Apache-2.0
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
-
16
- import re
17
-
18
- import numpy as np
19
-
20
- from aiq.data_models.intermediate_step import IntermediateStep
21
- from aiq.profiler.inference_optimization.data_models import LLMUniquenessMetrics
22
- from aiq.profiler.inference_optimization.data_models import LLMUniquenessMetricsByLLM
23
- from aiq.profiler.utils import create_standardized_dataframe
24
-
25
-
26
- # ----------------------------------------------------------------
27
- # 1. Main Function
28
- # ----------------------------------------------------------------
29
- def compute_inter_query_token_uniqueness_by_llm(all_steps: list[list[IntermediateStep]]) -> LLMUniquenessMetricsByLLM:
30
- """
31
- Computes p90, p95, and p99 of 'new words added' between consecutive llm_start events,
32
- grouped by (llm_name, example_number).
33
-
34
- Steps:
35
-
36
- 1. Filter df to only llm_start events.
37
- 2. Group first by (llm_name, example_number), then sort by event_timestamp in each group.
38
- 3. Compare each llm_text_input to the previous one in the same group to find how many new words appear.
39
- 4. Aggregate all 'new words count' across each llm_name, compute p90/p95/p99 for each LLM.
40
- 5. Return a Pydantic RootModel containing a dictionary::
41
-
42
- { llm_name -> LLMUniquenessMetrics(p90, p95, p99) }.
43
- """
44
- df = create_standardized_dataframe(all_steps)
45
- # Validate that the necessary columns exist
46
- required_cols = {'event_type', 'llm_name', 'example_number', 'event_timestamp', 'llm_text_input'}
47
- missing = required_cols - set(df.columns)
48
- if missing:
49
- raise ValueError(f"DataFrame missing required columns: {missing}")
50
-
51
- # 1) Filter to llm_start events
52
- cdf = df[df['event_type'] == 'LLM_START'].copy()
53
- if cdf.empty:
54
- # Return an empty dictionary if no llm_start events
55
- return LLMUniquenessMetricsByLLM(root={})
56
-
57
- # Helper to tokenize text into a set of words
58
- def tokenize_to_set(text: str) -> set:
59
- if not isinstance(text, str):
60
- return set()
61
- return set(re.findall(r"\w+", text.lower()))
62
-
63
- # We'll store new_words counts for each llm_name
64
- llm_to_counts: dict[str, list[int]] = {}
65
-
66
- # 2) Group by (llm_name, example_number), then sort each group
67
- grouped = cdf.groupby(['llm_name', 'example_number'], as_index=False, group_keys=True)
68
-
69
- for (llm, ex_num), group_df in grouped: # pylint: disable=unused-variable
70
- # Sort by event_timestamp
71
- group_df = group_df.sort_values('event_timestamp', ascending=True)
72
-
73
- # Shift the llm_text_input to compare consecutive calls
74
- group_df['prev_llm_text_input'] = group_df['llm_text_input'].shift(1)
75
-
76
- # Compute new words for each row (excluding the first in the group)
77
- def compute_new_words(row):
78
- current_tokens = tokenize_to_set(row['llm_text_input'])
79
- prev_tokens = tokenize_to_set(row['prev_llm_text_input'])
80
- return len(current_tokens - prev_tokens)
81
-
82
- group_df['new_words_count'] = group_df.apply(compute_new_words, axis=1)
83
-
84
- # Drop rows where there's no 'previous' call
85
- valid_rows = group_df.dropna(subset=['prev_llm_text_input'])
86
-
87
- # Gather the new_words_count
88
- counts = valid_rows['new_words_count'].tolist()
89
- if counts:
90
- # Accumulate them in llm_to_counts
91
- if llm not in llm_to_counts:
92
- llm_to_counts[llm] = []
93
- llm_to_counts[llm].extend(counts)
94
-
95
- # 4) For each llm_name, compute p90, p95, p99
96
- output_dict = {}
97
- for llm_name, counts_list in llm_to_counts.items():
98
- arr = np.array(counts_list)
99
- p90_val = float(np.percentile(arr, 90))
100
- p95_val = float(np.percentile(arr, 95))
101
- p99_val = float(np.percentile(arr, 99))
102
-
103
- output_dict[llm_name] = LLMUniquenessMetrics(p90=p90_val, p95=p95_val, p99=p99_val)
104
-
105
- ret_val = LLMUniquenessMetricsByLLM(root=output_dict)
106
- # Validate & return as a RootModel
107
- return ret_val
@@ -1,72 +0,0 @@
1
- # SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
2
- # SPDX-License-Identifier: Apache-2.0
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
-
16
- import numpy as np
17
-
18
- from aiq.data_models.intermediate_step import IntermediateStep
19
- from aiq.profiler.inference_optimization.data_models import WorkflowRuntimeMetrics
20
- from aiq.profiler.utils import create_standardized_dataframe
21
-
22
-
23
- def compute_workflow_runtime_metrics(all_steps: list[list[IntermediateStep]]) -> WorkflowRuntimeMetrics:
24
- """
25
- Computes the p90, p95, and p99 of workflow runtime for each example_number.
26
-
27
- The 'workflow runtime' per example is::
28
-
29
- max(event_timestamp) - min(event_timestamp)
30
-
31
- for that example_number.
32
-
33
- Parameters
34
- ----------
35
- all_steps : IntermediateStep
36
- Must contain at least two columns:
37
- - 'example_number'
38
- - 'event_timestamp'
39
-
40
- Returns
41
- -------
42
- WorkflowRuntimeMetrics
43
- A Pydantic model with 'p90', 'p95', and 'p99' attributes.
44
- """
45
- df = create_standardized_dataframe(all_steps)
46
- required_cols = {"example_number", "event_timestamp"}
47
- missing = required_cols - set(df.columns)
48
- if missing:
49
- raise ValueError(f"DataFrame is missing required columns: {missing}")
50
-
51
- # Group by example_number, then find min and max timestamp
52
- grouped = df.groupby("example_number")["event_timestamp"]
53
- min_timestamps = grouped.min()
54
- max_timestamps = grouped.max()
55
-
56
- # Workflow runtime is difference between max and min
57
- runtimes = max_timestamps - min_timestamps
58
-
59
- # Convert to a NumPy array for percentile calculations
60
- runtimes_arr = runtimes.values
61
-
62
- # Edge case: if there's only one example or no data
63
- # (NumPy percentile can handle 1-element arrays, but let's guard for empties)
64
- if len(runtimes_arr) == 0:
65
- return WorkflowRuntimeMetrics(p90=0.0, p95=0.0, p99=0.0)
66
-
67
- # Compute p90, p95, p99
68
- p90_val = float(np.percentile(runtimes_arr, 90))
69
- p95_val = float(np.percentile(runtimes_arr, 95))
70
- p99_val = float(np.percentile(runtimes_arr, 99))
71
-
72
- return WorkflowRuntimeMetrics(p90=p90_val, p95=p95_val, p99=p99_val)
@@ -1,102 +0,0 @@
1
- # SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
2
- # SPDX-License-Identifier: Apache-2.0
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
-
16
- from aiq.data_models.intermediate_step import IntermediateStep
17
- from aiq.data_models.intermediate_step import IntermediateStepType
18
- from aiq.data_models.intermediate_step import TokenUsageBaseModel
19
-
20
-
21
- class IntermediatePropertyAdaptor(IntermediateStep):
22
-
23
- @classmethod
24
- def from_intermediate_step(cls, step: IntermediateStep) -> "IntermediatePropertyAdaptor":
25
- """
26
- Create an adaptor instance from an existing IntermediateStep.
27
- Uses the dict() representation of the instance to initialize the adaptor.
28
- """
29
- return cls(**step.model_dump())
30
-
31
- @property
32
- def token_usage(self) -> TokenUsageBaseModel:
33
- return self.payload.usage_info.token_usage if self.payload.usage_info else TokenUsageBaseModel()
34
-
35
- @property
36
- def seconds_between_calls(self) -> int:
37
- return self.payload.usage_info.seconds_between_calls if self.payload.usage_info else 0
38
-
39
- @property
40
- def llm_text_input(self) -> str:
41
- ret = ""
42
- if self.payload.data and self.event_type == IntermediateStepType.LLM_START:
43
- ret = self.payload.data.input
44
- return ret
45
-
46
- @property
47
- def llm_text_output(self) -> str:
48
- ret = ""
49
- if self.payload.data and self.event_type == IntermediateStepType.LLM_END:
50
- ret = self.payload.data.output
51
- return ret
52
-
53
- @property
54
- def llm_text_chunk(self) -> str:
55
- ret = ""
56
- if self.payload.data and self.event_type == IntermediateStepType.LLM_NEW_TOKEN:
57
- ret = self.payload.data.chunk
58
- return ret
59
-
60
- @property
61
- def tool_input(self) -> str:
62
- ret = ""
63
- if self.payload.data and self.event_type == IntermediateStepType.TOOL_START:
64
- ret = self.payload.data.input
65
- return ret
66
-
67
- @property
68
- def tool_output(self) -> str:
69
- ret = ""
70
- if self.payload.data and self.event_type == IntermediateStepType.TOOL_END:
71
- ret = self.payload.data.output
72
- return ret
73
-
74
- @property
75
- def llm_name(self) -> str:
76
- ret = ""
77
- if self.payload.name and self.event_type in [IntermediateStepType.LLM_START, IntermediateStepType.LLM_END]:
78
- ret = self.payload.name
79
- return ret
80
-
81
- @property
82
- def tool_name(self) -> str:
83
- ret = ""
84
- if self.payload.name and self.event_type in [IntermediateStepType.TOOL_START, IntermediateStepType.TOOL_END]:
85
- ret = self.payload.name
86
- return ret
87
-
88
- @property
89
- def function_name(self) -> str:
90
- return self.function_ancestry.function_name
91
-
92
- @property
93
- def function_id(self) -> str:
94
- return self.function_ancestry.function_id
95
-
96
- @property
97
- def parent_function_id(self) -> str:
98
- return self.function_ancestry.parent_id
99
-
100
- @property
101
- def parent_function_name(self) -> str:
102
- return self.function_ancestry.parent_name