nvidia-nat 1.2.0rc5__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- aiq/agent/__init__.py +0 -0
- aiq/agent/base.py +239 -0
- aiq/agent/dual_node.py +67 -0
- aiq/agent/react_agent/__init__.py +0 -0
- aiq/agent/react_agent/agent.py +355 -0
- aiq/agent/react_agent/output_parser.py +104 -0
- aiq/agent/react_agent/prompt.py +41 -0
- aiq/agent/react_agent/register.py +149 -0
- aiq/agent/reasoning_agent/__init__.py +0 -0
- aiq/agent/reasoning_agent/reasoning_agent.py +225 -0
- aiq/agent/register.py +23 -0
- aiq/agent/rewoo_agent/__init__.py +0 -0
- aiq/agent/rewoo_agent/agent.py +411 -0
- aiq/agent/rewoo_agent/prompt.py +108 -0
- aiq/agent/rewoo_agent/register.py +158 -0
- aiq/agent/tool_calling_agent/__init__.py +0 -0
- aiq/agent/tool_calling_agent/agent.py +119 -0
- aiq/agent/tool_calling_agent/register.py +106 -0
- aiq/authentication/__init__.py +14 -0
- aiq/authentication/api_key/__init__.py +14 -0
- aiq/authentication/api_key/api_key_auth_provider.py +96 -0
- aiq/authentication/api_key/api_key_auth_provider_config.py +124 -0
- aiq/authentication/api_key/register.py +26 -0
- aiq/authentication/exceptions/__init__.py +14 -0
- aiq/authentication/exceptions/api_key_exceptions.py +38 -0
- aiq/authentication/http_basic_auth/__init__.py +0 -0
- aiq/authentication/http_basic_auth/http_basic_auth_provider.py +81 -0
- aiq/authentication/http_basic_auth/register.py +30 -0
- aiq/authentication/interfaces.py +93 -0
- aiq/authentication/oauth2/__init__.py +14 -0
- aiq/authentication/oauth2/oauth2_auth_code_flow_provider.py +107 -0
- aiq/authentication/oauth2/oauth2_auth_code_flow_provider_config.py +39 -0
- aiq/authentication/oauth2/register.py +25 -0
- aiq/authentication/register.py +21 -0
- aiq/builder/__init__.py +0 -0
- aiq/builder/builder.py +285 -0
- aiq/builder/component_utils.py +316 -0
- aiq/builder/context.py +264 -0
- aiq/builder/embedder.py +24 -0
- aiq/builder/eval_builder.py +161 -0
- aiq/builder/evaluator.py +29 -0
- aiq/builder/framework_enum.py +24 -0
- aiq/builder/front_end.py +73 -0
- aiq/builder/function.py +344 -0
- aiq/builder/function_base.py +380 -0
- aiq/builder/function_info.py +627 -0
- aiq/builder/intermediate_step_manager.py +174 -0
- aiq/builder/llm.py +25 -0
- aiq/builder/retriever.py +25 -0
- aiq/builder/user_interaction_manager.py +74 -0
- aiq/builder/workflow.py +148 -0
- aiq/builder/workflow_builder.py +1117 -0
- aiq/cli/__init__.py +14 -0
- aiq/cli/cli_utils/__init__.py +0 -0
- aiq/cli/cli_utils/config_override.py +231 -0
- aiq/cli/cli_utils/validation.py +37 -0
- aiq/cli/commands/__init__.py +0 -0
- aiq/cli/commands/configure/__init__.py +0 -0
- aiq/cli/commands/configure/channel/__init__.py +0 -0
- aiq/cli/commands/configure/channel/add.py +28 -0
- aiq/cli/commands/configure/channel/channel.py +36 -0
- aiq/cli/commands/configure/channel/remove.py +30 -0
- aiq/cli/commands/configure/channel/update.py +30 -0
- aiq/cli/commands/configure/configure.py +33 -0
- aiq/cli/commands/evaluate.py +139 -0
- aiq/cli/commands/info/__init__.py +14 -0
- aiq/cli/commands/info/info.py +39 -0
- aiq/cli/commands/info/list_channels.py +32 -0
- aiq/cli/commands/info/list_components.py +129 -0
- aiq/cli/commands/info/list_mcp.py +213 -0
- aiq/cli/commands/registry/__init__.py +14 -0
- aiq/cli/commands/registry/publish.py +88 -0
- aiq/cli/commands/registry/pull.py +118 -0
- aiq/cli/commands/registry/registry.py +38 -0
- aiq/cli/commands/registry/remove.py +108 -0
- aiq/cli/commands/registry/search.py +155 -0
- aiq/cli/commands/sizing/__init__.py +14 -0
- aiq/cli/commands/sizing/calc.py +297 -0
- aiq/cli/commands/sizing/sizing.py +27 -0
- aiq/cli/commands/start.py +246 -0
- aiq/cli/commands/uninstall.py +81 -0
- aiq/cli/commands/validate.py +47 -0
- aiq/cli/commands/workflow/__init__.py +14 -0
- aiq/cli/commands/workflow/templates/__init__.py.j2 +0 -0
- aiq/cli/commands/workflow/templates/config.yml.j2 +16 -0
- aiq/cli/commands/workflow/templates/pyproject.toml.j2 +22 -0
- aiq/cli/commands/workflow/templates/register.py.j2 +5 -0
- aiq/cli/commands/workflow/templates/workflow.py.j2 +36 -0
- aiq/cli/commands/workflow/workflow.py +37 -0
- aiq/cli/commands/workflow/workflow_commands.py +313 -0
- aiq/cli/entrypoint.py +135 -0
- aiq/cli/main.py +44 -0
- aiq/cli/register_workflow.py +488 -0
- aiq/cli/type_registry.py +1000 -0
- aiq/data_models/__init__.py +14 -0
- aiq/data_models/api_server.py +694 -0
- aiq/data_models/authentication.py +231 -0
- aiq/data_models/common.py +171 -0
- aiq/data_models/component.py +54 -0
- aiq/data_models/component_ref.py +168 -0
- aiq/data_models/config.py +406 -0
- aiq/data_models/dataset_handler.py +123 -0
- aiq/data_models/discovery_metadata.py +335 -0
- aiq/data_models/embedder.py +27 -0
- aiq/data_models/evaluate.py +127 -0
- aiq/data_models/evaluator.py +26 -0
- aiq/data_models/front_end.py +26 -0
- aiq/data_models/function.py +30 -0
- aiq/data_models/function_dependencies.py +72 -0
- aiq/data_models/interactive.py +246 -0
- aiq/data_models/intermediate_step.py +302 -0
- aiq/data_models/invocation_node.py +38 -0
- aiq/data_models/llm.py +27 -0
- aiq/data_models/logging.py +26 -0
- aiq/data_models/memory.py +27 -0
- aiq/data_models/object_store.py +44 -0
- aiq/data_models/profiler.py +54 -0
- aiq/data_models/registry_handler.py +26 -0
- aiq/data_models/retriever.py +30 -0
- aiq/data_models/retry_mixin.py +35 -0
- aiq/data_models/span.py +187 -0
- aiq/data_models/step_adaptor.py +64 -0
- aiq/data_models/streaming.py +33 -0
- aiq/data_models/swe_bench_model.py +54 -0
- aiq/data_models/telemetry_exporter.py +26 -0
- aiq/data_models/ttc_strategy.py +30 -0
- aiq/embedder/__init__.py +0 -0
- aiq/embedder/langchain_client.py +41 -0
- aiq/embedder/nim_embedder.py +59 -0
- aiq/embedder/openai_embedder.py +43 -0
- aiq/embedder/register.py +24 -0
- aiq/eval/__init__.py +14 -0
- aiq/eval/config.py +60 -0
- aiq/eval/dataset_handler/__init__.py +0 -0
- aiq/eval/dataset_handler/dataset_downloader.py +106 -0
- aiq/eval/dataset_handler/dataset_filter.py +52 -0
- aiq/eval/dataset_handler/dataset_handler.py +254 -0
- aiq/eval/evaluate.py +506 -0
- aiq/eval/evaluator/__init__.py +14 -0
- aiq/eval/evaluator/base_evaluator.py +73 -0
- aiq/eval/evaluator/evaluator_model.py +45 -0
- aiq/eval/intermediate_step_adapter.py +99 -0
- aiq/eval/rag_evaluator/__init__.py +0 -0
- aiq/eval/rag_evaluator/evaluate.py +178 -0
- aiq/eval/rag_evaluator/register.py +143 -0
- aiq/eval/register.py +23 -0
- aiq/eval/remote_workflow.py +133 -0
- aiq/eval/runners/__init__.py +14 -0
- aiq/eval/runners/config.py +39 -0
- aiq/eval/runners/multi_eval_runner.py +54 -0
- aiq/eval/runtime_event_subscriber.py +52 -0
- aiq/eval/swe_bench_evaluator/__init__.py +0 -0
- aiq/eval/swe_bench_evaluator/evaluate.py +215 -0
- aiq/eval/swe_bench_evaluator/register.py +36 -0
- aiq/eval/trajectory_evaluator/__init__.py +0 -0
- aiq/eval/trajectory_evaluator/evaluate.py +75 -0
- aiq/eval/trajectory_evaluator/register.py +40 -0
- aiq/eval/tunable_rag_evaluator/__init__.py +0 -0
- aiq/eval/tunable_rag_evaluator/evaluate.py +245 -0
- aiq/eval/tunable_rag_evaluator/register.py +52 -0
- aiq/eval/usage_stats.py +41 -0
- aiq/eval/utils/__init__.py +0 -0
- aiq/eval/utils/output_uploader.py +140 -0
- aiq/eval/utils/tqdm_position_registry.py +40 -0
- aiq/eval/utils/weave_eval.py +184 -0
- aiq/experimental/__init__.py +0 -0
- aiq/experimental/decorators/__init__.py +0 -0
- aiq/experimental/decorators/experimental_warning_decorator.py +130 -0
- aiq/experimental/test_time_compute/__init__.py +0 -0
- aiq/experimental/test_time_compute/editing/__init__.py +0 -0
- aiq/experimental/test_time_compute/editing/iterative_plan_refinement_editor.py +147 -0
- aiq/experimental/test_time_compute/editing/llm_as_a_judge_editor.py +204 -0
- aiq/experimental/test_time_compute/editing/motivation_aware_summarization.py +107 -0
- aiq/experimental/test_time_compute/functions/__init__.py +0 -0
- aiq/experimental/test_time_compute/functions/execute_score_select_function.py +105 -0
- aiq/experimental/test_time_compute/functions/its_tool_orchestration_function.py +205 -0
- aiq/experimental/test_time_compute/functions/its_tool_wrapper_function.py +146 -0
- aiq/experimental/test_time_compute/functions/plan_select_execute_function.py +224 -0
- aiq/experimental/test_time_compute/models/__init__.py +0 -0
- aiq/experimental/test_time_compute/models/editor_config.py +132 -0
- aiq/experimental/test_time_compute/models/scoring_config.py +112 -0
- aiq/experimental/test_time_compute/models/search_config.py +120 -0
- aiq/experimental/test_time_compute/models/selection_config.py +154 -0
- aiq/experimental/test_time_compute/models/stage_enums.py +43 -0
- aiq/experimental/test_time_compute/models/strategy_base.py +66 -0
- aiq/experimental/test_time_compute/models/tool_use_config.py +41 -0
- aiq/experimental/test_time_compute/models/ttc_item.py +48 -0
- aiq/experimental/test_time_compute/register.py +36 -0
- aiq/experimental/test_time_compute/scoring/__init__.py +0 -0
- aiq/experimental/test_time_compute/scoring/llm_based_agent_scorer.py +168 -0
- aiq/experimental/test_time_compute/scoring/llm_based_plan_scorer.py +168 -0
- aiq/experimental/test_time_compute/scoring/motivation_aware_scorer.py +111 -0
- aiq/experimental/test_time_compute/search/__init__.py +0 -0
- aiq/experimental/test_time_compute/search/multi_llm_planner.py +128 -0
- aiq/experimental/test_time_compute/search/multi_query_retrieval_search.py +122 -0
- aiq/experimental/test_time_compute/search/single_shot_multi_plan_planner.py +128 -0
- aiq/experimental/test_time_compute/selection/__init__.py +0 -0
- aiq/experimental/test_time_compute/selection/best_of_n_selector.py +63 -0
- aiq/experimental/test_time_compute/selection/llm_based_agent_output_selector.py +131 -0
- aiq/experimental/test_time_compute/selection/llm_based_output_merging_selector.py +159 -0
- aiq/experimental/test_time_compute/selection/llm_based_plan_selector.py +128 -0
- aiq/experimental/test_time_compute/selection/threshold_selector.py +58 -0
- aiq/front_ends/__init__.py +14 -0
- aiq/front_ends/console/__init__.py +14 -0
- aiq/front_ends/console/authentication_flow_handler.py +233 -0
- aiq/front_ends/console/console_front_end_config.py +32 -0
- aiq/front_ends/console/console_front_end_plugin.py +96 -0
- aiq/front_ends/console/register.py +25 -0
- aiq/front_ends/cron/__init__.py +14 -0
- aiq/front_ends/fastapi/__init__.py +14 -0
- aiq/front_ends/fastapi/auth_flow_handlers/__init__.py +0 -0
- aiq/front_ends/fastapi/auth_flow_handlers/http_flow_handler.py +27 -0
- aiq/front_ends/fastapi/auth_flow_handlers/websocket_flow_handler.py +107 -0
- aiq/front_ends/fastapi/fastapi_front_end_config.py +234 -0
- aiq/front_ends/fastapi/fastapi_front_end_controller.py +68 -0
- aiq/front_ends/fastapi/fastapi_front_end_plugin.py +116 -0
- aiq/front_ends/fastapi/fastapi_front_end_plugin_worker.py +1092 -0
- aiq/front_ends/fastapi/html_snippets/__init__.py +14 -0
- aiq/front_ends/fastapi/html_snippets/auth_code_grant_success.py +35 -0
- aiq/front_ends/fastapi/intermediate_steps_subscriber.py +80 -0
- aiq/front_ends/fastapi/job_store.py +183 -0
- aiq/front_ends/fastapi/main.py +72 -0
- aiq/front_ends/fastapi/message_handler.py +298 -0
- aiq/front_ends/fastapi/message_validator.py +345 -0
- aiq/front_ends/fastapi/register.py +25 -0
- aiq/front_ends/fastapi/response_helpers.py +195 -0
- aiq/front_ends/fastapi/step_adaptor.py +321 -0
- aiq/front_ends/mcp/__init__.py +14 -0
- aiq/front_ends/mcp/mcp_front_end_config.py +32 -0
- aiq/front_ends/mcp/mcp_front_end_plugin.py +93 -0
- aiq/front_ends/mcp/register.py +27 -0
- aiq/front_ends/mcp/tool_converter.py +242 -0
- aiq/front_ends/register.py +22 -0
- aiq/front_ends/simple_base/__init__.py +14 -0
- aiq/front_ends/simple_base/simple_front_end_plugin_base.py +54 -0
- aiq/llm/__init__.py +0 -0
- aiq/llm/aws_bedrock_llm.py +57 -0
- aiq/llm/nim_llm.py +46 -0
- aiq/llm/openai_llm.py +46 -0
- aiq/llm/register.py +23 -0
- aiq/llm/utils/__init__.py +14 -0
- aiq/llm/utils/env_config_value.py +94 -0
- aiq/llm/utils/error.py +17 -0
- aiq/memory/__init__.py +20 -0
- aiq/memory/interfaces.py +183 -0
- aiq/memory/models.py +112 -0
- aiq/meta/module_to_distro.json +3 -0
- aiq/meta/pypi.md +58 -0
- aiq/object_store/__init__.py +20 -0
- aiq/object_store/in_memory_object_store.py +76 -0
- aiq/object_store/interfaces.py +84 -0
- aiq/object_store/models.py +36 -0
- aiq/object_store/register.py +20 -0
- aiq/observability/__init__.py +14 -0
- aiq/observability/exporter/__init__.py +14 -0
- aiq/observability/exporter/base_exporter.py +449 -0
- aiq/observability/exporter/exporter.py +78 -0
- aiq/observability/exporter/file_exporter.py +33 -0
- aiq/observability/exporter/processing_exporter.py +322 -0
- aiq/observability/exporter/raw_exporter.py +52 -0
- aiq/observability/exporter/span_exporter.py +265 -0
- aiq/observability/exporter_manager.py +335 -0
- aiq/observability/mixin/__init__.py +14 -0
- aiq/observability/mixin/batch_config_mixin.py +26 -0
- aiq/observability/mixin/collector_config_mixin.py +23 -0
- aiq/observability/mixin/file_mixin.py +288 -0
- aiq/observability/mixin/file_mode.py +23 -0
- aiq/observability/mixin/resource_conflict_mixin.py +134 -0
- aiq/observability/mixin/serialize_mixin.py +61 -0
- aiq/observability/mixin/type_introspection_mixin.py +183 -0
- aiq/observability/processor/__init__.py +14 -0
- aiq/observability/processor/batching_processor.py +310 -0
- aiq/observability/processor/callback_processor.py +42 -0
- aiq/observability/processor/intermediate_step_serializer.py +28 -0
- aiq/observability/processor/processor.py +71 -0
- aiq/observability/register.py +96 -0
- aiq/observability/utils/__init__.py +14 -0
- aiq/observability/utils/dict_utils.py +236 -0
- aiq/observability/utils/time_utils.py +31 -0
- aiq/plugins/.namespace +1 -0
- aiq/profiler/__init__.py +0 -0
- aiq/profiler/calc/__init__.py +14 -0
- aiq/profiler/calc/calc_runner.py +627 -0
- aiq/profiler/calc/calculations.py +288 -0
- aiq/profiler/calc/data_models.py +188 -0
- aiq/profiler/calc/plot.py +345 -0
- aiq/profiler/callbacks/__init__.py +0 -0
- aiq/profiler/callbacks/agno_callback_handler.py +295 -0
- aiq/profiler/callbacks/base_callback_class.py +20 -0
- aiq/profiler/callbacks/langchain_callback_handler.py +290 -0
- aiq/profiler/callbacks/llama_index_callback_handler.py +205 -0
- aiq/profiler/callbacks/semantic_kernel_callback_handler.py +238 -0
- aiq/profiler/callbacks/token_usage_base_model.py +27 -0
- aiq/profiler/data_frame_row.py +51 -0
- aiq/profiler/data_models.py +24 -0
- aiq/profiler/decorators/__init__.py +0 -0
- aiq/profiler/decorators/framework_wrapper.py +131 -0
- aiq/profiler/decorators/function_tracking.py +254 -0
- aiq/profiler/forecasting/__init__.py +0 -0
- aiq/profiler/forecasting/config.py +18 -0
- aiq/profiler/forecasting/model_trainer.py +75 -0
- aiq/profiler/forecasting/models/__init__.py +22 -0
- aiq/profiler/forecasting/models/forecasting_base_model.py +40 -0
- aiq/profiler/forecasting/models/linear_model.py +196 -0
- aiq/profiler/forecasting/models/random_forest_regressor.py +268 -0
- aiq/profiler/inference_metrics_model.py +28 -0
- aiq/profiler/inference_optimization/__init__.py +0 -0
- aiq/profiler/inference_optimization/bottleneck_analysis/__init__.py +0 -0
- aiq/profiler/inference_optimization/bottleneck_analysis/nested_stack_analysis.py +460 -0
- aiq/profiler/inference_optimization/bottleneck_analysis/simple_stack_analysis.py +258 -0
- aiq/profiler/inference_optimization/data_models.py +386 -0
- aiq/profiler/inference_optimization/experimental/__init__.py +0 -0
- aiq/profiler/inference_optimization/experimental/concurrency_spike_analysis.py +468 -0
- aiq/profiler/inference_optimization/experimental/prefix_span_analysis.py +405 -0
- aiq/profiler/inference_optimization/llm_metrics.py +212 -0
- aiq/profiler/inference_optimization/prompt_caching.py +163 -0
- aiq/profiler/inference_optimization/token_uniqueness.py +107 -0
- aiq/profiler/inference_optimization/workflow_runtimes.py +72 -0
- aiq/profiler/intermediate_property_adapter.py +102 -0
- aiq/profiler/profile_runner.py +473 -0
- aiq/profiler/utils.py +184 -0
- aiq/registry_handlers/__init__.py +0 -0
- aiq/registry_handlers/local/__init__.py +0 -0
- aiq/registry_handlers/local/local_handler.py +176 -0
- aiq/registry_handlers/local/register_local.py +37 -0
- aiq/registry_handlers/metadata_factory.py +60 -0
- aiq/registry_handlers/package_utils.py +567 -0
- aiq/registry_handlers/pypi/__init__.py +0 -0
- aiq/registry_handlers/pypi/pypi_handler.py +251 -0
- aiq/registry_handlers/pypi/register_pypi.py +40 -0
- aiq/registry_handlers/register.py +21 -0
- aiq/registry_handlers/registry_handler_base.py +157 -0
- aiq/registry_handlers/rest/__init__.py +0 -0
- aiq/registry_handlers/rest/register_rest.py +56 -0
- aiq/registry_handlers/rest/rest_handler.py +237 -0
- aiq/registry_handlers/schemas/__init__.py +0 -0
- aiq/registry_handlers/schemas/headers.py +42 -0
- aiq/registry_handlers/schemas/package.py +68 -0
- aiq/registry_handlers/schemas/publish.py +63 -0
- aiq/registry_handlers/schemas/pull.py +82 -0
- aiq/registry_handlers/schemas/remove.py +36 -0
- aiq/registry_handlers/schemas/search.py +91 -0
- aiq/registry_handlers/schemas/status.py +47 -0
- aiq/retriever/__init__.py +0 -0
- aiq/retriever/interface.py +37 -0
- aiq/retriever/milvus/__init__.py +14 -0
- aiq/retriever/milvus/register.py +81 -0
- aiq/retriever/milvus/retriever.py +228 -0
- aiq/retriever/models.py +74 -0
- aiq/retriever/nemo_retriever/__init__.py +14 -0
- aiq/retriever/nemo_retriever/register.py +60 -0
- aiq/retriever/nemo_retriever/retriever.py +190 -0
- aiq/retriever/register.py +22 -0
- aiq/runtime/__init__.py +14 -0
- aiq/runtime/loader.py +215 -0
- aiq/runtime/runner.py +190 -0
- aiq/runtime/session.py +158 -0
- aiq/runtime/user_metadata.py +130 -0
- aiq/settings/__init__.py +0 -0
- aiq/settings/global_settings.py +318 -0
- aiq/test/.namespace +1 -0
- aiq/tool/__init__.py +0 -0
- aiq/tool/chat_completion.py +74 -0
- aiq/tool/code_execution/README.md +151 -0
- aiq/tool/code_execution/__init__.py +0 -0
- aiq/tool/code_execution/code_sandbox.py +267 -0
- aiq/tool/code_execution/local_sandbox/.gitignore +1 -0
- aiq/tool/code_execution/local_sandbox/Dockerfile.sandbox +60 -0
- aiq/tool/code_execution/local_sandbox/__init__.py +13 -0
- aiq/tool/code_execution/local_sandbox/local_sandbox_server.py +198 -0
- aiq/tool/code_execution/local_sandbox/sandbox.requirements.txt +6 -0
- aiq/tool/code_execution/local_sandbox/start_local_sandbox.sh +50 -0
- aiq/tool/code_execution/register.py +74 -0
- aiq/tool/code_execution/test_code_execution_sandbox.py +414 -0
- aiq/tool/code_execution/utils.py +100 -0
- aiq/tool/datetime_tools.py +42 -0
- aiq/tool/document_search.py +141 -0
- aiq/tool/github_tools/__init__.py +0 -0
- aiq/tool/github_tools/create_github_commit.py +133 -0
- aiq/tool/github_tools/create_github_issue.py +87 -0
- aiq/tool/github_tools/create_github_pr.py +106 -0
- aiq/tool/github_tools/get_github_file.py +106 -0
- aiq/tool/github_tools/get_github_issue.py +166 -0
- aiq/tool/github_tools/get_github_pr.py +256 -0
- aiq/tool/github_tools/update_github_issue.py +100 -0
- aiq/tool/mcp/__init__.py +14 -0
- aiq/tool/mcp/exceptions.py +142 -0
- aiq/tool/mcp/mcp_client.py +255 -0
- aiq/tool/mcp/mcp_tool.py +96 -0
- aiq/tool/memory_tools/__init__.py +0 -0
- aiq/tool/memory_tools/add_memory_tool.py +79 -0
- aiq/tool/memory_tools/delete_memory_tool.py +67 -0
- aiq/tool/memory_tools/get_memory_tool.py +72 -0
- aiq/tool/nvidia_rag.py +95 -0
- aiq/tool/register.py +38 -0
- aiq/tool/retriever.py +89 -0
- aiq/tool/server_tools.py +66 -0
- aiq/utils/__init__.py +0 -0
- aiq/utils/data_models/__init__.py +0 -0
- aiq/utils/data_models/schema_validator.py +58 -0
- aiq/utils/debugging_utils.py +43 -0
- aiq/utils/dump_distro_mapping.py +32 -0
- aiq/utils/exception_handlers/__init__.py +0 -0
- aiq/utils/exception_handlers/automatic_retries.py +289 -0
- aiq/utils/exception_handlers/mcp.py +211 -0
- aiq/utils/exception_handlers/schemas.py +114 -0
- aiq/utils/io/__init__.py +0 -0
- aiq/utils/io/model_processing.py +28 -0
- aiq/utils/io/yaml_tools.py +119 -0
- aiq/utils/log_utils.py +37 -0
- aiq/utils/metadata_utils.py +74 -0
- aiq/utils/optional_imports.py +142 -0
- aiq/utils/producer_consumer_queue.py +178 -0
- aiq/utils/reactive/__init__.py +0 -0
- aiq/utils/reactive/base/__init__.py +0 -0
- aiq/utils/reactive/base/observable_base.py +65 -0
- aiq/utils/reactive/base/observer_base.py +55 -0
- aiq/utils/reactive/base/subject_base.py +79 -0
- aiq/utils/reactive/observable.py +59 -0
- aiq/utils/reactive/observer.py +76 -0
- aiq/utils/reactive/subject.py +131 -0
- aiq/utils/reactive/subscription.py +49 -0
- aiq/utils/settings/__init__.py +0 -0
- aiq/utils/settings/global_settings.py +197 -0
- aiq/utils/string_utils.py +38 -0
- aiq/utils/type_converter.py +290 -0
- aiq/utils/type_utils.py +484 -0
- aiq/utils/url_utils.py +27 -0
- nvidia_nat-1.2.0rc5.dist-info/METADATA +363 -0
- nvidia_nat-1.2.0rc5.dist-info/RECORD +435 -0
- nvidia_nat-1.2.0rc5.dist-info/WHEEL +5 -0
- nvidia_nat-1.2.0rc5.dist-info/entry_points.txt +20 -0
- nvidia_nat-1.2.0rc5.dist-info/licenses/LICENSE-3rd-party.txt +3686 -0
- nvidia_nat-1.2.0rc5.dist-info/licenses/LICENSE.md +201 -0
- nvidia_nat-1.2.0rc5.dist-info/top_level.txt +1 -0
|
@@ -0,0 +1,460 @@
|
|
|
1
|
+
# SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
|
|
2
|
+
# SPDX-License-Identifier: Apache-2.0
|
|
3
|
+
#
|
|
4
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
5
|
+
# you may not use this file except in compliance with the License.
|
|
6
|
+
# You may obtain a copy of the License at
|
|
7
|
+
#
|
|
8
|
+
# http://www.apache.org/licenses/LICENSE-2.0
|
|
9
|
+
#
|
|
10
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
11
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
12
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
13
|
+
# See the License for the specific language governing permissions and
|
|
14
|
+
# limitations under the License.
|
|
15
|
+
"""
|
|
16
|
+
An enhanced script that:
|
|
17
|
+
|
|
18
|
+
1. Groups events by example_number.
|
|
19
|
+
2. Builds a nested call tree (stack-based) for each example_number, so calls from different examples never nest.
|
|
20
|
+
3. Combines all calls into one global list for concurrency analysis.
|
|
21
|
+
4. Computes:
|
|
22
|
+
|
|
23
|
+
- self_time, subtree_time for each call
|
|
24
|
+
- concurrency distribution (p50, p90, p95, p99) across all examples
|
|
25
|
+
- each node's midpoint concurrency
|
|
26
|
+
- a custom 'bottleneck_score' (here = subtree_time)
|
|
27
|
+
|
|
28
|
+
5. Optionally saves a Gantt chart.
|
|
29
|
+
6. Returns a Pydantic object with concurrency stats, node metrics, top bottlenecks, and a textual report.
|
|
30
|
+
"""
|
|
31
|
+
|
|
32
|
+
import logging
|
|
33
|
+
import os
|
|
34
|
+
|
|
35
|
+
import pandas as pd
|
|
36
|
+
|
|
37
|
+
from aiq.data_models.intermediate_step import IntermediateStep
|
|
38
|
+
from aiq.profiler.inference_optimization.data_models import CallNode
|
|
39
|
+
from aiq.profiler.inference_optimization.data_models import ConcurrencyDistribution
|
|
40
|
+
from aiq.profiler.inference_optimization.data_models import NestedCallProfilingResult
|
|
41
|
+
from aiq.profiler.inference_optimization.data_models import NodeMetrics
|
|
42
|
+
from aiq.profiler.utils import create_standardized_dataframe
|
|
43
|
+
|
|
44
|
+
logger = logging.getLogger(__name__)
|
|
45
|
+
|
|
46
|
+
# --------------------------------------------------------------------------------
|
|
47
|
+
# 1) Build the Nested Call Tree PER EXAMPLE
|
|
48
|
+
# --------------------------------------------------------------------------------
|
|
49
|
+
|
|
50
|
+
|
|
51
|
+
def build_call_tree_for_example(example_df: pd.DataFrame) -> list[CallNode]:
|
|
52
|
+
"""
|
|
53
|
+
Stack-based approach for a single example:
|
|
54
|
+
|
|
55
|
+
1. Sort events by timestamp ascending.
|
|
56
|
+
2. On `*_START` => push a new node, attach to parent's children if stack not empty.
|
|
57
|
+
3. On `*_END` => pop from stack if matches the top's UUID, finalize end_time/duration.
|
|
58
|
+
|
|
59
|
+
Returns:
|
|
60
|
+
A list of top-level calls for this example.
|
|
61
|
+
"""
|
|
62
|
+
stack: list[CallNode] = []
|
|
63
|
+
top_level_dict: dict[str, CallNode] = {}
|
|
64
|
+
partial_map: dict[str, CallNode] = {}
|
|
65
|
+
|
|
66
|
+
def parse_op_type(evt: str) -> str | None:
|
|
67
|
+
evt = evt.upper()
|
|
68
|
+
if evt.startswith("LLM_"):
|
|
69
|
+
return "LLM"
|
|
70
|
+
if evt.startswith("TOOL_"):
|
|
71
|
+
return "TOOL"
|
|
72
|
+
if evt.startswith("FUNCTION_"):
|
|
73
|
+
return "FUNCTION"
|
|
74
|
+
if evt.startswith("SPAN_"):
|
|
75
|
+
return "FUNCTION"
|
|
76
|
+
return None
|
|
77
|
+
|
|
78
|
+
def get_op_name(row: pd.Series, op_type: str) -> str:
|
|
79
|
+
if op_type == "LLM":
|
|
80
|
+
return row.get("llm_name") or "unknown_llm"
|
|
81
|
+
if op_type == "FUNCTION":
|
|
82
|
+
return row.get("function_name") or "unknown_function"
|
|
83
|
+
if op_type == "TOOL":
|
|
84
|
+
return row.get("tool_name") or "unknown_tool"
|
|
85
|
+
|
|
86
|
+
return "unknown_op"
|
|
87
|
+
|
|
88
|
+
for _, row in example_df.iterrows():
|
|
89
|
+
et = row["event_type"].value.upper()
|
|
90
|
+
uuid = str(row["UUID"])
|
|
91
|
+
ts = float(row["event_timestamp"])
|
|
92
|
+
|
|
93
|
+
op_type = parse_op_type(et)
|
|
94
|
+
if not op_type:
|
|
95
|
+
# not an LLM_/TOOL_ event => skip
|
|
96
|
+
continue
|
|
97
|
+
|
|
98
|
+
if et.endswith("_START"):
|
|
99
|
+
name = get_op_name(row, op_type)
|
|
100
|
+
node = CallNode(uuid=uuid,
|
|
101
|
+
operation_type=op_type,
|
|
102
|
+
operation_name=name,
|
|
103
|
+
start_time=ts,
|
|
104
|
+
end_time=ts,
|
|
105
|
+
duration=0.0,
|
|
106
|
+
children=[],
|
|
107
|
+
parent=None)
|
|
108
|
+
if stack:
|
|
109
|
+
parent = stack[-1]
|
|
110
|
+
node.parent = parent
|
|
111
|
+
parent.children.append(node)
|
|
112
|
+
else:
|
|
113
|
+
# top-level
|
|
114
|
+
top_level_dict[uuid] = node
|
|
115
|
+
|
|
116
|
+
stack.append(node)
|
|
117
|
+
partial_map[uuid] = node
|
|
118
|
+
|
|
119
|
+
elif et.endswith("_END"):
|
|
120
|
+
if uuid not in partial_map:
|
|
121
|
+
# no known start => skip
|
|
122
|
+
continue
|
|
123
|
+
node = partial_map[uuid]
|
|
124
|
+
if stack and stack[-1].uuid == uuid:
|
|
125
|
+
stack.pop()
|
|
126
|
+
|
|
127
|
+
node.end_time = ts
|
|
128
|
+
node.duration = max(0.0, ts - node.start_time)
|
|
129
|
+
del partial_map[uuid]
|
|
130
|
+
|
|
131
|
+
# partial calls remain in stack => they have no final end_time
|
|
132
|
+
# we won't forcibly remove them
|
|
133
|
+
|
|
134
|
+
# collect top-level nodes
|
|
135
|
+
roots = []
|
|
136
|
+
for _, node in top_level_dict.items():
|
|
137
|
+
if node.parent is None:
|
|
138
|
+
roots.append(node)
|
|
139
|
+
|
|
140
|
+
return roots
|
|
141
|
+
|
|
142
|
+
|
|
143
|
+
def build_call_tree_per_example(all_steps: list[list[IntermediateStep]]) -> list[CallNode]:
|
|
144
|
+
"""
|
|
145
|
+
1) Group the DataFrame by example_number.
|
|
146
|
+
2) For each example, build a separate stack-based call tree.
|
|
147
|
+
3) Return a combined list of all top-level calls from all examples.
|
|
148
|
+
|
|
149
|
+
This ensures no cross-example nesting.
|
|
150
|
+
"""
|
|
151
|
+
df = create_standardized_dataframe(all_steps)
|
|
152
|
+
required = {"example_number", "event_type", "UUID", "event_timestamp"}
|
|
153
|
+
missing = required - set(df.columns)
|
|
154
|
+
if missing:
|
|
155
|
+
raise ValueError(f"DataFrame missing required columns: {missing}")
|
|
156
|
+
|
|
157
|
+
# Sort globally first (so each example is also in ascending time)
|
|
158
|
+
dfc = df.copy()
|
|
159
|
+
dfc.sort_values(["example_number", "event_timestamp"], inplace=True)
|
|
160
|
+
|
|
161
|
+
# We'll collect top-level calls for each example
|
|
162
|
+
all_roots: list[CallNode] = []
|
|
163
|
+
|
|
164
|
+
for _, group_df in dfc.groupby("example_number"):
|
|
165
|
+
# Build the call tree for this single example
|
|
166
|
+
# group_df is already sorted within this example
|
|
167
|
+
roots_for_example = build_call_tree_for_example(group_df)
|
|
168
|
+
all_roots.extend(roots_for_example)
|
|
169
|
+
|
|
170
|
+
return all_roots
|
|
171
|
+
|
|
172
|
+
|
|
173
|
+
# --------------------------------------------------------------------------------
|
|
174
|
+
# 2) Concurrency Computation
|
|
175
|
+
# --------------------------------------------------------------------------------
|
|
176
|
+
|
|
177
|
+
|
|
178
|
+
def compute_time_based_concurrency(roots: list[CallNode]) -> ConcurrencyDistribution:
|
|
179
|
+
"""
|
|
180
|
+
Build a timeline of (start, +1), (end, -1) from all calls, then:
|
|
181
|
+
- Sort events by time
|
|
182
|
+
- Create segments [ (t_i, t_{i+1}, concurrency) ]
|
|
183
|
+
- Compute concurrency percentiles (p50, p90, p95, p99) based on total time spent at each concurrency.
|
|
184
|
+
- This concurrency is across ALL calls from ALL examples.
|
|
185
|
+
|
|
186
|
+
Returns:
|
|
187
|
+
--------
|
|
188
|
+
ConcurrencyDistribution
|
|
189
|
+
with the piecewise segments + concurrency percentiles.
|
|
190
|
+
"""
|
|
191
|
+
# Flatten
|
|
192
|
+
all_nodes = []
|
|
193
|
+
|
|
194
|
+
def dfs(n: CallNode):
|
|
195
|
+
all_nodes.append(n)
|
|
196
|
+
for c in n.children:
|
|
197
|
+
dfs(c)
|
|
198
|
+
|
|
199
|
+
for r in roots:
|
|
200
|
+
dfs(r)
|
|
201
|
+
|
|
202
|
+
if not all_nodes:
|
|
203
|
+
return ConcurrencyDistribution(timeline_segments=[], p50=0, p90=0, p95=0, p99=0)
|
|
204
|
+
|
|
205
|
+
events = []
|
|
206
|
+
for n in all_nodes:
|
|
207
|
+
st = n.start_time
|
|
208
|
+
et = n.end_time
|
|
209
|
+
if st > et:
|
|
210
|
+
# partial or invalid => skip
|
|
211
|
+
continue
|
|
212
|
+
events.append((st, +1))
|
|
213
|
+
events.append((et, -1))
|
|
214
|
+
|
|
215
|
+
events.sort(key=lambda x: x[0])
|
|
216
|
+
timeline_segments: list[tuple[float, float, int]] = []
|
|
217
|
+
curr_concurrency = 0
|
|
218
|
+
prev_time = events[0][0]
|
|
219
|
+
|
|
220
|
+
for _, (t, delta) in enumerate(events):
|
|
221
|
+
if t > prev_time:
|
|
222
|
+
# segment is [prev_time, t) at concurrency=curr_concurrency
|
|
223
|
+
timeline_segments.append((prev_time, t, curr_concurrency))
|
|
224
|
+
curr_concurrency += delta
|
|
225
|
+
prev_time = t
|
|
226
|
+
|
|
227
|
+
# Summaries
|
|
228
|
+
total_time = 0.0
|
|
229
|
+
concurrency_durations: dict[int, float] = {}
|
|
230
|
+
|
|
231
|
+
for (seg_start, seg_end, c_val) in timeline_segments:
|
|
232
|
+
length = seg_end - seg_start
|
|
233
|
+
if length <= 0:
|
|
234
|
+
continue
|
|
235
|
+
total_time += length
|
|
236
|
+
concurrency_durations[c_val] = concurrency_durations.get(c_val, 0) + length
|
|
237
|
+
|
|
238
|
+
if total_time <= 0:
|
|
239
|
+
return ConcurrencyDistribution(timeline_segments=timeline_segments, p50=0, p90=0, p95=0, p99=0)
|
|
240
|
+
|
|
241
|
+
# Build concurrency-level distribution
|
|
242
|
+
sorted_levels = sorted(concurrency_durations.items(), key=lambda x: x[0]) # ascending concurrency
|
|
243
|
+
|
|
244
|
+
def concurrency_at_percentile(p: float) -> float:
|
|
245
|
+
threshold = total_time * (p / 100.0)
|
|
246
|
+
accum = 0.0
|
|
247
|
+
last_c = 0
|
|
248
|
+
for c_val, c_dur in sorted_levels:
|
|
249
|
+
accum += c_dur
|
|
250
|
+
if accum >= threshold:
|
|
251
|
+
return float(c_val)
|
|
252
|
+
last_c = c_val
|
|
253
|
+
return float(last_c)
|
|
254
|
+
|
|
255
|
+
p50_val = concurrency_at_percentile(50)
|
|
256
|
+
p90_val = concurrency_at_percentile(90)
|
|
257
|
+
p95_val = concurrency_at_percentile(95)
|
|
258
|
+
p99_val = concurrency_at_percentile(99)
|
|
259
|
+
|
|
260
|
+
return ConcurrencyDistribution(timeline_segments=timeline_segments,
|
|
261
|
+
p50=p50_val,
|
|
262
|
+
p90=p90_val,
|
|
263
|
+
p95=p95_val,
|
|
264
|
+
p99=p99_val)
|
|
265
|
+
|
|
266
|
+
|
|
267
|
+
def find_midpoint_concurrency(node: CallNode, segments: list[tuple[float, float, int]]) -> float:
|
|
268
|
+
"""
|
|
269
|
+
Approximate concurrency for a node by finding the concurrency in timeline_segments
|
|
270
|
+
at the node's midpoint (or start if zero-length).
|
|
271
|
+
"""
|
|
272
|
+
if node.start_time >= node.end_time:
|
|
273
|
+
mid = node.start_time
|
|
274
|
+
else:
|
|
275
|
+
mid = 0.5 * (node.start_time + node.end_time)
|
|
276
|
+
|
|
277
|
+
# Binary search in segments
|
|
278
|
+
left, right = 0, len(segments) - 1
|
|
279
|
+
while left <= right:
|
|
280
|
+
mid_idx = (left + right) // 2
|
|
281
|
+
seg_start, seg_end, seg_conc = segments[mid_idx]
|
|
282
|
+
if seg_start <= mid < seg_end:
|
|
283
|
+
return float(seg_conc)
|
|
284
|
+
if mid < seg_start:
|
|
285
|
+
right = mid_idx - 1
|
|
286
|
+
else:
|
|
287
|
+
left = mid_idx + 1
|
|
288
|
+
return 0.0
|
|
289
|
+
|
|
290
|
+
|
|
291
|
+
# --------------------------------------------------------------------------------
|
|
292
|
+
# 3) Gantt Chart
|
|
293
|
+
# --------------------------------------------------------------------------------
|
|
294
|
+
|
|
295
|
+
|
|
296
|
+
def save_gantt_chart(all_nodes: list[CallNode], output_path: str) -> None:
|
|
297
|
+
"""
|
|
298
|
+
Save a Gantt chart as a PNG, color-coded by operation_type.
|
|
299
|
+
Each node is displayed as a horizontal bar from start_time to end_time.
|
|
300
|
+
The y-axis is the node index (sorted by start_time).
|
|
301
|
+
"""
|
|
302
|
+
|
|
303
|
+
try:
|
|
304
|
+
import matplotlib.pyplot as plt
|
|
305
|
+
except ImportError:
|
|
306
|
+
logger.error("matplotlib is not installed. Please install matplotlib to use generate plots for the profiler "
|
|
307
|
+
"or install `aiq[profiler]` to install all necessary profiling packages.")
|
|
308
|
+
|
|
309
|
+
raise
|
|
310
|
+
|
|
311
|
+
# Sort calls by start_time
|
|
312
|
+
sorted_nodes = sorted(all_nodes, key=lambda x: x.start_time)
|
|
313
|
+
min_start = sorted_nodes[0].start_time
|
|
314
|
+
max_end = max(node.end_time for node in sorted_nodes)
|
|
315
|
+
|
|
316
|
+
color_map = {
|
|
317
|
+
"LLM": "tab:blue",
|
|
318
|
+
"TOOL": "tab:green",
|
|
319
|
+
"FUNCTION": "tab:orange",
|
|
320
|
+
}
|
|
321
|
+
default_color = "tab:gray"
|
|
322
|
+
|
|
323
|
+
fig, ax = plt.subplots(figsize=(20, 15))
|
|
324
|
+
|
|
325
|
+
y_positions = range(len(sorted_nodes))
|
|
326
|
+
labels = []
|
|
327
|
+
for i, node in enumerate(sorted_nodes):
|
|
328
|
+
start = node.start_time
|
|
329
|
+
width = node.end_time - node.start_time
|
|
330
|
+
c = color_map.get(node.operation_type, default_color)
|
|
331
|
+
ax.barh(y=i, width=width, left=start - min_start, height=0.6, color=c, edgecolor="black")
|
|
332
|
+
labels.append(f"{node.operation_type}:{node.operation_name}")
|
|
333
|
+
|
|
334
|
+
ax.set_yticks(list(y_positions))
|
|
335
|
+
ax.set_yticklabels(labels)
|
|
336
|
+
ax.invert_yaxis()
|
|
337
|
+
ax.set_xlim(0, max_end - min_start)
|
|
338
|
+
ax.set_xlabel("Time")
|
|
339
|
+
ax.set_title("Gantt Chart of Nested Calls (All Examples)")
|
|
340
|
+
plt.tight_layout()
|
|
341
|
+
plt.savefig(output_path, dpi=150)
|
|
342
|
+
plt.close(fig)
|
|
343
|
+
|
|
344
|
+
|
|
345
|
+
# --------------------------------------------------------------------------------
|
|
346
|
+
# 4) Analysis & Final Pydantic Result
|
|
347
|
+
# --------------------------------------------------------------------------------
|
|
348
|
+
|
|
349
|
+
|
|
350
|
+
def analyze_calls_and_build_result(roots: list[CallNode], output_dir: str | None = None) -> NestedCallProfilingResult:
|
|
351
|
+
"""
|
|
352
|
+
1. Compute concurrency distribution (p50, p90, p95, p99) across ALL calls in all examples.
|
|
353
|
+
2. For each node, compute self_time, subtree_time, concurrency at midpoint, bottleneck_score.
|
|
354
|
+
3. Identify top 5 bottlenecks (by subtree_time).
|
|
355
|
+
4. Build a textual report.
|
|
356
|
+
5. Optionally save a Gantt chart to 'output_dir'.
|
|
357
|
+
|
|
358
|
+
Returns NestedCallProfilingResult.
|
|
359
|
+
"""
|
|
360
|
+
if not roots:
|
|
361
|
+
empty_concurrency = ConcurrencyDistribution(timeline_segments=[], p50=0, p90=0, p95=0, p99=0)
|
|
362
|
+
return NestedCallProfilingResult(concurrency=empty_concurrency,
|
|
363
|
+
node_metrics={},
|
|
364
|
+
top_bottlenecks=[],
|
|
365
|
+
textual_report="No calls found.")
|
|
366
|
+
|
|
367
|
+
# Flatten all calls
|
|
368
|
+
all_nodes: list[CallNode] = []
|
|
369
|
+
|
|
370
|
+
def dfs(n: CallNode):
|
|
371
|
+
all_nodes.append(n)
|
|
372
|
+
for c in n.children:
|
|
373
|
+
dfs(c)
|
|
374
|
+
|
|
375
|
+
for r in roots:
|
|
376
|
+
dfs(r)
|
|
377
|
+
|
|
378
|
+
# 1) concurrency across all calls
|
|
379
|
+
concurrency_info = compute_time_based_concurrency(roots)
|
|
380
|
+
|
|
381
|
+
# 2) build NodeMetrics
|
|
382
|
+
node_metrics_map: dict[str, NodeMetrics] = {}
|
|
383
|
+
for node in all_nodes:
|
|
384
|
+
self_t = node.compute_self_time()
|
|
385
|
+
subtree_t = node.compute_subtree_time()
|
|
386
|
+
bscore = subtree_t
|
|
387
|
+
mid_conc = find_midpoint_concurrency(node, concurrency_info.timeline_segments)
|
|
388
|
+
|
|
389
|
+
m = NodeMetrics(uuid=node.uuid,
|
|
390
|
+
operation_type=node.operation_type,
|
|
391
|
+
operation_name=node.operation_name,
|
|
392
|
+
start_time=node.start_time,
|
|
393
|
+
end_time=node.end_time,
|
|
394
|
+
duration=node.duration,
|
|
395
|
+
self_time=self_t,
|
|
396
|
+
subtree_time=subtree_t,
|
|
397
|
+
concurrency_midpoint=mid_conc,
|
|
398
|
+
bottleneck_score=bscore)
|
|
399
|
+
node_metrics_map[node.uuid] = m
|
|
400
|
+
|
|
401
|
+
# 3) top 5
|
|
402
|
+
all_metrics = list(node_metrics_map.values())
|
|
403
|
+
sorted_metrics = sorted(all_metrics, key=lambda x: x.bottleneck_score, reverse=True)
|
|
404
|
+
top_5 = sorted_metrics[:5]
|
|
405
|
+
|
|
406
|
+
# 4) textual report
|
|
407
|
+
lines = []
|
|
408
|
+
lines.append("=== Multi-Example Nested Call Profiling Report ===")
|
|
409
|
+
lines.append(f"Total calls (across all examples): {len(all_nodes)}")
|
|
410
|
+
|
|
411
|
+
lines.append("\n-- Concurrency Distribution (all examples) --")
|
|
412
|
+
lines.append(f"p50={concurrency_info.p50:.1f}, p90={concurrency_info.p90:.1f}, "
|
|
413
|
+
f"p95={concurrency_info.p95:.1f}, p99={concurrency_info.p99:.1f}")
|
|
414
|
+
|
|
415
|
+
lines.append("\n-- Top 5 Calls by Bottleneck Score (subtree_time) --")
|
|
416
|
+
for i, tm in enumerate(top_5, start=1):
|
|
417
|
+
lines.append(f"{i}) UUID={tm.uuid}, {tm.operation_type} '{tm.operation_name}', "
|
|
418
|
+
f"dur={tm.duration:.2f}, self_time={tm.self_time:.2f}, "
|
|
419
|
+
f"subtree_time={tm.subtree_time:.2f}, concurrency={tm.concurrency_midpoint:.1f}, "
|
|
420
|
+
f"score={tm.bottleneck_score:.2f}")
|
|
421
|
+
|
|
422
|
+
lines.append("\n-- Full Tree(s) (All Examples) --")
|
|
423
|
+
|
|
424
|
+
for root in roots:
|
|
425
|
+
lines.append(str(root))
|
|
426
|
+
|
|
427
|
+
report_text = "\n".join(lines)
|
|
428
|
+
|
|
429
|
+
# 5) optional Gantt chart
|
|
430
|
+
if output_dir:
|
|
431
|
+
os.makedirs(output_dir, exist_ok=True)
|
|
432
|
+
chart_path = os.path.join(output_dir, "gantt_chart.png")
|
|
433
|
+
save_gantt_chart(all_nodes, chart_path)
|
|
434
|
+
|
|
435
|
+
# Return the final Pydantic result
|
|
436
|
+
return NestedCallProfilingResult(concurrency=concurrency_info,
|
|
437
|
+
node_metrics=node_metrics_map,
|
|
438
|
+
top_bottlenecks=top_5,
|
|
439
|
+
textual_report=report_text)
|
|
440
|
+
|
|
441
|
+
|
|
442
|
+
def multi_example_call_profiling(all_steps: list[list[IntermediateStep]],
|
|
443
|
+
output_dir: str | None = None) -> NestedCallProfilingResult:
|
|
444
|
+
"""
|
|
445
|
+
The high-level function:
|
|
446
|
+
|
|
447
|
+
1. Build a forest of calls by grouping by example_number (so no cross-example nesting).
|
|
448
|
+
2. Analyze concurrency across all calls in all examples.
|
|
449
|
+
3. Return a NestedCallProfilingResult with concurrency distribution, node metrics, top bottlenecks, and textual
|
|
450
|
+
report. Optionally saves a Gantt chart.
|
|
451
|
+
|
|
452
|
+
:param all_steps: Intermediate steps for each example.
|
|
453
|
+
:param output_dir: Directory path to save gantt_chart.png (if provided)
|
|
454
|
+
:return: NestedCallProfilingResult (pydantic)
|
|
455
|
+
"""
|
|
456
|
+
# Build the forest (all examples combined)
|
|
457
|
+
roots = build_call_tree_per_example(all_steps)
|
|
458
|
+
# Analyze calls
|
|
459
|
+
result = analyze_calls_and_build_result(roots, output_dir=output_dir)
|
|
460
|
+
return result
|