nvidia-nat 1.1.0a20251020__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- aiq/__init__.py +66 -0
- nat/agent/__init__.py +0 -0
- nat/agent/base.py +265 -0
- nat/agent/dual_node.py +72 -0
- nat/agent/prompt_optimizer/__init__.py +0 -0
- nat/agent/prompt_optimizer/prompt.py +68 -0
- nat/agent/prompt_optimizer/register.py +149 -0
- nat/agent/react_agent/__init__.py +0 -0
- nat/agent/react_agent/agent.py +394 -0
- nat/agent/react_agent/output_parser.py +104 -0
- nat/agent/react_agent/prompt.py +44 -0
- nat/agent/react_agent/register.py +168 -0
- nat/agent/reasoning_agent/__init__.py +0 -0
- nat/agent/reasoning_agent/reasoning_agent.py +227 -0
- nat/agent/register.py +23 -0
- nat/agent/rewoo_agent/__init__.py +0 -0
- nat/agent/rewoo_agent/agent.py +593 -0
- nat/agent/rewoo_agent/prompt.py +107 -0
- nat/agent/rewoo_agent/register.py +175 -0
- nat/agent/tool_calling_agent/__init__.py +0 -0
- nat/agent/tool_calling_agent/agent.py +246 -0
- nat/agent/tool_calling_agent/register.py +129 -0
- nat/authentication/__init__.py +14 -0
- nat/authentication/api_key/__init__.py +14 -0
- nat/authentication/api_key/api_key_auth_provider.py +96 -0
- nat/authentication/api_key/api_key_auth_provider_config.py +124 -0
- nat/authentication/api_key/register.py +26 -0
- nat/authentication/credential_validator/__init__.py +14 -0
- nat/authentication/credential_validator/bearer_token_validator.py +557 -0
- nat/authentication/exceptions/__init__.py +14 -0
- nat/authentication/exceptions/api_key_exceptions.py +38 -0
- nat/authentication/http_basic_auth/__init__.py +0 -0
- nat/authentication/http_basic_auth/http_basic_auth_provider.py +81 -0
- nat/authentication/http_basic_auth/register.py +30 -0
- nat/authentication/interfaces.py +96 -0
- nat/authentication/oauth2/__init__.py +14 -0
- nat/authentication/oauth2/oauth2_auth_code_flow_provider.py +140 -0
- nat/authentication/oauth2/oauth2_auth_code_flow_provider_config.py +39 -0
- nat/authentication/oauth2/oauth2_resource_server_config.py +124 -0
- nat/authentication/oauth2/register.py +25 -0
- nat/authentication/register.py +20 -0
- nat/builder/__init__.py +0 -0
- nat/builder/builder.py +317 -0
- nat/builder/component_utils.py +320 -0
- nat/builder/context.py +321 -0
- nat/builder/embedder.py +24 -0
- nat/builder/eval_builder.py +166 -0
- nat/builder/evaluator.py +29 -0
- nat/builder/framework_enum.py +25 -0
- nat/builder/front_end.py +73 -0
- nat/builder/function.py +714 -0
- nat/builder/function_base.py +380 -0
- nat/builder/function_info.py +625 -0
- nat/builder/intermediate_step_manager.py +206 -0
- nat/builder/llm.py +25 -0
- nat/builder/retriever.py +25 -0
- nat/builder/user_interaction_manager.py +78 -0
- nat/builder/workflow.py +160 -0
- nat/builder/workflow_builder.py +1365 -0
- nat/cli/__init__.py +14 -0
- nat/cli/cli_utils/__init__.py +0 -0
- nat/cli/cli_utils/config_override.py +231 -0
- nat/cli/cli_utils/validation.py +37 -0
- nat/cli/commands/__init__.py +0 -0
- nat/cli/commands/configure/__init__.py +0 -0
- nat/cli/commands/configure/channel/__init__.py +0 -0
- nat/cli/commands/configure/channel/add.py +28 -0
- nat/cli/commands/configure/channel/channel.py +34 -0
- nat/cli/commands/configure/channel/remove.py +30 -0
- nat/cli/commands/configure/channel/update.py +30 -0
- nat/cli/commands/configure/configure.py +33 -0
- nat/cli/commands/evaluate.py +139 -0
- nat/cli/commands/info/__init__.py +14 -0
- nat/cli/commands/info/info.py +47 -0
- nat/cli/commands/info/list_channels.py +32 -0
- nat/cli/commands/info/list_components.py +128 -0
- nat/cli/commands/mcp/__init__.py +14 -0
- nat/cli/commands/mcp/mcp.py +986 -0
- nat/cli/commands/object_store/__init__.py +14 -0
- nat/cli/commands/object_store/object_store.py +227 -0
- nat/cli/commands/optimize.py +90 -0
- nat/cli/commands/registry/__init__.py +14 -0
- nat/cli/commands/registry/publish.py +88 -0
- nat/cli/commands/registry/pull.py +118 -0
- nat/cli/commands/registry/registry.py +36 -0
- nat/cli/commands/registry/remove.py +108 -0
- nat/cli/commands/registry/search.py +153 -0
- nat/cli/commands/sizing/__init__.py +14 -0
- nat/cli/commands/sizing/calc.py +297 -0
- nat/cli/commands/sizing/sizing.py +27 -0
- nat/cli/commands/start.py +257 -0
- nat/cli/commands/uninstall.py +81 -0
- nat/cli/commands/validate.py +47 -0
- nat/cli/commands/workflow/__init__.py +14 -0
- nat/cli/commands/workflow/templates/__init__.py.j2 +0 -0
- nat/cli/commands/workflow/templates/config.yml.j2 +17 -0
- nat/cli/commands/workflow/templates/pyproject.toml.j2 +25 -0
- nat/cli/commands/workflow/templates/register.py.j2 +4 -0
- nat/cli/commands/workflow/templates/workflow.py.j2 +50 -0
- nat/cli/commands/workflow/workflow.py +37 -0
- nat/cli/commands/workflow/workflow_commands.py +403 -0
- nat/cli/entrypoint.py +141 -0
- nat/cli/main.py +60 -0
- nat/cli/register_workflow.py +522 -0
- nat/cli/type_registry.py +1069 -0
- nat/control_flow/__init__.py +0 -0
- nat/control_flow/register.py +20 -0
- nat/control_flow/router_agent/__init__.py +0 -0
- nat/control_flow/router_agent/agent.py +329 -0
- nat/control_flow/router_agent/prompt.py +48 -0
- nat/control_flow/router_agent/register.py +91 -0
- nat/control_flow/sequential_executor.py +166 -0
- nat/data_models/__init__.py +14 -0
- nat/data_models/agent.py +34 -0
- nat/data_models/api_server.py +843 -0
- nat/data_models/authentication.py +245 -0
- nat/data_models/common.py +171 -0
- nat/data_models/component.py +60 -0
- nat/data_models/component_ref.py +179 -0
- nat/data_models/config.py +434 -0
- nat/data_models/dataset_handler.py +169 -0
- nat/data_models/discovery_metadata.py +305 -0
- nat/data_models/embedder.py +27 -0
- nat/data_models/evaluate.py +130 -0
- nat/data_models/evaluator.py +26 -0
- nat/data_models/front_end.py +26 -0
- nat/data_models/function.py +64 -0
- nat/data_models/function_dependencies.py +80 -0
- nat/data_models/gated_field_mixin.py +242 -0
- nat/data_models/interactive.py +246 -0
- nat/data_models/intermediate_step.py +302 -0
- nat/data_models/invocation_node.py +38 -0
- nat/data_models/llm.py +27 -0
- nat/data_models/logging.py +26 -0
- nat/data_models/memory.py +27 -0
- nat/data_models/object_store.py +44 -0
- nat/data_models/optimizable.py +119 -0
- nat/data_models/optimizer.py +149 -0
- nat/data_models/profiler.py +54 -0
- nat/data_models/registry_handler.py +26 -0
- nat/data_models/retriever.py +30 -0
- nat/data_models/retry_mixin.py +35 -0
- nat/data_models/span.py +228 -0
- nat/data_models/step_adaptor.py +64 -0
- nat/data_models/streaming.py +33 -0
- nat/data_models/swe_bench_model.py +54 -0
- nat/data_models/telemetry_exporter.py +26 -0
- nat/data_models/temperature_mixin.py +44 -0
- nat/data_models/thinking_mixin.py +86 -0
- nat/data_models/top_p_mixin.py +44 -0
- nat/data_models/ttc_strategy.py +30 -0
- nat/embedder/__init__.py +0 -0
- nat/embedder/azure_openai_embedder.py +46 -0
- nat/embedder/nim_embedder.py +59 -0
- nat/embedder/openai_embedder.py +42 -0
- nat/embedder/register.py +22 -0
- nat/eval/__init__.py +14 -0
- nat/eval/config.py +62 -0
- nat/eval/dataset_handler/__init__.py +0 -0
- nat/eval/dataset_handler/dataset_downloader.py +106 -0
- nat/eval/dataset_handler/dataset_filter.py +52 -0
- nat/eval/dataset_handler/dataset_handler.py +431 -0
- nat/eval/evaluate.py +565 -0
- nat/eval/evaluator/__init__.py +14 -0
- nat/eval/evaluator/base_evaluator.py +77 -0
- nat/eval/evaluator/evaluator_model.py +58 -0
- nat/eval/intermediate_step_adapter.py +99 -0
- nat/eval/rag_evaluator/__init__.py +0 -0
- nat/eval/rag_evaluator/evaluate.py +178 -0
- nat/eval/rag_evaluator/register.py +143 -0
- nat/eval/register.py +26 -0
- nat/eval/remote_workflow.py +133 -0
- nat/eval/runners/__init__.py +14 -0
- nat/eval/runners/config.py +39 -0
- nat/eval/runners/multi_eval_runner.py +54 -0
- nat/eval/runtime_evaluator/__init__.py +14 -0
- nat/eval/runtime_evaluator/evaluate.py +123 -0
- nat/eval/runtime_evaluator/register.py +100 -0
- nat/eval/runtime_event_subscriber.py +52 -0
- nat/eval/swe_bench_evaluator/__init__.py +0 -0
- nat/eval/swe_bench_evaluator/evaluate.py +215 -0
- nat/eval/swe_bench_evaluator/register.py +36 -0
- nat/eval/trajectory_evaluator/__init__.py +0 -0
- nat/eval/trajectory_evaluator/evaluate.py +75 -0
- nat/eval/trajectory_evaluator/register.py +40 -0
- nat/eval/tunable_rag_evaluator/__init__.py +0 -0
- nat/eval/tunable_rag_evaluator/evaluate.py +242 -0
- nat/eval/tunable_rag_evaluator/register.py +52 -0
- nat/eval/usage_stats.py +41 -0
- nat/eval/utils/__init__.py +0 -0
- nat/eval/utils/eval_trace_ctx.py +89 -0
- nat/eval/utils/output_uploader.py +140 -0
- nat/eval/utils/tqdm_position_registry.py +40 -0
- nat/eval/utils/weave_eval.py +193 -0
- nat/experimental/__init__.py +0 -0
- nat/experimental/decorators/__init__.py +0 -0
- nat/experimental/decorators/experimental_warning_decorator.py +154 -0
- nat/experimental/test_time_compute/__init__.py +0 -0
- nat/experimental/test_time_compute/editing/__init__.py +0 -0
- nat/experimental/test_time_compute/editing/iterative_plan_refinement_editor.py +147 -0
- nat/experimental/test_time_compute/editing/llm_as_a_judge_editor.py +204 -0
- nat/experimental/test_time_compute/editing/motivation_aware_summarization.py +107 -0
- nat/experimental/test_time_compute/functions/__init__.py +0 -0
- nat/experimental/test_time_compute/functions/execute_score_select_function.py +105 -0
- nat/experimental/test_time_compute/functions/plan_select_execute_function.py +228 -0
- nat/experimental/test_time_compute/functions/ttc_tool_orchestration_function.py +205 -0
- nat/experimental/test_time_compute/functions/ttc_tool_wrapper_function.py +146 -0
- nat/experimental/test_time_compute/models/__init__.py +0 -0
- nat/experimental/test_time_compute/models/editor_config.py +132 -0
- nat/experimental/test_time_compute/models/scoring_config.py +112 -0
- nat/experimental/test_time_compute/models/search_config.py +120 -0
- nat/experimental/test_time_compute/models/selection_config.py +154 -0
- nat/experimental/test_time_compute/models/stage_enums.py +43 -0
- nat/experimental/test_time_compute/models/strategy_base.py +67 -0
- nat/experimental/test_time_compute/models/tool_use_config.py +41 -0
- nat/experimental/test_time_compute/models/ttc_item.py +48 -0
- nat/experimental/test_time_compute/register.py +35 -0
- nat/experimental/test_time_compute/scoring/__init__.py +0 -0
- nat/experimental/test_time_compute/scoring/llm_based_agent_scorer.py +168 -0
- nat/experimental/test_time_compute/scoring/llm_based_plan_scorer.py +168 -0
- nat/experimental/test_time_compute/scoring/motivation_aware_scorer.py +111 -0
- nat/experimental/test_time_compute/search/__init__.py +0 -0
- nat/experimental/test_time_compute/search/multi_llm_planner.py +128 -0
- nat/experimental/test_time_compute/search/multi_query_retrieval_search.py +122 -0
- nat/experimental/test_time_compute/search/single_shot_multi_plan_planner.py +128 -0
- nat/experimental/test_time_compute/selection/__init__.py +0 -0
- nat/experimental/test_time_compute/selection/best_of_n_selector.py +63 -0
- nat/experimental/test_time_compute/selection/llm_based_agent_output_selector.py +131 -0
- nat/experimental/test_time_compute/selection/llm_based_output_merging_selector.py +157 -0
- nat/experimental/test_time_compute/selection/llm_based_plan_selector.py +128 -0
- nat/experimental/test_time_compute/selection/threshold_selector.py +58 -0
- nat/front_ends/__init__.py +14 -0
- nat/front_ends/console/__init__.py +14 -0
- nat/front_ends/console/authentication_flow_handler.py +285 -0
- nat/front_ends/console/console_front_end_config.py +32 -0
- nat/front_ends/console/console_front_end_plugin.py +108 -0
- nat/front_ends/console/register.py +25 -0
- nat/front_ends/cron/__init__.py +14 -0
- nat/front_ends/fastapi/__init__.py +14 -0
- nat/front_ends/fastapi/auth_flow_handlers/__init__.py +0 -0
- nat/front_ends/fastapi/auth_flow_handlers/http_flow_handler.py +27 -0
- nat/front_ends/fastapi/auth_flow_handlers/websocket_flow_handler.py +142 -0
- nat/front_ends/fastapi/dask_client_mixin.py +65 -0
- nat/front_ends/fastapi/fastapi_front_end_config.py +272 -0
- nat/front_ends/fastapi/fastapi_front_end_controller.py +68 -0
- nat/front_ends/fastapi/fastapi_front_end_plugin.py +247 -0
- nat/front_ends/fastapi/fastapi_front_end_plugin_worker.py +1257 -0
- nat/front_ends/fastapi/html_snippets/__init__.py +14 -0
- nat/front_ends/fastapi/html_snippets/auth_code_grant_success.py +35 -0
- nat/front_ends/fastapi/intermediate_steps_subscriber.py +80 -0
- nat/front_ends/fastapi/job_store.py +602 -0
- nat/front_ends/fastapi/main.py +64 -0
- nat/front_ends/fastapi/message_handler.py +344 -0
- nat/front_ends/fastapi/message_validator.py +351 -0
- nat/front_ends/fastapi/register.py +25 -0
- nat/front_ends/fastapi/response_helpers.py +195 -0
- nat/front_ends/fastapi/step_adaptor.py +319 -0
- nat/front_ends/fastapi/utils.py +57 -0
- nat/front_ends/mcp/__init__.py +14 -0
- nat/front_ends/mcp/introspection_token_verifier.py +73 -0
- nat/front_ends/mcp/mcp_front_end_config.py +90 -0
- nat/front_ends/mcp/mcp_front_end_plugin.py +113 -0
- nat/front_ends/mcp/mcp_front_end_plugin_worker.py +268 -0
- nat/front_ends/mcp/memory_profiler.py +320 -0
- nat/front_ends/mcp/register.py +27 -0
- nat/front_ends/mcp/tool_converter.py +290 -0
- nat/front_ends/register.py +21 -0
- nat/front_ends/simple_base/__init__.py +14 -0
- nat/front_ends/simple_base/simple_front_end_plugin_base.py +56 -0
- nat/llm/__init__.py +0 -0
- nat/llm/aws_bedrock_llm.py +69 -0
- nat/llm/azure_openai_llm.py +57 -0
- nat/llm/litellm_llm.py +69 -0
- nat/llm/nim_llm.py +58 -0
- nat/llm/openai_llm.py +54 -0
- nat/llm/register.py +27 -0
- nat/llm/utils/__init__.py +14 -0
- nat/llm/utils/env_config_value.py +93 -0
- nat/llm/utils/error.py +17 -0
- nat/llm/utils/thinking.py +215 -0
- nat/memory/__init__.py +20 -0
- nat/memory/interfaces.py +183 -0
- nat/memory/models.py +112 -0
- nat/meta/pypi.md +58 -0
- nat/object_store/__init__.py +20 -0
- nat/object_store/in_memory_object_store.py +76 -0
- nat/object_store/interfaces.py +84 -0
- nat/object_store/models.py +38 -0
- nat/object_store/register.py +19 -0
- nat/observability/__init__.py +14 -0
- nat/observability/exporter/__init__.py +14 -0
- nat/observability/exporter/base_exporter.py +449 -0
- nat/observability/exporter/exporter.py +78 -0
- nat/observability/exporter/file_exporter.py +33 -0
- nat/observability/exporter/processing_exporter.py +550 -0
- nat/observability/exporter/raw_exporter.py +52 -0
- nat/observability/exporter/span_exporter.py +308 -0
- nat/observability/exporter_manager.py +335 -0
- nat/observability/mixin/__init__.py +14 -0
- nat/observability/mixin/batch_config_mixin.py +26 -0
- nat/observability/mixin/collector_config_mixin.py +23 -0
- nat/observability/mixin/file_mixin.py +288 -0
- nat/observability/mixin/file_mode.py +23 -0
- nat/observability/mixin/redaction_config_mixin.py +42 -0
- nat/observability/mixin/resource_conflict_mixin.py +134 -0
- nat/observability/mixin/serialize_mixin.py +61 -0
- nat/observability/mixin/tagging_config_mixin.py +62 -0
- nat/observability/mixin/type_introspection_mixin.py +496 -0
- nat/observability/processor/__init__.py +14 -0
- nat/observability/processor/batching_processor.py +308 -0
- nat/observability/processor/callback_processor.py +42 -0
- nat/observability/processor/falsy_batch_filter_processor.py +55 -0
- nat/observability/processor/intermediate_step_serializer.py +28 -0
- nat/observability/processor/processor.py +74 -0
- nat/observability/processor/processor_factory.py +70 -0
- nat/observability/processor/redaction/__init__.py +24 -0
- nat/observability/processor/redaction/contextual_redaction_processor.py +125 -0
- nat/observability/processor/redaction/contextual_span_redaction_processor.py +66 -0
- nat/observability/processor/redaction/redaction_processor.py +177 -0
- nat/observability/processor/redaction/span_header_redaction_processor.py +92 -0
- nat/observability/processor/span_tagging_processor.py +68 -0
- nat/observability/register.py +114 -0
- nat/observability/utils/__init__.py +14 -0
- nat/observability/utils/dict_utils.py +236 -0
- nat/observability/utils/time_utils.py +31 -0
- nat/plugins/.namespace +1 -0
- nat/profiler/__init__.py +0 -0
- nat/profiler/calc/__init__.py +14 -0
- nat/profiler/calc/calc_runner.py +626 -0
- nat/profiler/calc/calculations.py +288 -0
- nat/profiler/calc/data_models.py +188 -0
- nat/profiler/calc/plot.py +345 -0
- nat/profiler/callbacks/__init__.py +0 -0
- nat/profiler/callbacks/agno_callback_handler.py +295 -0
- nat/profiler/callbacks/base_callback_class.py +20 -0
- nat/profiler/callbacks/langchain_callback_handler.py +297 -0
- nat/profiler/callbacks/llama_index_callback_handler.py +205 -0
- nat/profiler/callbacks/semantic_kernel_callback_handler.py +238 -0
- nat/profiler/callbacks/token_usage_base_model.py +27 -0
- nat/profiler/data_frame_row.py +51 -0
- nat/profiler/data_models.py +24 -0
- nat/profiler/decorators/__init__.py +0 -0
- nat/profiler/decorators/framework_wrapper.py +180 -0
- nat/profiler/decorators/function_tracking.py +411 -0
- nat/profiler/forecasting/__init__.py +0 -0
- nat/profiler/forecasting/config.py +18 -0
- nat/profiler/forecasting/model_trainer.py +75 -0
- nat/profiler/forecasting/models/__init__.py +22 -0
- nat/profiler/forecasting/models/forecasting_base_model.py +42 -0
- nat/profiler/forecasting/models/linear_model.py +197 -0
- nat/profiler/forecasting/models/random_forest_regressor.py +269 -0
- nat/profiler/inference_metrics_model.py +28 -0
- nat/profiler/inference_optimization/__init__.py +0 -0
- nat/profiler/inference_optimization/bottleneck_analysis/__init__.py +0 -0
- nat/profiler/inference_optimization/bottleneck_analysis/nested_stack_analysis.py +460 -0
- nat/profiler/inference_optimization/bottleneck_analysis/simple_stack_analysis.py +258 -0
- nat/profiler/inference_optimization/data_models.py +386 -0
- nat/profiler/inference_optimization/experimental/__init__.py +0 -0
- nat/profiler/inference_optimization/experimental/concurrency_spike_analysis.py +468 -0
- nat/profiler/inference_optimization/experimental/prefix_span_analysis.py +404 -0
- nat/profiler/inference_optimization/llm_metrics.py +212 -0
- nat/profiler/inference_optimization/prompt_caching.py +163 -0
- nat/profiler/inference_optimization/token_uniqueness.py +107 -0
- nat/profiler/inference_optimization/workflow_runtimes.py +72 -0
- nat/profiler/intermediate_property_adapter.py +102 -0
- nat/profiler/parameter_optimization/__init__.py +0 -0
- nat/profiler/parameter_optimization/optimizable_utils.py +93 -0
- nat/profiler/parameter_optimization/optimizer_runtime.py +67 -0
- nat/profiler/parameter_optimization/parameter_optimizer.py +153 -0
- nat/profiler/parameter_optimization/parameter_selection.py +107 -0
- nat/profiler/parameter_optimization/pareto_visualizer.py +380 -0
- nat/profiler/parameter_optimization/prompt_optimizer.py +384 -0
- nat/profiler/parameter_optimization/update_helpers.py +66 -0
- nat/profiler/profile_runner.py +478 -0
- nat/profiler/utils.py +186 -0
- nat/registry_handlers/__init__.py +0 -0
- nat/registry_handlers/local/__init__.py +0 -0
- nat/registry_handlers/local/local_handler.py +176 -0
- nat/registry_handlers/local/register_local.py +37 -0
- nat/registry_handlers/metadata_factory.py +60 -0
- nat/registry_handlers/package_utils.py +570 -0
- nat/registry_handlers/pypi/__init__.py +0 -0
- nat/registry_handlers/pypi/pypi_handler.py +248 -0
- nat/registry_handlers/pypi/register_pypi.py +40 -0
- nat/registry_handlers/register.py +20 -0
- nat/registry_handlers/registry_handler_base.py +157 -0
- nat/registry_handlers/rest/__init__.py +0 -0
- nat/registry_handlers/rest/register_rest.py +56 -0
- nat/registry_handlers/rest/rest_handler.py +236 -0
- nat/registry_handlers/schemas/__init__.py +0 -0
- nat/registry_handlers/schemas/headers.py +42 -0
- nat/registry_handlers/schemas/package.py +68 -0
- nat/registry_handlers/schemas/publish.py +68 -0
- nat/registry_handlers/schemas/pull.py +82 -0
- nat/registry_handlers/schemas/remove.py +36 -0
- nat/registry_handlers/schemas/search.py +91 -0
- nat/registry_handlers/schemas/status.py +47 -0
- nat/retriever/__init__.py +0 -0
- nat/retriever/interface.py +41 -0
- nat/retriever/milvus/__init__.py +14 -0
- nat/retriever/milvus/register.py +81 -0
- nat/retriever/milvus/retriever.py +228 -0
- nat/retriever/models.py +77 -0
- nat/retriever/nemo_retriever/__init__.py +14 -0
- nat/retriever/nemo_retriever/register.py +60 -0
- nat/retriever/nemo_retriever/retriever.py +190 -0
- nat/retriever/register.py +21 -0
- nat/runtime/__init__.py +14 -0
- nat/runtime/loader.py +220 -0
- nat/runtime/runner.py +292 -0
- nat/runtime/session.py +223 -0
- nat/runtime/user_metadata.py +130 -0
- nat/settings/__init__.py +0 -0
- nat/settings/global_settings.py +329 -0
- nat/test/.namespace +1 -0
- nat/tool/__init__.py +0 -0
- nat/tool/chat_completion.py +77 -0
- nat/tool/code_execution/README.md +151 -0
- nat/tool/code_execution/__init__.py +0 -0
- nat/tool/code_execution/code_sandbox.py +267 -0
- nat/tool/code_execution/local_sandbox/.gitignore +1 -0
- nat/tool/code_execution/local_sandbox/Dockerfile.sandbox +60 -0
- nat/tool/code_execution/local_sandbox/__init__.py +13 -0
- nat/tool/code_execution/local_sandbox/local_sandbox_server.py +198 -0
- nat/tool/code_execution/local_sandbox/sandbox.requirements.txt +6 -0
- nat/tool/code_execution/local_sandbox/start_local_sandbox.sh +50 -0
- nat/tool/code_execution/register.py +74 -0
- nat/tool/code_execution/test_code_execution_sandbox.py +414 -0
- nat/tool/code_execution/utils.py +100 -0
- nat/tool/datetime_tools.py +82 -0
- nat/tool/document_search.py +141 -0
- nat/tool/github_tools.py +450 -0
- nat/tool/memory_tools/__init__.py +0 -0
- nat/tool/memory_tools/add_memory_tool.py +79 -0
- nat/tool/memory_tools/delete_memory_tool.py +66 -0
- nat/tool/memory_tools/get_memory_tool.py +72 -0
- nat/tool/nvidia_rag.py +95 -0
- nat/tool/register.py +31 -0
- nat/tool/retriever.py +95 -0
- nat/tool/server_tools.py +66 -0
- nat/utils/__init__.py +0 -0
- nat/utils/callable_utils.py +70 -0
- nat/utils/data_models/__init__.py +0 -0
- nat/utils/data_models/schema_validator.py +58 -0
- nat/utils/debugging_utils.py +43 -0
- nat/utils/decorators.py +210 -0
- nat/utils/dump_distro_mapping.py +32 -0
- nat/utils/exception_handlers/__init__.py +0 -0
- nat/utils/exception_handlers/automatic_retries.py +342 -0
- nat/utils/exception_handlers/schemas.py +114 -0
- nat/utils/io/__init__.py +0 -0
- nat/utils/io/model_processing.py +28 -0
- nat/utils/io/yaml_tools.py +119 -0
- nat/utils/log_levels.py +25 -0
- nat/utils/log_utils.py +37 -0
- nat/utils/metadata_utils.py +74 -0
- nat/utils/optional_imports.py +142 -0
- nat/utils/producer_consumer_queue.py +178 -0
- nat/utils/reactive/__init__.py +0 -0
- nat/utils/reactive/base/__init__.py +0 -0
- nat/utils/reactive/base/observable_base.py +65 -0
- nat/utils/reactive/base/observer_base.py +55 -0
- nat/utils/reactive/base/subject_base.py +79 -0
- nat/utils/reactive/observable.py +59 -0
- nat/utils/reactive/observer.py +76 -0
- nat/utils/reactive/subject.py +131 -0
- nat/utils/reactive/subscription.py +49 -0
- nat/utils/settings/__init__.py +0 -0
- nat/utils/settings/global_settings.py +195 -0
- nat/utils/string_utils.py +38 -0
- nat/utils/type_converter.py +299 -0
- nat/utils/type_utils.py +488 -0
- nat/utils/url_utils.py +27 -0
- nvidia_nat-1.1.0a20251020.dist-info/METADATA +195 -0
- nvidia_nat-1.1.0a20251020.dist-info/RECORD +480 -0
- nvidia_nat-1.1.0a20251020.dist-info/WHEEL +5 -0
- nvidia_nat-1.1.0a20251020.dist-info/entry_points.txt +22 -0
- nvidia_nat-1.1.0a20251020.dist-info/licenses/LICENSE-3rd-party.txt +5478 -0
- nvidia_nat-1.1.0a20251020.dist-info/licenses/LICENSE.md +201 -0
- nvidia_nat-1.1.0a20251020.dist-info/top_level.txt +2 -0
|
@@ -0,0 +1,140 @@
|
|
|
1
|
+
# SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
|
|
2
|
+
# SPDX-License-Identifier: Apache-2.0
|
|
3
|
+
#
|
|
4
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
5
|
+
# you may not use this file except in compliance with the License.
|
|
6
|
+
# You may obtain a copy of the License at
|
|
7
|
+
#
|
|
8
|
+
# http://www.apache.org/licenses/LICENSE-2.0
|
|
9
|
+
#
|
|
10
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
11
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
12
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
13
|
+
# See the License for the specific language governing permissions and
|
|
14
|
+
# limitations under the License.
|
|
15
|
+
|
|
16
|
+
import asyncio
|
|
17
|
+
import logging
|
|
18
|
+
import os
|
|
19
|
+
import subprocess
|
|
20
|
+
import sys
|
|
21
|
+
from pathlib import Path
|
|
22
|
+
|
|
23
|
+
import aioboto3
|
|
24
|
+
from botocore.exceptions import NoCredentialsError
|
|
25
|
+
from tqdm import tqdm
|
|
26
|
+
|
|
27
|
+
from nat.data_models.evaluate import EvalOutputConfig
|
|
28
|
+
|
|
29
|
+
logger = logging.getLogger(__name__)
|
|
30
|
+
|
|
31
|
+
|
|
32
|
+
class OutputUploader:
|
|
33
|
+
"""
|
|
34
|
+
Run custom scripts and upload evaluation outputs using the configured s3
|
|
35
|
+
credentials.
|
|
36
|
+
"""
|
|
37
|
+
|
|
38
|
+
def __init__(self, output_config: EvalOutputConfig, job_id: str | None = None):
|
|
39
|
+
self.output_config = output_config
|
|
40
|
+
self._s3_client = None
|
|
41
|
+
self.job_id = job_id
|
|
42
|
+
|
|
43
|
+
@property
|
|
44
|
+
def s3_config(self):
|
|
45
|
+
return self.output_config.s3
|
|
46
|
+
|
|
47
|
+
async def _upload_file(self, s3_client, bucket, s3_key, local_path, pbar):
|
|
48
|
+
try:
|
|
49
|
+
await s3_client.upload_file(str(local_path), bucket, s3_key)
|
|
50
|
+
logger.info("Uploaded %s to s3://%s/%s", local_path, bucket, s3_key)
|
|
51
|
+
pbar.update(1)
|
|
52
|
+
except Exception as e:
|
|
53
|
+
logger.error("Failed to upload %s to s3://%s/%s: %s", local_path, bucket, s3_key, e)
|
|
54
|
+
raise
|
|
55
|
+
|
|
56
|
+
async def upload_directory(self):
|
|
57
|
+
"""
|
|
58
|
+
Upload the contents of the local output directory to the remote S3 bucket in parallel.
|
|
59
|
+
"""
|
|
60
|
+
if not self.output_config.s3:
|
|
61
|
+
logger.info("No S3 config provided; skipping upload.")
|
|
62
|
+
return
|
|
63
|
+
|
|
64
|
+
local_dir = self.output_config.dir
|
|
65
|
+
bucket = self.s3_config.bucket
|
|
66
|
+
remote_prefix = self.output_config.remote_dir or ""
|
|
67
|
+
if self.job_id:
|
|
68
|
+
remote_prefix = str(Path(remote_prefix) / f"jobs/{self.job_id}")
|
|
69
|
+
|
|
70
|
+
file_entries = []
|
|
71
|
+
for root, _, files in os.walk(local_dir):
|
|
72
|
+
for file in files:
|
|
73
|
+
local_path = Path(root) / file
|
|
74
|
+
relative_path = local_path.relative_to(local_dir)
|
|
75
|
+
s3_path = Path(remote_prefix) / relative_path
|
|
76
|
+
s3_key = str(s3_path).replace("\\", "/") # Normalize for S3
|
|
77
|
+
file_entries.append((local_path, s3_key))
|
|
78
|
+
|
|
79
|
+
session = aioboto3.Session()
|
|
80
|
+
try:
|
|
81
|
+
if self.s3_config.endpoint_url:
|
|
82
|
+
region_name = None
|
|
83
|
+
endpoint_url = self.s3_config.endpoint_url
|
|
84
|
+
elif self.s3_config.region_name:
|
|
85
|
+
region_name = self.s3_config.region_name
|
|
86
|
+
endpoint_url = None
|
|
87
|
+
else:
|
|
88
|
+
raise ValueError("No endpoint_url or region_name provided in the config: eval.general.output.s3")
|
|
89
|
+
async with session.client(
|
|
90
|
+
"s3",
|
|
91
|
+
endpoint_url=endpoint_url,
|
|
92
|
+
region_name=region_name,
|
|
93
|
+
aws_access_key_id=self.s3_config.access_key,
|
|
94
|
+
aws_secret_access_key=self.s3_config.secret_key,
|
|
95
|
+
) as s3_client:
|
|
96
|
+
with tqdm(total=len(file_entries), desc="Uploading files to S3") as pbar:
|
|
97
|
+
upload_tasks = [
|
|
98
|
+
self._upload_file(s3_client, bucket, s3_key, local_path, pbar)
|
|
99
|
+
for local_path, s3_key in file_entries
|
|
100
|
+
]
|
|
101
|
+
await asyncio.gather(*upload_tasks)
|
|
102
|
+
|
|
103
|
+
except NoCredentialsError as e:
|
|
104
|
+
logger.error("AWS credentials not available: %s", e)
|
|
105
|
+
raise
|
|
106
|
+
except Exception as e:
|
|
107
|
+
logger.error("Failed to upload files to S3: %s", e)
|
|
108
|
+
raise
|
|
109
|
+
|
|
110
|
+
def run_custom_scripts(self):
|
|
111
|
+
"""
|
|
112
|
+
Run custom Python scripts defined in the EvalOutputConfig.
|
|
113
|
+
Each script is run with its kwargs passed as command-line arguments.
|
|
114
|
+
The output directory is passed as the first argument.
|
|
115
|
+
"""
|
|
116
|
+
for _, script_config in self.output_config.custom_scripts.items():
|
|
117
|
+
script_path = script_config.script
|
|
118
|
+
if not script_path.exists():
|
|
119
|
+
logger.error("Custom script %s does not exist.", script_path)
|
|
120
|
+
continue
|
|
121
|
+
|
|
122
|
+
# use python interpreter
|
|
123
|
+
args = [sys.executable, str(script_path)]
|
|
124
|
+
# add output directory as first keyword argument
|
|
125
|
+
args.append("--output_dir")
|
|
126
|
+
args.append(str(self.output_config.dir))
|
|
127
|
+
if script_config.kwargs:
|
|
128
|
+
for key, value in script_config.kwargs.items():
|
|
129
|
+
args.append(f"--{key}")
|
|
130
|
+
args.append(str(value))
|
|
131
|
+
|
|
132
|
+
display_args = " ".join(f'"{arg}"' if " " in arg else arg for arg in args[1:])
|
|
133
|
+
|
|
134
|
+
try:
|
|
135
|
+
logger.info("Running custom script: %s %s", script_path, display_args)
|
|
136
|
+
subprocess.run(args, check=True, text=True)
|
|
137
|
+
logger.info("Custom script %s completed successfully.", script_path)
|
|
138
|
+
except subprocess.CalledProcessError as e:
|
|
139
|
+
logger.error("Custom script %s failed with return code %s", script_path, e.returncode)
|
|
140
|
+
raise
|
|
@@ -0,0 +1,40 @@
|
|
|
1
|
+
# SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
|
|
2
|
+
# SPDX-License-Identifier: Apache-2.0
|
|
3
|
+
#
|
|
4
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
5
|
+
# you may not use this file except in compliance with the License.
|
|
6
|
+
# You may obtain a copy of the License at
|
|
7
|
+
#
|
|
8
|
+
# http://www.apache.org/licenses/LICENSE-2.0
|
|
9
|
+
#
|
|
10
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
11
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
12
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
13
|
+
# See the License for the specific language governing permissions and
|
|
14
|
+
# limitations under the License.
|
|
15
|
+
|
|
16
|
+
|
|
17
|
+
class TqdmPositionRegistry:
|
|
18
|
+
"""
|
|
19
|
+
A simple registry for tqdm positions.
|
|
20
|
+
"""
|
|
21
|
+
_positions = set()
|
|
22
|
+
_max_positions = 100
|
|
23
|
+
|
|
24
|
+
@classmethod
|
|
25
|
+
def claim(cls) -> int:
|
|
26
|
+
"""
|
|
27
|
+
Claim a tqdm position in the range of 0-99.
|
|
28
|
+
"""
|
|
29
|
+
for i in range(cls._max_positions):
|
|
30
|
+
if i not in cls._positions:
|
|
31
|
+
cls._positions.add(i)
|
|
32
|
+
return i
|
|
33
|
+
raise RuntimeError("No available tqdm positions.")
|
|
34
|
+
|
|
35
|
+
@classmethod
|
|
36
|
+
def release(cls, pos: int):
|
|
37
|
+
"""
|
|
38
|
+
Release a tqdm position.
|
|
39
|
+
"""
|
|
40
|
+
cls._positions.discard(pos)
|
|
@@ -0,0 +1,193 @@
|
|
|
1
|
+
# SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
|
|
2
|
+
# SPDX-License-Identifier: Apache-2.0
|
|
3
|
+
#
|
|
4
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
5
|
+
# you may not use this file except in compliance with the License.
|
|
6
|
+
# You may obtain a copy of the License at
|
|
7
|
+
#
|
|
8
|
+
# http://www.apache.org/licenses/LICENSE-2.0
|
|
9
|
+
#
|
|
10
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
11
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
12
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
13
|
+
# See the License for the specific language governing permissions and
|
|
14
|
+
# limitations under the License.
|
|
15
|
+
|
|
16
|
+
import asyncio
|
|
17
|
+
import logging
|
|
18
|
+
from typing import TYPE_CHECKING
|
|
19
|
+
from typing import Any
|
|
20
|
+
|
|
21
|
+
from nat.eval.evaluator.evaluator_model import EvalInput
|
|
22
|
+
from nat.eval.evaluator.evaluator_model import EvalInputItem
|
|
23
|
+
from nat.eval.evaluator.evaluator_model import EvalOutput
|
|
24
|
+
from nat.eval.usage_stats import UsageStats
|
|
25
|
+
from nat.eval.usage_stats import UsageStatsItem
|
|
26
|
+
from nat.profiler.data_models import ProfilerResults
|
|
27
|
+
|
|
28
|
+
if TYPE_CHECKING:
|
|
29
|
+
from nat.eval.utils.eval_trace_ctx import EvalTraceContext
|
|
30
|
+
|
|
31
|
+
logger = logging.getLogger(__name__)
|
|
32
|
+
|
|
33
|
+
|
|
34
|
+
class WeaveEvaluationIntegration:
|
|
35
|
+
"""
|
|
36
|
+
Class to handle all Weave integration functionality.
|
|
37
|
+
"""
|
|
38
|
+
|
|
39
|
+
def __init__(self, eval_trace_context: "EvalTraceContext"):
|
|
40
|
+
self.available = False
|
|
41
|
+
self.client = None
|
|
42
|
+
self.eval_logger = None
|
|
43
|
+
self.pred_loggers = {}
|
|
44
|
+
self.eval_trace_context = eval_trace_context
|
|
45
|
+
|
|
46
|
+
try:
|
|
47
|
+
from weave import EvaluationLogger
|
|
48
|
+
from weave.trace.context import weave_client_context
|
|
49
|
+
self.evaluation_logger_cls = EvaluationLogger
|
|
50
|
+
self.weave_client_context = weave_client_context
|
|
51
|
+
self.available = True
|
|
52
|
+
except ImportError:
|
|
53
|
+
self.available = False
|
|
54
|
+
# we simply don't do anything if weave is not available
|
|
55
|
+
pass
|
|
56
|
+
|
|
57
|
+
def initialize_client(self):
|
|
58
|
+
"""Initialize the Weave client if available."""
|
|
59
|
+
if not self.available:
|
|
60
|
+
return False
|
|
61
|
+
|
|
62
|
+
try:
|
|
63
|
+
self.client = self.weave_client_context.require_weave_client()
|
|
64
|
+
return self.client is not None
|
|
65
|
+
except Exception:
|
|
66
|
+
self.client = None
|
|
67
|
+
return False
|
|
68
|
+
|
|
69
|
+
def _get_prediction_inputs(self, item: EvalInputItem):
|
|
70
|
+
"""Get the inputs for displaying in the UI.
|
|
71
|
+
The following fields are excluded as they are too large to display in the UI:
|
|
72
|
+
- full_dataset_entry
|
|
73
|
+
- expected_trajectory
|
|
74
|
+
- trajectory
|
|
75
|
+
|
|
76
|
+
output_obj is excluded because it is displayed separately.
|
|
77
|
+
"""
|
|
78
|
+
include = {"id", "input_obj", "expected_output_obj"}
|
|
79
|
+
return item.model_dump(include=include)
|
|
80
|
+
|
|
81
|
+
def _get_weave_dataset(self, eval_input: EvalInput):
|
|
82
|
+
"""Get the full dataset for Weave."""
|
|
83
|
+
return [item.full_dataset_entry for item in eval_input.eval_input_items]
|
|
84
|
+
|
|
85
|
+
def initialize_logger(self, workflow_alias: str, eval_input: EvalInput, config: Any):
|
|
86
|
+
"""Initialize the Weave evaluation logger."""
|
|
87
|
+
if not self.client and not self.initialize_client():
|
|
88
|
+
# lazy init the client
|
|
89
|
+
return False
|
|
90
|
+
|
|
91
|
+
try:
|
|
92
|
+
weave_dataset = self._get_weave_dataset(eval_input)
|
|
93
|
+
config_dict = config.model_dump(mode="json")
|
|
94
|
+
config_dict["name"] = workflow_alias
|
|
95
|
+
self.eval_logger = self.evaluation_logger_cls(model=config_dict,
|
|
96
|
+
dataset=weave_dataset,
|
|
97
|
+
name=workflow_alias,
|
|
98
|
+
eval_attributes={})
|
|
99
|
+
self.pred_loggers = {}
|
|
100
|
+
|
|
101
|
+
# Capture the current evaluation call for context propagation
|
|
102
|
+
self.eval_trace_context.set_eval_call(self.eval_logger._evaluate_call)
|
|
103
|
+
|
|
104
|
+
return True
|
|
105
|
+
except Exception as e:
|
|
106
|
+
self.eval_logger = None
|
|
107
|
+
logger.warning("Failed to initialize Weave `EvaluationLogger`: %s", e)
|
|
108
|
+
|
|
109
|
+
return False
|
|
110
|
+
|
|
111
|
+
def log_prediction(self, item: EvalInputItem, output: Any):
|
|
112
|
+
"""Log a prediction to Weave."""
|
|
113
|
+
if not self.eval_logger:
|
|
114
|
+
return
|
|
115
|
+
|
|
116
|
+
pred_logger = self.eval_logger.log_prediction(inputs=self._get_prediction_inputs(item), output=output)
|
|
117
|
+
self.pred_loggers[item.id] = pred_logger
|
|
118
|
+
|
|
119
|
+
async def log_usage_stats(self, item: EvalInputItem, usage_stats_item: UsageStatsItem):
|
|
120
|
+
"""Log usage stats to Weave."""
|
|
121
|
+
if not self.eval_logger:
|
|
122
|
+
return
|
|
123
|
+
|
|
124
|
+
# log each usage stat as a score
|
|
125
|
+
await self.pred_loggers[item.id].alog_score(scorer="wf_runtime", score=usage_stats_item.runtime)
|
|
126
|
+
|
|
127
|
+
# log the total tokens for this item, per-llm tokens can be exported later if needed
|
|
128
|
+
await self.pred_loggers[item.id].alog_score(scorer="wf_tokens", score=usage_stats_item.total_tokens)
|
|
129
|
+
|
|
130
|
+
async def alog_score(self, eval_output: EvalOutput, evaluator_name: str):
|
|
131
|
+
"""Log scores for evaluation outputs."""
|
|
132
|
+
if not self.eval_logger:
|
|
133
|
+
return
|
|
134
|
+
|
|
135
|
+
# Create coroutines for all score logging operations
|
|
136
|
+
coros = []
|
|
137
|
+
for eval_output_item in eval_output.eval_output_items:
|
|
138
|
+
if eval_output_item.id in self.pred_loggers:
|
|
139
|
+
coros.append(self.pred_loggers[eval_output_item.id].alog_score(
|
|
140
|
+
scorer=evaluator_name,
|
|
141
|
+
score=eval_output_item.score,
|
|
142
|
+
))
|
|
143
|
+
|
|
144
|
+
# Execute all coroutines concurrently
|
|
145
|
+
if coros:
|
|
146
|
+
await asyncio.gather(*coros)
|
|
147
|
+
|
|
148
|
+
async def afinish_loggers(self):
|
|
149
|
+
"""Finish all prediction loggers and wait for exports."""
|
|
150
|
+
if not self.eval_logger:
|
|
151
|
+
return
|
|
152
|
+
|
|
153
|
+
async def _finish_one(pred_logger):
|
|
154
|
+
if hasattr(pred_logger, '_has_finished') and not pred_logger._has_finished:
|
|
155
|
+
return
|
|
156
|
+
# run the *blocking* finish() in a thread so we don't nest loops
|
|
157
|
+
await asyncio.to_thread(pred_logger.finish)
|
|
158
|
+
|
|
159
|
+
await asyncio.gather(*[_finish_one(pl) for pl in self.pred_loggers.values()])
|
|
160
|
+
|
|
161
|
+
def _log_profiler_metrics(self, profiler_results: ProfilerResults, usage_stats: UsageStats) -> dict[str, Any]:
|
|
162
|
+
"""Log profiler metrics to Weave."""
|
|
163
|
+
profile_metrics = {}
|
|
164
|
+
if profiler_results.llm_latency_ci:
|
|
165
|
+
profile_metrics["llm_latency_p95"] = profiler_results.llm_latency_ci.p95
|
|
166
|
+
if profiler_results.workflow_runtime_metrics:
|
|
167
|
+
profile_metrics["wf_runtime_p95"] = profiler_results.workflow_runtime_metrics.p95
|
|
168
|
+
|
|
169
|
+
profile_metrics["total_runtime"] = usage_stats.total_runtime
|
|
170
|
+
|
|
171
|
+
return profile_metrics
|
|
172
|
+
|
|
173
|
+
def log_summary(self,
|
|
174
|
+
usage_stats: UsageStats,
|
|
175
|
+
evaluation_results: list[tuple[str, EvalOutput]],
|
|
176
|
+
profiler_results: ProfilerResults):
|
|
177
|
+
"""Log summary statistics to Weave."""
|
|
178
|
+
if not self.eval_logger:
|
|
179
|
+
return
|
|
180
|
+
|
|
181
|
+
summary = {}
|
|
182
|
+
# add evaluation results to the summary
|
|
183
|
+
for evaluator_name, eval_output in evaluation_results:
|
|
184
|
+
summary[evaluator_name] = eval_output.average_score
|
|
185
|
+
|
|
186
|
+
# add profiler metrics to the summary
|
|
187
|
+
profile_metrics = self._log_profiler_metrics(profiler_results, usage_stats)
|
|
188
|
+
summary.update(profile_metrics)
|
|
189
|
+
|
|
190
|
+
# Log the summary to finish the evaluation, disable auto-summarize
|
|
191
|
+
# as we will be adding profiler metrics to the summary
|
|
192
|
+
self.eval_logger.log_summary(summary, auto_summarize=False)
|
|
193
|
+
logger.info("Logged Evaluation Summary to Weave")
|
|
File without changes
|
|
File without changes
|
|
@@ -0,0 +1,154 @@
|
|
|
1
|
+
# SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
|
|
2
|
+
# SPDX-License-Identifier: Apache-2.0
|
|
3
|
+
#
|
|
4
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
5
|
+
# you may not use this file except in compliance with the License.
|
|
6
|
+
# You may obtain a copy of the License at
|
|
7
|
+
#
|
|
8
|
+
# http://www.apache.org/licenses/LICENSE-2.0
|
|
9
|
+
#
|
|
10
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
11
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
12
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
13
|
+
# See the License for the specific language governing permissions and
|
|
14
|
+
# limitations under the License.
|
|
15
|
+
|
|
16
|
+
import functools
|
|
17
|
+
import inspect
|
|
18
|
+
import logging
|
|
19
|
+
from collections.abc import AsyncGenerator
|
|
20
|
+
from collections.abc import Callable
|
|
21
|
+
from collections.abc import Generator
|
|
22
|
+
from typing import Any
|
|
23
|
+
from typing import TypeVar
|
|
24
|
+
from typing import overload
|
|
25
|
+
|
|
26
|
+
logger = logging.getLogger(__name__)
|
|
27
|
+
|
|
28
|
+
BASE_WARNING_MESSAGE = ("is experimental and the API may change in future releases. "
|
|
29
|
+
"Future versions may introduce breaking changes without notice.")
|
|
30
|
+
|
|
31
|
+
_warning_issued = set()
|
|
32
|
+
|
|
33
|
+
# Type variables for overloads
|
|
34
|
+
F = TypeVar('F', bound=Callable[..., Any])
|
|
35
|
+
|
|
36
|
+
|
|
37
|
+
def issue_experimental_warning(function_name: str,
|
|
38
|
+
feature_name: str | None = None,
|
|
39
|
+
metadata: dict[str, Any] | None = None):
|
|
40
|
+
"""
|
|
41
|
+
Log a warning message that the function is experimental.
|
|
42
|
+
|
|
43
|
+
A warning is emitted only once per function. When a ``metadata`` dict
|
|
44
|
+
is supplied, it is appended to the log entry to provide extra context
|
|
45
|
+
(e.g., version, author, feature flag).
|
|
46
|
+
"""
|
|
47
|
+
if function_name not in _warning_issued:
|
|
48
|
+
if (feature_name):
|
|
49
|
+
warning_message = f"The {feature_name} feature {BASE_WARNING_MESSAGE}"
|
|
50
|
+
else:
|
|
51
|
+
warning_message = f"This function {BASE_WARNING_MESSAGE}"
|
|
52
|
+
|
|
53
|
+
warning_message += f" Function: {function_name}"
|
|
54
|
+
|
|
55
|
+
if (metadata):
|
|
56
|
+
warning_message += f" | Metadata: {metadata}"
|
|
57
|
+
|
|
58
|
+
# Issue warning and save function name to avoid duplicate warnings
|
|
59
|
+
logger.warning(warning_message)
|
|
60
|
+
|
|
61
|
+
_warning_issued.add(function_name)
|
|
62
|
+
|
|
63
|
+
|
|
64
|
+
# Overloads for different function types
|
|
65
|
+
@overload
|
|
66
|
+
def experimental(func: F, *, feature_name: str | None = None, metadata: dict[str, Any] | None = None) -> F:
|
|
67
|
+
"""Overload for when a function is passed directly."""
|
|
68
|
+
...
|
|
69
|
+
|
|
70
|
+
|
|
71
|
+
@overload
|
|
72
|
+
def experimental(*, feature_name: str | None = None, metadata: dict[str, Any] | None = None) -> Callable[[F], F]:
|
|
73
|
+
"""Overload for decorator factory usage (when called with parentheses)."""
|
|
74
|
+
...
|
|
75
|
+
|
|
76
|
+
|
|
77
|
+
def experimental(func: Any = None, *, feature_name: str | None = None, metadata: dict[str, Any] | None = None) -> Any:
|
|
78
|
+
"""
|
|
79
|
+
Decorator that can wrap any type of function (sync, async, generator,
|
|
80
|
+
async generator) and logs a warning that the function is experimental.
|
|
81
|
+
|
|
82
|
+
Args:
|
|
83
|
+
func: The function to be decorated.
|
|
84
|
+
feature_name: Optional name of the feature that is experimental. If provided, the warning will be
|
|
85
|
+
prefixed with "The <feature_name> feature is experimental".
|
|
86
|
+
metadata: Optional dictionary of metadata to log with the warning. This can include information
|
|
87
|
+
like version, author, etc. If provided, the metadata will be
|
|
88
|
+
logged alongside the experimental warning.
|
|
89
|
+
"""
|
|
90
|
+
function_name: str = f"{func.__module__}.{func.__qualname__}" if func else "<unknown_function>"
|
|
91
|
+
|
|
92
|
+
# If called as @track_function(...) but not immediately passed a function
|
|
93
|
+
if func is None:
|
|
94
|
+
|
|
95
|
+
def decorator_wrapper(actual_func):
|
|
96
|
+
return experimental(actual_func, feature_name=feature_name, metadata=metadata)
|
|
97
|
+
|
|
98
|
+
return decorator_wrapper
|
|
99
|
+
|
|
100
|
+
# --- Validate metadata ---
|
|
101
|
+
if metadata is not None:
|
|
102
|
+
if not isinstance(metadata, dict):
|
|
103
|
+
raise TypeError("metadata must be a dict[str, Any].")
|
|
104
|
+
if any(not isinstance(k, str) for k in metadata.keys()):
|
|
105
|
+
raise TypeError("All metadata keys must be strings.")
|
|
106
|
+
|
|
107
|
+
# --- Now detect the function type and wrap accordingly ---
|
|
108
|
+
if inspect.isasyncgenfunction(func):
|
|
109
|
+
# ---------------------
|
|
110
|
+
# ASYNC GENERATOR
|
|
111
|
+
# ---------------------
|
|
112
|
+
|
|
113
|
+
@functools.wraps(func)
|
|
114
|
+
async def async_gen_wrapper(*args, **kwargs) -> AsyncGenerator[Any, Any]:
|
|
115
|
+
issue_experimental_warning(function_name, feature_name, metadata)
|
|
116
|
+
async for item in func(*args, **kwargs):
|
|
117
|
+
yield item # yield the original item
|
|
118
|
+
|
|
119
|
+
return async_gen_wrapper
|
|
120
|
+
|
|
121
|
+
if inspect.iscoroutinefunction(func):
|
|
122
|
+
# ---------------------
|
|
123
|
+
# ASYNC FUNCTION
|
|
124
|
+
# ---------------------
|
|
125
|
+
@functools.wraps(func)
|
|
126
|
+
async def async_wrapper(*args, **kwargs) -> Any:
|
|
127
|
+
issue_experimental_warning(function_name, feature_name, metadata)
|
|
128
|
+
result = await func(*args, **kwargs)
|
|
129
|
+
return result
|
|
130
|
+
|
|
131
|
+
return async_wrapper
|
|
132
|
+
|
|
133
|
+
if inspect.isgeneratorfunction(func):
|
|
134
|
+
# ---------------------
|
|
135
|
+
# SYNC GENERATOR
|
|
136
|
+
# ---------------------
|
|
137
|
+
@functools.wraps(func)
|
|
138
|
+
def sync_gen_wrapper(*args, **kwargs) -> Generator[Any, Any, Any]:
|
|
139
|
+
issue_experimental_warning(function_name, feature_name, metadata)
|
|
140
|
+
yield from func(*args, **kwargs) # yield the original item
|
|
141
|
+
|
|
142
|
+
return sync_gen_wrapper
|
|
143
|
+
|
|
144
|
+
@functools.wraps(func)
|
|
145
|
+
def sync_wrapper(*args, **kwargs) -> Any:
|
|
146
|
+
issue_experimental_warning(function_name, feature_name, metadata)
|
|
147
|
+
result = func(*args, **kwargs)
|
|
148
|
+
return result
|
|
149
|
+
|
|
150
|
+
return sync_wrapper
|
|
151
|
+
|
|
152
|
+
|
|
153
|
+
# Compatibility aliases with previous releases
|
|
154
|
+
aiq_experimental = experimental
|
|
File without changes
|
|
File without changes
|
|
@@ -0,0 +1,147 @@
|
|
|
1
|
+
# SPDX-FileCopyrightText: Copyright (c) 2025, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
|
|
2
|
+
# SPDX-License-Identifier: Apache-2.0
|
|
3
|
+
#
|
|
4
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
5
|
+
# you may not use this file except in compliance with the License.
|
|
6
|
+
# You may obtain a copy of the License at
|
|
7
|
+
#
|
|
8
|
+
# http://www.apache.org/licenses/LICENSE-2.0
|
|
9
|
+
#
|
|
10
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
11
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
12
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
13
|
+
# See the License for the specific language governing permissions and
|
|
14
|
+
# limitations under the License.
|
|
15
|
+
|
|
16
|
+
import asyncio
|
|
17
|
+
import logging
|
|
18
|
+
import re
|
|
19
|
+
|
|
20
|
+
from nat.builder.builder import Builder
|
|
21
|
+
from nat.builder.framework_enum import LLMFrameworkEnum
|
|
22
|
+
from nat.cli.register_workflow import register_ttc_strategy
|
|
23
|
+
from nat.data_models.ttc_strategy import TTCStrategyBaseConfig
|
|
24
|
+
from nat.experimental.test_time_compute.models.editor_config import IterativePlanRefinementConfig
|
|
25
|
+
from nat.experimental.test_time_compute.models.stage_enums import PipelineTypeEnum
|
|
26
|
+
from nat.experimental.test_time_compute.models.stage_enums import StageTypeEnum
|
|
27
|
+
from nat.experimental.test_time_compute.models.strategy_base import StrategyBase
|
|
28
|
+
from nat.experimental.test_time_compute.models.ttc_item import TTCItem
|
|
29
|
+
from nat.utils.io.model_processing import remove_r1_think_tags
|
|
30
|
+
|
|
31
|
+
logger = logging.getLogger(__name__)
|
|
32
|
+
|
|
33
|
+
|
|
34
|
+
class IterativePlanRefinementEditor(StrategyBase):
|
|
35
|
+
"""
|
|
36
|
+
A planner that generates an initial plan, then refines it multiple times
|
|
37
|
+
using the same LLM. Each iteration updates the plan to (hopefully) be better.
|
|
38
|
+
"""
|
|
39
|
+
|
|
40
|
+
def __init__(self, config: TTCStrategyBaseConfig) -> None:
|
|
41
|
+
super().__init__(config)
|
|
42
|
+
self.llm_bound = None
|
|
43
|
+
|
|
44
|
+
def supported_pipeline_types(self) -> [PipelineTypeEnum]:
|
|
45
|
+
return [PipelineTypeEnum.PLANNING]
|
|
46
|
+
|
|
47
|
+
def stage_type(self) -> StageTypeEnum:
|
|
48
|
+
return StageTypeEnum.EDITING
|
|
49
|
+
|
|
50
|
+
async def build_components(self, builder: Builder) -> None:
|
|
51
|
+
"""
|
|
52
|
+
Build the components required for the iterative planner.
|
|
53
|
+
"""
|
|
54
|
+
logger.debug("Building components for IterativePlanRefinementEditor")
|
|
55
|
+
self.llm_bound = await builder.get_llm(self.config.editor_llm, wrapper_type=LLMFrameworkEnum.LANGCHAIN)
|
|
56
|
+
|
|
57
|
+
async def refine_single(self, prompt: str, context: str, ttc_item: TTCItem, prompt_idx: int) -> TTCItem:
|
|
58
|
+
from langchain_core.language_models import BaseChatModel
|
|
59
|
+
from langchain_core.prompts import PromptTemplate
|
|
60
|
+
|
|
61
|
+
if not isinstance(self.llm_bound, BaseChatModel):
|
|
62
|
+
raise ValueError("editor_llm must be a BaseChatModel instance for iterative plan refinement.")
|
|
63
|
+
|
|
64
|
+
llm: BaseChatModel = self.llm_bound
|
|
65
|
+
|
|
66
|
+
# Refinement loop
|
|
67
|
+
refinement_template = PromptTemplate(
|
|
68
|
+
template=self.config.refinement_template,
|
|
69
|
+
input_variables=["current_plan", "context", "original_prompt"],
|
|
70
|
+
validate_template=True,
|
|
71
|
+
)
|
|
72
|
+
|
|
73
|
+
current_plan = ttc_item.plan
|
|
74
|
+
for iteration in range(1, self.config.num_iterations + 1):
|
|
75
|
+
logger.info("Refinement iteration %d / %d for prompt %d", iteration, self.config.num_iterations, prompt_idx)
|
|
76
|
+
refine_prompt = (await refinement_template.ainvoke({
|
|
77
|
+
"current_plan": current_plan, "context": context, "original_prompt": prompt
|
|
78
|
+
})).to_string()
|
|
79
|
+
|
|
80
|
+
refine_response = await llm.ainvoke(refine_prompt)
|
|
81
|
+
refined_plan = remove_r1_think_tags(
|
|
82
|
+
refine_response.content if hasattr(refine_response, 'content') else str(refine_response))
|
|
83
|
+
refined_plan = re.sub(r'(?i)^\s*EDITED PLAN:\s*', '', refined_plan).strip()
|
|
84
|
+
if refined_plan:
|
|
85
|
+
current_plan = refined_plan
|
|
86
|
+
else:
|
|
87
|
+
logger.warning("Refinement iteration %d for prompt %d produced an empty plan; keeping existing plan.",
|
|
88
|
+
iteration,
|
|
89
|
+
prompt_idx)
|
|
90
|
+
|
|
91
|
+
logger.info("IterativePlanRefinementPlanner produced a final plan after %d iterations.",
|
|
92
|
+
self.config.num_iterations)
|
|
93
|
+
|
|
94
|
+
ttc_item.plan = current_plan
|
|
95
|
+
# Return a single final plan
|
|
96
|
+
return ttc_item
|
|
97
|
+
|
|
98
|
+
async def ainvoke(self,
|
|
99
|
+
items: list[TTCItem],
|
|
100
|
+
original_prompt: str | None = None,
|
|
101
|
+
agent_context: str | None = None,
|
|
102
|
+
**kwargs) -> list[TTCItem]:
|
|
103
|
+
"""
|
|
104
|
+
Runs the iterative plan refinement process on the provided planning items.
|
|
105
|
+
|
|
106
|
+
Each planning item is refined in parallel the configured number of times. Default is 3.
|
|
107
|
+
|
|
108
|
+
Args:
|
|
109
|
+
items (list[TTCItem]): The planning items to refine.
|
|
110
|
+
original_prompt (str): The original prompt used to generate the plans.
|
|
111
|
+
agent_context (str): The context for the agent.
|
|
112
|
+
|
|
113
|
+
Returns:
|
|
114
|
+
list[TTCItem]: The refined planning items.
|
|
115
|
+
"""
|
|
116
|
+
|
|
117
|
+
if not original_prompt or not agent_context:
|
|
118
|
+
raise ValueError("Arguments original_prompt and agent_context must be provdied.")
|
|
119
|
+
|
|
120
|
+
# Generate feedback for each planning item concurrently
|
|
121
|
+
tasks = [
|
|
122
|
+
self.refine_single(prompt=original_prompt, context=agent_context, ttc_item=item, prompt_idx=i + 1)
|
|
123
|
+
for i, item in enumerate(items)
|
|
124
|
+
]
|
|
125
|
+
|
|
126
|
+
# Run the tasks concurrently and gather results
|
|
127
|
+
refined_planning_items = await asyncio.gather(*tasks)
|
|
128
|
+
|
|
129
|
+
return refined_planning_items
|
|
130
|
+
|
|
131
|
+
|
|
132
|
+
@register_ttc_strategy(config_type=IterativePlanRefinementConfig)
|
|
133
|
+
async def register_iterative_plan_refinement_editor(config: IterativePlanRefinementConfig, builder: Builder):
|
|
134
|
+
"""
|
|
135
|
+
Register the IterativePlanRefinementEditor strategy.
|
|
136
|
+
|
|
137
|
+
Args:
|
|
138
|
+
config (IterativePlanRefinementConfig): The configuration for the strategy.
|
|
139
|
+
|
|
140
|
+
Returns:
|
|
141
|
+
IterativePlanRefinementEditor: The registered strategy instance.
|
|
142
|
+
"""
|
|
143
|
+
|
|
144
|
+
editor = IterativePlanRefinementEditor(config)
|
|
145
|
+
await editor.build_components(builder=builder)
|
|
146
|
+
|
|
147
|
+
yield editor
|