unique_toolkit 1.43.11__py3-none-any.whl → 1.45.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3,7 +3,10 @@ from unique_toolkit.agentic.loop_runner._iteration_handler_utils import (
3
3
  handle_last_iteration,
4
4
  handle_normal_iteration,
5
5
  )
6
- from unique_toolkit.agentic.loop_runner.base import LoopIterationRunner
6
+ from unique_toolkit.agentic.loop_runner.base import (
7
+ LoopIterationRunner,
8
+ ResponsesLoopIterationRunner,
9
+ )
7
10
  from unique_toolkit.agentic.loop_runner.middleware import (
8
11
  PlanningConfig,
9
12
  PlanningMiddleware,
@@ -16,6 +19,7 @@ from unique_toolkit.agentic.loop_runner.runners import (
16
19
  BasicLoopIterationRunner,
17
20
  BasicLoopIterationRunnerConfig,
18
21
  QwenLoopIterationRunner,
22
+ ResponsesBasicLoopIterationRunner,
19
23
  is_qwen_model,
20
24
  )
21
25
 
@@ -34,4 +38,6 @@ __all__ = [
34
38
  "QWEN_FORCED_TOOL_CALL_INSTRUCTION",
35
39
  "QWEN_LAST_ITERATION_INSTRUCTION",
36
40
  "QWEN_MAX_LOOP_ITERATIONS",
41
+ "ResponsesLoopIterationRunner",
42
+ "ResponsesBasicLoopIterationRunner",
37
43
  ]
@@ -0,0 +1,63 @@
1
+ import logging
2
+ from typing import Unpack
3
+
4
+ from unique_toolkit.agentic.loop_runner._responses_stream_handler_utils import (
5
+ responses_stream_response,
6
+ )
7
+ from unique_toolkit.agentic.loop_runner.base import (
8
+ _ResponsesLoopIterationRunnerKwargs,
9
+ )
10
+ from unique_toolkit.language_model.schemas import ResponsesLanguageModelStreamResponse
11
+
12
+ _LOGGER = logging.getLogger(__name__)
13
+
14
+
15
+ async def handle_responses_last_iteration(
16
+ **kwargs: Unpack[_ResponsesLoopIterationRunnerKwargs],
17
+ ) -> ResponsesLanguageModelStreamResponse:
18
+ _LOGGER.info("Reached last iteration, removing tools and producing final response")
19
+
20
+ return await responses_stream_response(
21
+ loop_runner_kwargs=kwargs,
22
+ tools=None,
23
+ )
24
+
25
+
26
+ async def handle_responses_normal_iteration(
27
+ **kwargs: Unpack[_ResponsesLoopIterationRunnerKwargs],
28
+ ) -> ResponsesLanguageModelStreamResponse:
29
+ _LOGGER.info("Running loop iteration %d", kwargs["iteration_index"])
30
+
31
+ return await responses_stream_response(loop_runner_kwargs=kwargs)
32
+
33
+
34
+ async def handle_responses_forced_tools_iteration(
35
+ **kwargs: Unpack[_ResponsesLoopIterationRunnerKwargs],
36
+ ) -> ResponsesLanguageModelStreamResponse:
37
+ assert "tool_choices" in kwargs
38
+
39
+ tool_choices = kwargs["tool_choices"]
40
+ assert len(tool_choices) > 0
41
+
42
+ _LOGGER.info("Forcing tools calls: %s", tool_choices)
43
+
44
+ responses: list[ResponsesLanguageModelStreamResponse] = []
45
+
46
+ for opt in tool_choices:
47
+ responses.append(
48
+ await responses_stream_response(loop_runner_kwargs=kwargs, tool_choice=opt)
49
+ )
50
+
51
+ # Merge responses and refs:
52
+ tool_calls = []
53
+ references = []
54
+ for r in responses:
55
+ if r.tool_calls:
56
+ tool_calls.extend(r.tool_calls)
57
+ references.extend(r.message.references)
58
+
59
+ response = responses[0]
60
+ response.tool_calls = tool_calls if len(tool_calls) > 0 else None
61
+ response.message.references = references
62
+
63
+ return response
@@ -0,0 +1,90 @@
1
+ from typing import Any, Required, Sequence
2
+
3
+ from openai.types.responses import (
4
+ ResponseIncludable,
5
+ ResponseInputItemParam,
6
+ ResponseOutputItem,
7
+ ResponseTextConfigParam,
8
+ ToolParam,
9
+ response_create_params,
10
+ )
11
+ from openai.types.shared_params import Metadata, Reasoning
12
+ from typing_extensions import TypedDict
13
+
14
+ from unique_toolkit import LanguageModelToolDescription
15
+ from unique_toolkit.agentic.loop_runner.base import _ResponsesLoopIterationRunnerKwargs
16
+ from unique_toolkit.chat.service import LanguageModelMessages
17
+ from unique_toolkit.content import ContentChunk
18
+ from unique_toolkit.language_model.schemas import (
19
+ LanguageModelMessageOptions,
20
+ ResponsesLanguageModelStreamResponse,
21
+ )
22
+
23
+
24
+ class _ResponsesStreamingHandlerKwargs(TypedDict, total=False):
25
+ messages: Required[
26
+ str
27
+ | LanguageModelMessages
28
+ | Sequence[
29
+ ResponseInputItemParam | LanguageModelMessageOptions | ResponseOutputItem
30
+ ]
31
+ ]
32
+ model_name: Required[str]
33
+ tools: Sequence[LanguageModelToolDescription | ToolParam]
34
+ content_chunks: list[ContentChunk]
35
+ start_text: str
36
+ debug_info: dict[str, Any]
37
+ temperature: float
38
+ include: list[ResponseIncludable]
39
+ instructions: str
40
+ max_output_tokens: int
41
+ metadata: Metadata
42
+ parallel_tool_calls: bool
43
+ text: ResponseTextConfigParam
44
+ tool_choice: response_create_params.ToolChoice
45
+ top_p: float
46
+ reasoning: Reasoning
47
+ other_options: dict[str, Any]
48
+
49
+
50
+ def _extract_responses_streaming_kwargs(
51
+ kwargs: _ResponsesLoopIterationRunnerKwargs,
52
+ ) -> _ResponsesStreamingHandlerKwargs:
53
+ res = _ResponsesStreamingHandlerKwargs(
54
+ messages=kwargs["messages"],
55
+ model_name=kwargs["model"].name,
56
+ )
57
+
58
+ for k in [
59
+ "tools",
60
+ "content_chunks",
61
+ "start_text",
62
+ "debug_info",
63
+ "temperature",
64
+ "include",
65
+ "instructions",
66
+ "max_output_tokens",
67
+ "metadata",
68
+ "parallel_tool_calls",
69
+ "text",
70
+ "top_p",
71
+ "reasoning",
72
+ "other_options",
73
+ ]:
74
+ if k in kwargs:
75
+ res[k] = kwargs[k]
76
+
77
+ return res
78
+
79
+
80
+ async def responses_stream_response(
81
+ loop_runner_kwargs: _ResponsesLoopIterationRunnerKwargs,
82
+ **kwargs,
83
+ ) -> ResponsesLanguageModelStreamResponse:
84
+ streaming_handler = loop_runner_kwargs["streaming_handler"]
85
+ streaming_handler_kwargs = _extract_responses_streaming_kwargs(loop_runner_kwargs)
86
+ streaming_handler_kwargs.update(**kwargs)
87
+
88
+ return await streaming_handler.complete_with_references_async(
89
+ **streaming_handler_kwargs
90
+ )
@@ -1,6 +1,15 @@
1
- from typing import Any, Protocol, Required, Unpack
1
+ from typing import Any, Protocol, Required, Sequence, Unpack
2
2
 
3
3
  from openai.types.chat import ChatCompletionNamedToolChoiceParam
4
+ from openai.types.responses import (
5
+ ResponseIncludable,
6
+ ResponseInputItemParam,
7
+ ResponseOutputItem,
8
+ ResponseTextConfigParam,
9
+ ToolParam,
10
+ response_create_params,
11
+ )
12
+ from openai.types.shared_params import Metadata, Reasoning
4
13
  from typing_extensions import TypedDict
5
14
 
6
15
  from unique_toolkit import LanguageModelToolDescription
@@ -8,7 +17,12 @@ from unique_toolkit.chat.functions import LanguageModelStreamResponse
8
17
  from unique_toolkit.chat.service import LanguageModelMessages
9
18
  from unique_toolkit.content import ContentChunk
10
19
  from unique_toolkit.language_model.infos import LanguageModelInfo
20
+ from unique_toolkit.language_model.schemas import (
21
+ LanguageModelMessageOptions,
22
+ ResponsesLanguageModelStreamResponse,
23
+ )
11
24
  from unique_toolkit.protocols.support import (
25
+ ResponsesSupportCompleteWithReferences,
12
26
  SupportCompleteWithReferences,
13
27
  )
14
28
 
@@ -36,3 +50,38 @@ class LoopIterationRunner(Protocol):
36
50
  self,
37
51
  **kwargs: Unpack[_LoopIterationRunnerKwargs],
38
52
  ) -> LanguageModelStreamResponse: ...
53
+
54
+
55
+ class _ResponsesLoopIterationRunnerKwargs(TypedDict, total=False):
56
+ iteration_index: Required[int]
57
+ streaming_handler: Required[ResponsesSupportCompleteWithReferences]
58
+ messages: Required[
59
+ str
60
+ | LanguageModelMessages
61
+ | Sequence[
62
+ ResponseInputItemParam | LanguageModelMessageOptions | ResponseOutputItem
63
+ ]
64
+ ]
65
+ model: Required[LanguageModelInfo]
66
+ tools: list[LanguageModelToolDescription | ToolParam]
67
+ content_chunks: list[ContentChunk]
68
+ start_text: str
69
+ debug_info: dict[str, Any]
70
+ temperature: float
71
+ include: list[ResponseIncludable]
72
+ instructions: str
73
+ max_output_tokens: int
74
+ metadata: Metadata
75
+ parallel_tool_calls: bool
76
+ text: ResponseTextConfigParam
77
+ tool_choices: list[response_create_params.ToolChoice]
78
+ top_p: float
79
+ reasoning: Reasoning
80
+ other_options: dict[str, Any]
81
+
82
+
83
+ class ResponsesLoopIterationRunner(Protocol):
84
+ async def __call__(
85
+ self,
86
+ **kwargs: Unpack[_ResponsesLoopIterationRunnerKwargs],
87
+ ) -> ResponsesLanguageModelStreamResponse: ...
@@ -1,6 +1,7 @@
1
1
  from unique_toolkit.agentic.loop_runner.runners.basic import (
2
2
  BasicLoopIterationRunner,
3
3
  BasicLoopIterationRunnerConfig,
4
+ ResponsesBasicLoopIterationRunner,
4
5
  )
5
6
  from unique_toolkit.agentic.loop_runner.runners.qwen import (
6
7
  QWEN_FORCED_TOOL_CALL_INSTRUCTION,
@@ -13,6 +14,7 @@ from unique_toolkit.agentic.loop_runner.runners.qwen import (
13
14
  __all__ = [
14
15
  "BasicLoopIterationRunnerConfig",
15
16
  "BasicLoopIterationRunner",
17
+ "ResponsesBasicLoopIterationRunner",
16
18
  "QwenLoopIterationRunner",
17
19
  "QWEN_FORCED_TOOL_CALL_INSTRUCTION",
18
20
  "QWEN_LAST_ITERATION_INSTRUCTION",
@@ -9,14 +9,19 @@ from unique_toolkit.agentic.loop_runner._iteration_handler_utils import (
9
9
  handle_last_iteration,
10
10
  handle_normal_iteration,
11
11
  )
12
+ from unique_toolkit.agentic.loop_runner._responses_iteration_handler_utils import (
13
+ handle_responses_forced_tools_iteration,
14
+ handle_responses_last_iteration,
15
+ handle_responses_normal_iteration,
16
+ )
12
17
  from unique_toolkit.agentic.loop_runner.base import (
13
18
  LoopIterationRunner,
19
+ ResponsesLoopIterationRunner,
14
20
  _LoopIterationRunnerKwargs,
21
+ _ResponsesLoopIterationRunnerKwargs,
15
22
  )
16
23
  from unique_toolkit.chat.functions import LanguageModelStreamResponse
17
- from unique_toolkit.protocols.support import (
18
- ResponsesLanguageModelStreamResponse,
19
- )
24
+ from unique_toolkit.language_model.schemas import ResponsesLanguageModelStreamResponse
20
25
 
21
26
  _LOGGER = logging.getLogger(__name__)
22
27
 
@@ -34,7 +39,7 @@ class BasicLoopIterationRunner(LoopIterationRunner):
34
39
  async def __call__(
35
40
  self,
36
41
  **kwargs: Unpack[_LoopIterationRunnerKwargs],
37
- ) -> LanguageModelStreamResponse | ResponsesLanguageModelStreamResponse:
42
+ ) -> LanguageModelStreamResponse:
38
43
  tool_choices = kwargs.get("tool_choices", [])
39
44
  iteration_index = kwargs["iteration_index"]
40
45
 
@@ -44,3 +49,23 @@ class BasicLoopIterationRunner(LoopIterationRunner):
44
49
  return await handle_last_iteration(**kwargs)
45
50
  else:
46
51
  return await handle_normal_iteration(**kwargs)
52
+
53
+
54
+ class ResponsesBasicLoopIterationRunner(ResponsesLoopIterationRunner):
55
+ def __init__(self, config: BasicLoopIterationRunnerConfig) -> None:
56
+ self._config = config
57
+
58
+ @override
59
+ async def __call__(
60
+ self,
61
+ **kwargs: Unpack[_ResponsesLoopIterationRunnerKwargs],
62
+ ) -> ResponsesLanguageModelStreamResponse:
63
+ tool_choices = kwargs.get("tool_choices", [])
64
+ iteration_index = kwargs["iteration_index"]
65
+
66
+ if len(tool_choices) > 0 and iteration_index == 0:
67
+ return await handle_responses_forced_tools_iteration(**kwargs)
68
+ elif iteration_index == self._config.max_loop_iterations - 1:
69
+ return await handle_responses_last_iteration(**kwargs)
70
+ else:
71
+ return await handle_responses_normal_iteration(**kwargs)
@@ -7,8 +7,6 @@ from unique_toolkit.agentic.tools.openai_builtin.base import (
7
7
  )
8
8
  from unique_toolkit.agentic.tools.schemas import BaseToolConfig
9
9
 
10
- DEFAULT_TOOL_DESCRIPTION = "Use this tool to run python code, e.g to generate plots, process excel files, perform calculations, etc."
11
-
12
10
  DEFAULT_TOOL_DESCRIPTION_FOR_SYSTEM_PROMPT = """
13
11
  Use this tool to run python code, e.g to generate plots, process excel files, perform calculations, etc.
14
12
 
@@ -48,11 +46,9 @@ Handling User Queries:
48
46
  - After exhausting all possible solutions without success, inform the user of what was tried and request clarification/help.
49
47
  """.strip()
50
48
 
51
- DEFAULT_TOOL_FORMAT_INFORMATION_FOR_SYSTEM_PROMPT = ""
52
-
53
- DEFAULT_TOOL_FORMAT_INFORMATION_FOR_USER_PROMPT = ""
54
49
 
55
50
  DEFAULT_TOOL_DESCRIPTION_FOR_USER_PROMPT = ""
51
+ DEFAULT_TOOL_DESCRIPTION = "Use this tool to run python code, e.g to generate plots, process excel files, perform calculations, etc."
56
52
 
57
53
 
58
54
  class OpenAICodeInterpreterConfig(BaseToolConfig):
@@ -60,33 +56,20 @@ class OpenAICodeInterpreterConfig(BaseToolConfig):
60
56
  default=True,
61
57
  description="If set, the files uploaded to the chat will be uploaded to the container where code is executed.",
62
58
  )
63
-
64
59
  tool_description: str = Field(
65
60
  default=DEFAULT_TOOL_DESCRIPTION,
66
- description="The description of the tool that will be sent to the model.",
61
+ description="The description of the tool that will be included in the system prompt.",
67
62
  )
68
63
  tool_description_for_system_prompt: str = Field(
69
64
  default=DEFAULT_TOOL_DESCRIPTION_FOR_SYSTEM_PROMPT,
70
65
  description="The description of the tool that will be included in the system prompt.",
71
66
  )
72
- tool_format_information_for_system_prompt: SkipJsonSchema[str] = Field(
73
- # Since the tool is executed in Azure, it's not always possible to have dynamic format information
74
- default=DEFAULT_TOOL_FORMAT_INFORMATION_FOR_SYSTEM_PROMPT,
75
- description="The format information of the tool that will be included in the system prompt.",
76
- )
77
67
  tool_description_for_user_prompt: SkipJsonSchema[str] = (
78
68
  Field( # At the moment, this is not appended to the user prompt
79
69
  default=DEFAULT_TOOL_DESCRIPTION_FOR_USER_PROMPT,
80
70
  description="The description of the tool that will be included in the user prompt.",
81
71
  )
82
72
  )
83
- tool_format_information_for_user_prompt: SkipJsonSchema[str] = (
84
- Field( # At the moment, this is not appended to the user prompt
85
- default=DEFAULT_TOOL_FORMAT_INFORMATION_FOR_USER_PROMPT,
86
- description="The format information of the tool that will be included in the user prompt.",
87
- )
88
- )
89
-
90
73
  expires_after_minutes: int = Field(
91
74
  default=20,
92
75
  description="The number of minutes after which the container will be deleted.",
@@ -239,9 +239,9 @@ class OpenAICodeInterpreterTool(OpenAIBuiltInTool[CodeInterpreter]):
239
239
  display_name=self.DISPLAY_NAME,
240
240
  tool_description=self._config.tool_description,
241
241
  tool_system_prompt=self._config.tool_description_for_system_prompt,
242
- tool_format_information_for_system_prompt=self._config.tool_format_information_for_system_prompt,
242
+ tool_format_information_for_system_prompt="",
243
243
  tool_user_prompt=self._config.tool_description_for_user_prompt,
244
- tool_format_information_for_user_prompt=self._config.tool_format_information_for_user_prompt,
244
+ tool_format_information_for_user_prompt="",
245
245
  input_model={},
246
246
  )
247
247
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: unique_toolkit
3
- Version: 1.43.11
3
+ Version: 1.45.0
4
4
  Summary:
5
5
  License: Proprietary
6
6
  Author: Cedric Klinkert
@@ -125,6 +125,12 @@ All notable changes to this project will be documented in this file.
125
125
  The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/),
126
126
  and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
127
127
 
128
+ ## [1.45.0] - 2026-01-23
129
+ - Remove unused code execution options from config
130
+
131
+ ## [1.44.0] - 2026-01-23
132
+ - Add LoopIterationRunner abstraction for the responses api
133
+
128
134
  ## [1.43.11] - 2026-01-23
129
135
  - Add `RESPONSES_API` Capablity to some models that were missing it
130
136
 
@@ -90,16 +90,18 @@ unique_toolkit/agentic/history_manager/history_construction_with_contents.py,sha
90
90
  unique_toolkit/agentic/history_manager/history_manager.py,sha256=7V7_173XkAjc8otBACF0G3dbqRs34FSlURbBPrE95Wk,9537
91
91
  unique_toolkit/agentic/history_manager/loop_token_reducer.py,sha256=3c-uonDovtanEJUpAO4zlA4-n9MS_Ws_V0Yb6G7hPM0,20172
92
92
  unique_toolkit/agentic/history_manager/utils.py,sha256=VIn_UmcR3jHtpux0qp5lQQzczgAm8XYSeQiPo87jC3A,3143
93
- unique_toolkit/agentic/loop_runner/__init__.py,sha256=26SAJn737LLmc3RoT6PPoPIeI80nlNWobHK6-Vz8HeI,1099
93
+ unique_toolkit/agentic/loop_runner/__init__.py,sha256=FUjG2vvcAPHfKyRn-EbNYw7n35z33maZWGP8C5_wevg,1258
94
94
  unique_toolkit/agentic/loop_runner/_iteration_handler_utils.py,sha256=czwTbMMmikppRjAA2wpx-UZGub8KrpSoWfojklr-5sE,3261
95
+ unique_toolkit/agentic/loop_runner/_responses_iteration_handler_utils.py,sha256=F_j81IPkGWBWMsti3fppSWlG1OAZWP5DlPsXsCefVzA,1936
96
+ unique_toolkit/agentic/loop_runner/_responses_stream_handler_utils.py,sha256=f9XM5EkAGl0XgKoAd6QnVo0OihL3IG5TawkBr4-IpeI,2604
95
97
  unique_toolkit/agentic/loop_runner/_stream_handler_utils.py,sha256=FTGc5y8wkDnwnRVSYEdandgKz-FiySOsrTFFMadwP6E,1706
96
- unique_toolkit/agentic/loop_runner/base.py,sha256=3g4PalzV00o8kcRwHds2c2rtxW4idD7_7vS2Z7GkMvQ,1370
98
+ unique_toolkit/agentic/loop_runner/base.py,sha256=bTTJwPpBHBHNCOu1Me62Iwtw2Kk4HBJzIZ5QnVVh7Cs,2883
97
99
  unique_toolkit/agentic/loop_runner/middleware/__init__.py,sha256=5yhFZ14_C1qAt-Mb3u3nZ1h6gxuSZ5Ts90-rbk2jjUM,232
98
100
  unique_toolkit/agentic/loop_runner/middleware/planning/__init__.py,sha256=Y9MlihNA8suNREixW98RF45bj0EMtD_tQuDrO2MEML4,304
99
101
  unique_toolkit/agentic/loop_runner/middleware/planning/planning.py,sha256=s6SAP3BCCExgwnyRj_bZTaWgTOiNVju5qcJA0WFUUoE,3216
100
102
  unique_toolkit/agentic/loop_runner/middleware/planning/schema.py,sha256=76C36CWCLfDAYYqtaQlhXsmkWM1fCqf8j-l5afQREKA,2869
101
- unique_toolkit/agentic/loop_runner/runners/__init__.py,sha256=JLBzWjOvTScIVHafeoqG1ZS3syLrYPBDzIfBuJFOw8s,598
102
- unique_toolkit/agentic/loop_runner/runners/basic.py,sha256=PKg7AXp5_2OJpJX8b7_5u0jBKffX5mg8X6DFk98sSwA,1535
103
+ unique_toolkit/agentic/loop_runner/runners/__init__.py,sha256=gQQIBBownqG9C6XNMUhowfKS4FNpEB3WhLThb4YcJSY,678
104
+ unique_toolkit/agentic/loop_runner/runners/basic.py,sha256=O-rIbO-sflvd3GwUq8sIKFEWivtw-wmV-iC82ZeMSiI,2573
103
105
  unique_toolkit/agentic/loop_runner/runners/qwen/__init__.py,sha256=RRT2Q_eeEjtAMsPJC7sYHSlY8clGYXRtQPQhA3HkHF0,481
104
106
  unique_toolkit/agentic/loop_runner/runners/qwen/helpers.py,sha256=JBnxeYKu8HiUq3VHjnb8XdHCe1c3cJ3OwagckF4UvnU,1763
105
107
  unique_toolkit/agentic/loop_runner/runners/qwen/qwen_runner.py,sha256=Qz_kwUUjLDUPBGM0hRVZSwVLYJ1fX5V646oU8AObIUw,4954
@@ -150,8 +152,8 @@ unique_toolkit/agentic/tools/mcp/tool_wrapper.py,sha256=JuNP_mild2vx6opIt5432y6m
150
152
  unique_toolkit/agentic/tools/openai_builtin/__init__.py,sha256=NdVjkTa3LbW-JHhzPRjinTmgOCtEv090Zr9jGZXmgqs,345
151
153
  unique_toolkit/agentic/tools/openai_builtin/base.py,sha256=qtUtv-iyid4GpTRKU7IwWuY_OCPHLKEi9x7MCftyLG0,1178
152
154
  unique_toolkit/agentic/tools/openai_builtin/code_interpreter/__init__.py,sha256=w2vONpnC6hKRPoJGwzDuRtNBsQd_o-gMUqArgIl_5KY,305
153
- unique_toolkit/agentic/tools/openai_builtin/code_interpreter/config.py,sha256=GS0WLrAo0lCW7d4ih5wheskNG7AsdXsbwEYEESX4K60,5327
154
- unique_toolkit/agentic/tools/openai_builtin/code_interpreter/service.py,sha256=KvJWpD-Qief_OONMs_WG8Ya2e9AnmT6Lzow-6lmnVcY,7998
155
+ unique_toolkit/agentic/tools/openai_builtin/code_interpreter/config.py,sha256=U5MHglk4wkVd2L4DxSiATZElr2RUUwKAmPrmt6n70PY,4538
156
+ unique_toolkit/agentic/tools/openai_builtin/code_interpreter/service.py,sha256=bzOU8rV-FUkgSpX-kWzd0_WDEP3sOtzPF2Hbodh0oLA,7896
155
157
  unique_toolkit/agentic/tools/openai_builtin/manager.py,sha256=QeDVgLfnCXrSmXI3b9bgQa9oyfQe_L15wa_YfhfNe9E,2633
156
158
  unique_toolkit/agentic/tools/schemas.py,sha256=TXshRvivr2hD-McXHumO0bp-Z0mz_GnAmQRiVjT59rU,5025
157
159
  unique_toolkit/agentic/tools/test/test_mcp_manager.py,sha256=VpB4k4Dh0lQWakilJMQSzO8sBXapuEC26cub_lorl-M,19221
@@ -242,7 +244,7 @@ unique_toolkit/short_term_memory/service.py,sha256=5PeVBu1ZCAfyDb2HLVvlmqSbyzBBu
242
244
  unique_toolkit/smart_rules/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
243
245
  unique_toolkit/smart_rules/compile.py,sha256=Ozhh70qCn2yOzRWr9d8WmJeTo7AQurwd3tStgBMPFLA,1246
244
246
  unique_toolkit/test_utilities/events.py,sha256=_mwV2bs5iLjxS1ynDCjaIq-gjjKhXYCK-iy3dRfvO3g,6410
245
- unique_toolkit-1.43.11.dist-info/LICENSE,sha256=GlN8wHNdh53xwOPg44URnwag6TEolCjoq3YD_KrWgss,193
246
- unique_toolkit-1.43.11.dist-info/METADATA,sha256=PROyDnl3op-5xx5BR_dqNW-FOt0JRqE5Ddhdx1XTBGA,48588
247
- unique_toolkit-1.43.11.dist-info/WHEEL,sha256=sP946D7jFCHeNz5Iq4fL4Lu-PrWrFsgfLXbbkciIZwg,88
248
- unique_toolkit-1.43.11.dist-info/RECORD,,
247
+ unique_toolkit-1.45.0.dist-info/LICENSE,sha256=GlN8wHNdh53xwOPg44URnwag6TEolCjoq3YD_KrWgss,193
248
+ unique_toolkit-1.45.0.dist-info/METADATA,sha256=wfMYzHsUT5U9jNgqX5uNaIFj_rE_1TfL-iySEmB6l-U,48751
249
+ unique_toolkit-1.45.0.dist-info/WHEEL,sha256=sP946D7jFCHeNz5Iq4fL4Lu-PrWrFsgfLXbbkciIZwg,88
250
+ unique_toolkit-1.45.0.dist-info/RECORD,,