unique_toolkit 1.15.0__py3-none-any.whl → 1.16.1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- unique_toolkit/agentic/debug_info_manager/debug_info_manager.py +16 -6
- unique_toolkit/agentic/debug_info_manager/test/test_debug_info_manager.py +278 -0
- unique_toolkit/agentic/postprocessor/postprocessor_manager.py +50 -11
- unique_toolkit/agentic/responses_api/__init__.py +19 -0
- unique_toolkit/agentic/responses_api/postprocessors/code_display.py +63 -0
- unique_toolkit/agentic/responses_api/postprocessors/generated_files.py +145 -0
- unique_toolkit/agentic/responses_api/stream_handler.py +15 -0
- unique_toolkit/agentic/tools/factory.py +4 -0
- unique_toolkit/agentic/tools/openai_builtin/__init__.py +11 -0
- unique_toolkit/agentic/tools/openai_builtin/base.py +30 -0
- unique_toolkit/agentic/tools/openai_builtin/code_interpreter/__init__.py +8 -0
- unique_toolkit/agentic/tools/openai_builtin/code_interpreter/config.py +57 -0
- unique_toolkit/agentic/tools/openai_builtin/code_interpreter/service.py +230 -0
- unique_toolkit/agentic/tools/openai_builtin/manager.py +62 -0
- unique_toolkit/agentic/tools/tool_manager.py +290 -125
- unique_toolkit/chat/functions.py +15 -6
- unique_toolkit/chat/responses_api.py +461 -0
- unique_toolkit/language_model/functions.py +25 -9
- unique_toolkit/language_model/schemas.py +222 -27
- unique_toolkit/protocols/support.py +91 -9
- unique_toolkit/services/__init__.py +7 -0
- unique_toolkit/services/chat_service.py +139 -7
- {unique_toolkit-1.15.0.dist-info → unique_toolkit-1.16.1.dist-info}/METADATA +8 -1
- {unique_toolkit-1.15.0.dist-info → unique_toolkit-1.16.1.dist-info}/RECORD +26 -13
- {unique_toolkit-1.15.0.dist-info → unique_toolkit-1.16.1.dist-info}/LICENSE +0 -0
- {unique_toolkit-1.15.0.dist-info → unique_toolkit-1.16.1.dist-info}/WHEEL +0 -0
@@ -1,8 +1,18 @@
|
|
1
1
|
import logging
|
2
|
-
from typing import Any
|
2
|
+
from typing import Any, Sequence
|
3
3
|
|
4
4
|
import unique_sdk
|
5
|
+
from openai.types.chat import ChatCompletionToolChoiceOptionParam
|
5
6
|
from openai.types.chat.chat_completion_message_param import ChatCompletionMessageParam
|
7
|
+
from openai.types.responses import (
|
8
|
+
ResponseIncludable,
|
9
|
+
ResponseInputItemParam,
|
10
|
+
ResponseOutputItem,
|
11
|
+
ResponseTextConfigParam,
|
12
|
+
ToolParam,
|
13
|
+
response_create_params,
|
14
|
+
)
|
15
|
+
from openai.types.shared_params import Metadata, Reasoning
|
6
16
|
from typing_extensions import deprecated
|
7
17
|
|
8
18
|
from unique_toolkit._common.utils.files import is_file_content, is_image_content
|
@@ -37,6 +47,10 @@ from unique_toolkit.chat.functions import (
|
|
37
47
|
update_message_log,
|
38
48
|
update_message_log_async,
|
39
49
|
)
|
50
|
+
from unique_toolkit.chat.responses_api import (
|
51
|
+
stream_responses_with_references,
|
52
|
+
stream_responses_with_references_async,
|
53
|
+
)
|
40
54
|
from unique_toolkit.chat.schemas import (
|
41
55
|
ChatMessage,
|
42
56
|
ChatMessageAssessment,
|
@@ -70,11 +84,13 @@ from unique_toolkit.language_model.infos import (
|
|
70
84
|
LanguageModelName,
|
71
85
|
)
|
72
86
|
from unique_toolkit.language_model.schemas import (
|
87
|
+
LanguageModelMessageOptions,
|
73
88
|
LanguageModelMessages,
|
74
89
|
LanguageModelResponse,
|
75
90
|
LanguageModelStreamResponse,
|
76
91
|
LanguageModelTool,
|
77
92
|
LanguageModelToolDescription,
|
93
|
+
ResponsesLanguageModelStreamResponse,
|
78
94
|
)
|
79
95
|
|
80
96
|
logger = logging.getLogger(f"toolkit.{DOMAIN_NAME}.{__name__}")
|
@@ -1292,8 +1308,9 @@ class ChatService(ChatServiceDeprecated):
|
|
1292
1308
|
debug_info: dict = {},
|
1293
1309
|
temperature: float = DEFAULT_COMPLETE_TEMPERATURE,
|
1294
1310
|
timeout: int = DEFAULT_COMPLETE_TIMEOUT,
|
1295
|
-
tools:
|
1311
|
+
tools: Sequence[LanguageModelTool | LanguageModelToolDescription] | None = None,
|
1296
1312
|
start_text: str | None = None,
|
1313
|
+
tool_choice: ChatCompletionToolChoiceOptionParam | None = None,
|
1297
1314
|
other_options: dict | None = None,
|
1298
1315
|
) -> LanguageModelStreamResponse:
|
1299
1316
|
return self.complete_with_references(
|
@@ -1305,6 +1322,7 @@ class ChatService(ChatServiceDeprecated):
|
|
1305
1322
|
timeout=timeout,
|
1306
1323
|
tools=tools,
|
1307
1324
|
start_text=start_text,
|
1325
|
+
tool_choice=tool_choice,
|
1308
1326
|
other_options=other_options,
|
1309
1327
|
)
|
1310
1328
|
|
@@ -1316,8 +1334,9 @@ class ChatService(ChatServiceDeprecated):
|
|
1316
1334
|
debug_info: dict | None = None,
|
1317
1335
|
temperature: float = DEFAULT_COMPLETE_TEMPERATURE,
|
1318
1336
|
timeout: int = DEFAULT_COMPLETE_TIMEOUT,
|
1319
|
-
tools:
|
1337
|
+
tools: Sequence[LanguageModelTool | LanguageModelToolDescription] | None = None,
|
1320
1338
|
start_text: str | None = None,
|
1339
|
+
tool_choice: ChatCompletionToolChoiceOptionParam | None = None,
|
1321
1340
|
other_options: dict | None = None,
|
1322
1341
|
) -> LanguageModelStreamResponse:
|
1323
1342
|
"""Streams a completion in the chat session synchronously."""
|
@@ -1336,6 +1355,7 @@ class ChatService(ChatServiceDeprecated):
|
|
1336
1355
|
timeout=timeout,
|
1337
1356
|
tools=tools,
|
1338
1357
|
start_text=start_text,
|
1358
|
+
tool_choice=tool_choice,
|
1339
1359
|
other_options=other_options,
|
1340
1360
|
)
|
1341
1361
|
|
@@ -1347,8 +1367,9 @@ class ChatService(ChatServiceDeprecated):
|
|
1347
1367
|
debug_info: dict | None = None,
|
1348
1368
|
temperature: float = DEFAULT_COMPLETE_TEMPERATURE,
|
1349
1369
|
timeout: int = DEFAULT_COMPLETE_TIMEOUT,
|
1350
|
-
tools:
|
1370
|
+
tools: Sequence[LanguageModelTool | LanguageModelToolDescription] | None = None,
|
1351
1371
|
start_text: str | None = None,
|
1372
|
+
tool_choice: ChatCompletionToolChoiceOptionParam | None = None,
|
1352
1373
|
other_options: dict | None = None,
|
1353
1374
|
) -> LanguageModelResponse:
|
1354
1375
|
response = self.complete_with_references(
|
@@ -1360,6 +1381,7 @@ class ChatService(ChatServiceDeprecated):
|
|
1360
1381
|
timeout=timeout,
|
1361
1382
|
tools=tools,
|
1362
1383
|
start_text=start_text,
|
1384
|
+
tool_choice=tool_choice,
|
1363
1385
|
other_options=other_options,
|
1364
1386
|
)
|
1365
1387
|
|
@@ -1374,8 +1396,9 @@ class ChatService(ChatServiceDeprecated):
|
|
1374
1396
|
debug_info: dict | None = None,
|
1375
1397
|
temperature: float = DEFAULT_COMPLETE_TEMPERATURE,
|
1376
1398
|
timeout: int = DEFAULT_COMPLETE_TIMEOUT,
|
1377
|
-
tools:
|
1399
|
+
tools: Sequence[LanguageModelTool | LanguageModelToolDescription] | None = None,
|
1378
1400
|
start_text: str | None = None,
|
1401
|
+
tool_choice: ChatCompletionToolChoiceOptionParam | None = None,
|
1379
1402
|
other_options: dict | None = None,
|
1380
1403
|
) -> LanguageModelStreamResponse:
|
1381
1404
|
"""Stream a completion in the chat session asynchronously."""
|
@@ -1388,6 +1411,7 @@ class ChatService(ChatServiceDeprecated):
|
|
1388
1411
|
timeout=timeout,
|
1389
1412
|
tools=tools,
|
1390
1413
|
start_text=start_text,
|
1414
|
+
tool_choice=tool_choice,
|
1391
1415
|
other_options=other_options,
|
1392
1416
|
)
|
1393
1417
|
|
@@ -1399,7 +1423,8 @@ class ChatService(ChatServiceDeprecated):
|
|
1399
1423
|
debug_info: dict | None = None,
|
1400
1424
|
temperature: float = DEFAULT_COMPLETE_TEMPERATURE,
|
1401
1425
|
timeout: int = DEFAULT_COMPLETE_TIMEOUT,
|
1402
|
-
tools:
|
1426
|
+
tools: Sequence[LanguageModelTool | LanguageModelToolDescription] | None = None,
|
1427
|
+
tool_choice: ChatCompletionToolChoiceOptionParam | None = None,
|
1403
1428
|
start_text: str | None = None,
|
1404
1429
|
other_options: dict | None = None,
|
1405
1430
|
) -> LanguageModelStreamResponse:
|
@@ -1418,6 +1443,7 @@ class ChatService(ChatServiceDeprecated):
|
|
1418
1443
|
timeout=timeout,
|
1419
1444
|
tools=tools,
|
1420
1445
|
start_text=start_text,
|
1446
|
+
tool_choice=tool_choice,
|
1421
1447
|
other_options=other_options,
|
1422
1448
|
)
|
1423
1449
|
|
@@ -1429,8 +1455,9 @@ class ChatService(ChatServiceDeprecated):
|
|
1429
1455
|
debug_info: dict | None = None,
|
1430
1456
|
temperature: float = DEFAULT_COMPLETE_TEMPERATURE,
|
1431
1457
|
timeout: int = DEFAULT_COMPLETE_TIMEOUT,
|
1432
|
-
tools:
|
1458
|
+
tools: Sequence[LanguageModelTool | LanguageModelToolDescription] | None = None,
|
1433
1459
|
start_text: str | None = None,
|
1460
|
+
tool_choice: ChatCompletionToolChoiceOptionParam | None = None,
|
1434
1461
|
other_options: dict | None = None,
|
1435
1462
|
) -> LanguageModelResponse:
|
1436
1463
|
response = self.complete_with_references_async(
|
@@ -1442,11 +1469,116 @@ class ChatService(ChatServiceDeprecated):
|
|
1442
1469
|
timeout=timeout,
|
1443
1470
|
tools=tools,
|
1444
1471
|
start_text=start_text,
|
1472
|
+
tool_choice=tool_choice,
|
1445
1473
|
other_options=other_options,
|
1446
1474
|
)
|
1447
1475
|
|
1448
1476
|
return LanguageModelResponse.from_stream_response(await response)
|
1449
1477
|
|
1478
|
+
def complete_responses_with_references(
|
1479
|
+
self,
|
1480
|
+
*,
|
1481
|
+
model_name: LanguageModelName | str,
|
1482
|
+
messages: str
|
1483
|
+
| LanguageModelMessages
|
1484
|
+
| Sequence[
|
1485
|
+
ResponseInputItemParam
|
1486
|
+
| LanguageModelMessageOptions
|
1487
|
+
| ResponseOutputItem # History is automatically convertible
|
1488
|
+
],
|
1489
|
+
content_chunks: list[ContentChunk] | None = None,
|
1490
|
+
tools: Sequence[LanguageModelToolDescription | ToolParam] | None = None,
|
1491
|
+
temperature: float = DEFAULT_COMPLETE_TEMPERATURE,
|
1492
|
+
debug_info: dict | None = None,
|
1493
|
+
start_text: str | None = None,
|
1494
|
+
include: list[ResponseIncludable] | None = None,
|
1495
|
+
instructions: str | None = None,
|
1496
|
+
max_output_tokens: int | None = None,
|
1497
|
+
metadata: Metadata | None = None,
|
1498
|
+
parallel_tool_calls: bool | None = None,
|
1499
|
+
text: ResponseTextConfigParam | None = None,
|
1500
|
+
tool_choice: response_create_params.ToolChoice | None = None,
|
1501
|
+
top_p: float | None = None,
|
1502
|
+
reasoning: Reasoning | None = None,
|
1503
|
+
other_options: dict | None = None,
|
1504
|
+
) -> ResponsesLanguageModelStreamResponse:
|
1505
|
+
return stream_responses_with_references(
|
1506
|
+
company_id=self._company_id,
|
1507
|
+
user_id=self._user_id,
|
1508
|
+
assistant_message_id=self._assistant_message_id,
|
1509
|
+
user_message_id=self._user_message_id,
|
1510
|
+
chat_id=self._chat_id,
|
1511
|
+
assistant_id=self._assistant_id,
|
1512
|
+
model_name=model_name,
|
1513
|
+
messages=messages,
|
1514
|
+
content_chunks=content_chunks,
|
1515
|
+
tools=tools,
|
1516
|
+
temperature=temperature,
|
1517
|
+
debug_info=debug_info,
|
1518
|
+
start_text=start_text,
|
1519
|
+
include=include,
|
1520
|
+
instructions=instructions,
|
1521
|
+
max_output_tokens=max_output_tokens,
|
1522
|
+
metadata=metadata,
|
1523
|
+
parallel_tool_calls=parallel_tool_calls,
|
1524
|
+
text=text,
|
1525
|
+
tool_choice=tool_choice,
|
1526
|
+
top_p=top_p,
|
1527
|
+
reasoning=reasoning,
|
1528
|
+
other_options=other_options,
|
1529
|
+
)
|
1530
|
+
|
1531
|
+
async def complete_responses_with_references_async(
|
1532
|
+
self,
|
1533
|
+
*,
|
1534
|
+
model_name: LanguageModelName | str,
|
1535
|
+
messages: str
|
1536
|
+
| LanguageModelMessages
|
1537
|
+
| Sequence[
|
1538
|
+
ResponseInputItemParam | LanguageModelMessageOptions | ResponseOutputItem
|
1539
|
+
],
|
1540
|
+
content_chunks: list[ContentChunk] | None = None,
|
1541
|
+
tools: Sequence[LanguageModelToolDescription | ToolParam] | None = None,
|
1542
|
+
temperature: float = DEFAULT_COMPLETE_TEMPERATURE,
|
1543
|
+
debug_info: dict | None = None,
|
1544
|
+
start_text: str | None = None,
|
1545
|
+
include: list[ResponseIncludable] | None = None,
|
1546
|
+
instructions: str | None = None,
|
1547
|
+
max_output_tokens: int | None = None,
|
1548
|
+
metadata: Metadata | None = None,
|
1549
|
+
parallel_tool_calls: bool | None = None,
|
1550
|
+
text: ResponseTextConfigParam | None = None,
|
1551
|
+
tool_choice: response_create_params.ToolChoice | None = None,
|
1552
|
+
top_p: float | None = None,
|
1553
|
+
reasoning: Reasoning | None = None,
|
1554
|
+
other_options: dict | None = None,
|
1555
|
+
) -> ResponsesLanguageModelStreamResponse:
|
1556
|
+
return await stream_responses_with_references_async(
|
1557
|
+
company_id=self._company_id,
|
1558
|
+
user_id=self._user_id,
|
1559
|
+
assistant_message_id=self._assistant_message_id,
|
1560
|
+
user_message_id=self._user_message_id,
|
1561
|
+
chat_id=self._chat_id,
|
1562
|
+
assistant_id=self._assistant_id,
|
1563
|
+
model_name=model_name,
|
1564
|
+
messages=messages,
|
1565
|
+
content_chunks=content_chunks,
|
1566
|
+
tools=tools,
|
1567
|
+
temperature=temperature,
|
1568
|
+
debug_info=debug_info,
|
1569
|
+
start_text=start_text,
|
1570
|
+
include=include,
|
1571
|
+
instructions=instructions,
|
1572
|
+
max_output_tokens=max_output_tokens,
|
1573
|
+
metadata=metadata,
|
1574
|
+
parallel_tool_calls=parallel_tool_calls,
|
1575
|
+
text=text,
|
1576
|
+
tool_choice=tool_choice,
|
1577
|
+
top_p=top_p,
|
1578
|
+
reasoning=reasoning,
|
1579
|
+
other_options=other_options,
|
1580
|
+
)
|
1581
|
+
|
1450
1582
|
# Chat Content Methods
|
1451
1583
|
############################################################################
|
1452
1584
|
|
@@ -1,6 +1,6 @@
|
|
1
1
|
Metadata-Version: 2.1
|
2
2
|
Name: unique_toolkit
|
3
|
-
Version: 1.
|
3
|
+
Version: 1.16.1
|
4
4
|
Summary:
|
5
5
|
License: Proprietary
|
6
6
|
Author: Cedric Klinkert
|
@@ -118,6 +118,13 @@ All notable changes to this project will be documented in this file.
|
|
118
118
|
The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/),
|
119
119
|
and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html).
|
120
120
|
|
121
|
+
## [1.16.1] - 2025-10-16
|
122
|
+
- Update debug info for better tool call tracking
|
123
|
+
|
124
|
+
## [1.16.0] - 2025-10-16
|
125
|
+
- Add responses api support.
|
126
|
+
- Add utilities for code execution.
|
127
|
+
|
121
128
|
## [1.15.0] - 2025-10-15
|
122
129
|
- Enable to distinguish between environment and modifiable payload parameters in human verification
|
123
130
|
|
@@ -26,7 +26,8 @@ unique_toolkit/_common/utils/write_configuration.py,sha256=fzvr4C-XBL3OSM3Od9Tbq
|
|
26
26
|
unique_toolkit/_common/validate_required_values.py,sha256=Y_M1ub9gIKP9qZ45F6Zq3ZHtuIqhmOjl8Z2Vd3avg8w,588
|
27
27
|
unique_toolkit/_common/validators.py,sha256=LFZmAalNa886EXm1VYamFvfBuUZjYKwDdT_HOYU0BtE,2934
|
28
28
|
unique_toolkit/agentic/__init__.py,sha256=AbpHGcgLb-kRsJGnwFEktk7uzpZOCcBY74-YBdrKVGs,1
|
29
|
-
unique_toolkit/agentic/debug_info_manager/debug_info_manager.py,sha256=
|
29
|
+
unique_toolkit/agentic/debug_info_manager/debug_info_manager.py,sha256=30ZZaw0vffjZjiu9AYdO1Sm8G9FN6XR2ehdOEUCKqh0,891
|
30
|
+
unique_toolkit/agentic/debug_info_manager/test/test_debug_info_manager.py,sha256=_fIS6_DHA8A3AB64-LPgHgUGa1w0CFUWwtgV-ZbhkzA,10535
|
30
31
|
unique_toolkit/agentic/evaluation/config.py,sha256=zcW7m63Yt5G39hN2If8slBl6Eu3jTRoRPjYaUMn54Uk,987
|
31
32
|
unique_toolkit/agentic/evaluation/context_relevancy/prompts.py,sha256=EdHFUOB581yVxcOL8482KUv_LzaRjuiem71EF8udYMc,1331
|
32
33
|
unique_toolkit/agentic/evaluation/context_relevancy/schema.py,sha256=lZd0TPzH43ifgWWGg3WO6b1AQX8aK2R9y51yH0d1DHM,2919
|
@@ -46,8 +47,12 @@ unique_toolkit/agentic/history_manager/history_construction_with_contents.py,sha
|
|
46
47
|
unique_toolkit/agentic/history_manager/history_manager.py,sha256=xWA8w5CWrzueGW05ekIhc_Y_-_380hFfIB4rB5wlito,9470
|
47
48
|
unique_toolkit/agentic/history_manager/loop_token_reducer.py,sha256=4XUX2-yVBnaYthV8p0zj2scVBUdK_3IhxBgoNlrytyQ,18498
|
48
49
|
unique_toolkit/agentic/history_manager/utils.py,sha256=NDSSz0Jp3oVJU3iKlVScmM1AOe-6hTiVjLr16DUPsV0,5656
|
49
|
-
unique_toolkit/agentic/postprocessor/postprocessor_manager.py,sha256=
|
50
|
+
unique_toolkit/agentic/postprocessor/postprocessor_manager.py,sha256=Z6rMQjhD0x6uC4p1cdxbUVv3jO-31hZTyNE1SiGYIu8,5680
|
50
51
|
unique_toolkit/agentic/reference_manager/reference_manager.py,sha256=x51CT0D8HHu2LzgXdHGy0leOYpjnsxVbPZ2nc28G9mA,4005
|
52
|
+
unique_toolkit/agentic/responses_api/__init__.py,sha256=9WTO-ef7fGE9Y1QtZJFm8Q_jkwK8Srtl-HWvpAD2Wxs,668
|
53
|
+
unique_toolkit/agentic/responses_api/postprocessors/code_display.py,sha256=qbrxXL_AQ3ufBOW2TuNgml7d8u6qY_WGriS5jyZlJlE,1902
|
54
|
+
unique_toolkit/agentic/responses_api/postprocessors/generated_files.py,sha256=WYOBZQz33xx2NZTI19_5DY6sH9XQ6N1t7BCZNXQT5NI,4983
|
55
|
+
unique_toolkit/agentic/responses_api/stream_handler.py,sha256=Y1IM0uiPBdlab5UuOTCsHTaVX-fd9MxfS3xkwhdFie4,647
|
51
56
|
unique_toolkit/agentic/short_term_memory_manager/persistent_short_term_memory_manager.py,sha256=uF3HSoZF0hBfuNhIE9N8KRtuwDfpoeXUFVrv_cyZ3Sw,5839
|
52
57
|
unique_toolkit/agentic/thinking_manager/thinking_manager.py,sha256=41QWFsdRrbWlQHBfYCFv726UDom4WbcvaRfjCmoUOQI,4183
|
53
58
|
unique_toolkit/agentic/tools/__init__.py,sha256=-ToY9-Xiz0K7qCUydH1h1yG6n4h1hQS8sBuSVPNEq2Y,43
|
@@ -75,16 +80,22 @@ unique_toolkit/agentic/tools/a2a/tool/config.py,sha256=07W8cZFc32yRvIeUFEQXi8R1c
|
|
75
80
|
unique_toolkit/agentic/tools/a2a/tool/service.py,sha256=6Eru-jqlt13lIBVC26LhH0Oxb3qVdDQIUWRoChf8ysw,10525
|
76
81
|
unique_toolkit/agentic/tools/agent_chunks_hanlder.py,sha256=x32Dp1Z8cVW5i-XzXbaMwX2KHPcNGmqEU-FB4AV9ZGo,1909
|
77
82
|
unique_toolkit/agentic/tools/config.py,sha256=QD83iy2xAJFyoPggYyLWq-MaSGSq00yS9iI31My1NBc,5387
|
78
|
-
unique_toolkit/agentic/tools/factory.py,sha256=
|
83
|
+
unique_toolkit/agentic/tools/factory.py,sha256=A1Aliwx037UAk9ADiDsg0zjCWWnvzV_PxwJNoPTvW6c,1434
|
79
84
|
unique_toolkit/agentic/tools/mcp/__init__.py,sha256=RLF_p-LDRC7GhiB3fdCi4u3bh6V9PY_w26fg61BLyco,122
|
80
85
|
unique_toolkit/agentic/tools/mcp/manager.py,sha256=DPYwwDe6RSZyuPaxn-je49fP_qOOs0ZV46EM6GZcV4c,2748
|
81
86
|
unique_toolkit/agentic/tools/mcp/models.py,sha256=OyCCb7Vwv1ftzC_OCpFkf3TX9Aeb74ZZagG-qK5peIU,722
|
82
87
|
unique_toolkit/agentic/tools/mcp/tool_wrapper.py,sha256=5UdtqFtZ0aqjae2eL3zVacDMfr1hu5KiQkaoI7VkhqA,9972
|
88
|
+
unique_toolkit/agentic/tools/openai_builtin/__init__.py,sha256=NdVjkTa3LbW-JHhzPRjinTmgOCtEv090Zr9jGZXmgqs,345
|
89
|
+
unique_toolkit/agentic/tools/openai_builtin/base.py,sha256=2Lw47XworwwkIQBQW5S1T6HS2mWqx13lx50moAMekRk,808
|
90
|
+
unique_toolkit/agentic/tools/openai_builtin/code_interpreter/__init__.py,sha256=w2vONpnC6hKRPoJGwzDuRtNBsQd_o-gMUqArgIl_5KY,305
|
91
|
+
unique_toolkit/agentic/tools/openai_builtin/code_interpreter/config.py,sha256=r-YIPBJ2V7YybBBb_sOvhxBaOrg1zcIgTe6UwfIwR2I,2575
|
92
|
+
unique_toolkit/agentic/tools/openai_builtin/code_interpreter/service.py,sha256=RKVExS1l3rJDVcu3mxKYA7SNV_Hphx2przhhiC-5PSo,7467
|
93
|
+
unique_toolkit/agentic/tools/openai_builtin/manager.py,sha256=kU4wGit9AnDbkijB7LJEHcGXG8UeTBhiZh4a7lxTGA8,2246
|
83
94
|
unique_toolkit/agentic/tools/schemas.py,sha256=0ZR8xCdGj1sEdPE0lfTIG2uSR5zqWoprUa3Seqez4C8,4837
|
84
95
|
unique_toolkit/agentic/tools/test/test_mcp_manager.py,sha256=PVRvkK3M21rzONpy5VE_i3vEbAGIz1haW_VPVwiPDI0,15724
|
85
96
|
unique_toolkit/agentic/tools/test/test_tool_progress_reporter.py,sha256=dod5QPqgGUInVAGXAbsAKNTEypIi6pUEWhDbJr9YfUU,6307
|
86
97
|
unique_toolkit/agentic/tools/tool.py,sha256=m56VLxiHuKU2_J5foZp00xhm5lTxWEW7zRLGbIE9ssU,6744
|
87
|
-
unique_toolkit/agentic/tools/tool_manager.py,sha256=
|
98
|
+
unique_toolkit/agentic/tools/tool_manager.py,sha256=DtxJobe_7QKFe6CjnMhCP-mnKO6MjnZeDXsO3jBoC9w,16283
|
88
99
|
unique_toolkit/agentic/tools/tool_progress_reporter.py,sha256=ixud9VoHey1vlU1t86cW0-WTvyTwMxNSWBon8I11SUk,7955
|
89
100
|
unique_toolkit/agentic/tools/utils/__init__.py,sha256=iD1YYzf9LcJFv95Z8BqCAFSewNBabybZRZyvPKGfvro,27
|
90
101
|
unique_toolkit/agentic/tools/utils/execution/__init__.py,sha256=OHiKpqBnfhBiEQagKVWJsZlHv8smPp5OI4dFIexzibw,37
|
@@ -105,7 +116,8 @@ unique_toolkit/app/verification.py,sha256=GxFFwcJMy25fCA_Xe89wKW7bgqOu8PAs5y8QpH
|
|
105
116
|
unique_toolkit/chat/__init__.py,sha256=uP7P6YPeOjEOvpX3bhcU6ND_m0QLr4wMklcrnAKK0q4,804
|
106
117
|
unique_toolkit/chat/constants.py,sha256=05kq6zjqUVB2d6_P7s-90nbljpB3ryxwCI-CAz0r2O4,83
|
107
118
|
unique_toolkit/chat/deprecated/service.py,sha256=CYwzXi7OB0RjHd73CO2jq8SlpdBmDYLatzPFkb5sA0k,6529
|
108
|
-
unique_toolkit/chat/functions.py,sha256=
|
119
|
+
unique_toolkit/chat/functions.py,sha256=1KQmqS5g3OgX8MJJmeL-_Cv4M6uJ3e0BCqkLDHl26dA,45580
|
120
|
+
unique_toolkit/chat/responses_api.py,sha256=MCI1MR_4wlo9xY1ifH2daNz4JvjX18uPdryQlemaeLw,14488
|
109
121
|
unique_toolkit/chat/schemas.py,sha256=u3WPdMkOlmwPGHUueQC-nk8k-QkM7ZSUcU0f-32g6Uc,6718
|
110
122
|
unique_toolkit/chat/service.py,sha256=6D00OL4QrGafbOhTaC5zNXaNgg7gS5W_2ePVa4LhqpE,4439
|
111
123
|
unique_toolkit/chat/state.py,sha256=Cjgwv_2vhDFbV69xxsn7SefhaoIAEqLx3ferdVFCnOg,1445
|
@@ -134,15 +146,16 @@ unique_toolkit/language_model/__init__.py,sha256=lRQyLlbwHbNFf4-0foBU13UGb09lwEe
|
|
134
146
|
unique_toolkit/language_model/builder.py,sha256=4OKfwJfj3TrgO1ezc_ewIue6W7BCQ2ZYQXUckWVPPTA,3369
|
135
147
|
unique_toolkit/language_model/constants.py,sha256=B-topqW0r83dkC_25DeQfnPk3n53qzIHUCBS7YJ0-1U,119
|
136
148
|
unique_toolkit/language_model/default_language_model.py,sha256=-_DBsJhLCsFdaU4ynAkyW0jYIl2lhrPybZm1K-GgVJs,125
|
137
|
-
unique_toolkit/language_model/functions.py,sha256=
|
149
|
+
unique_toolkit/language_model/functions.py,sha256=LGX3rR-XjkB-R520jp4w_Azgqf7BsIAo7E_VWoqA5xY,17260
|
138
150
|
unique_toolkit/language_model/infos.py,sha256=pGd4I7fAuy8D8iqEgKC7-grQZe4AIMebonkBBv1k37Q,59363
|
139
151
|
unique_toolkit/language_model/prompt.py,sha256=JSawaLjQg3VR-E2fK8engFyJnNdk21zaO8pPIodzN4Q,3991
|
140
152
|
unique_toolkit/language_model/reference.py,sha256=nkX2VFz-IrUz8yqyc3G5jUMNwrNpxITBrMEKkbqqYoI,8583
|
141
|
-
unique_toolkit/language_model/schemas.py,sha256=
|
153
|
+
unique_toolkit/language_model/schemas.py,sha256=k99I9GAR6e2rdVfbfl49ojbT7sk5xCMNYmME22RGJkk,23423
|
142
154
|
unique_toolkit/language_model/service.py,sha256=JkYGtCug8POQskTv_aoYkzTMOaPCWRM94y73o3bUttQ,12019
|
143
155
|
unique_toolkit/language_model/utils.py,sha256=bPQ4l6_YO71w-zaIPanUUmtbXC1_hCvLK0tAFc3VCRc,1902
|
144
|
-
unique_toolkit/protocols/support.py,sha256=
|
145
|
-
unique_toolkit/services/
|
156
|
+
unique_toolkit/protocols/support.py,sha256=ZEnbQL5w2-T_1AeM8OHycZJ3qbdfVI1nXe0nL9esQEw,5544
|
157
|
+
unique_toolkit/services/__init__.py,sha256=90-IT5FjMcnlqxjp5kme9Fqgp_on46rggctIqHMdqsw,195
|
158
|
+
unique_toolkit/services/chat_service.py,sha256=eGJKrsE4ZB6mXb-d6-9eRoM0_auXPorI-sADl6F593I,59954
|
146
159
|
unique_toolkit/services/knowledge_base.py,sha256=N6_PocSG6EW5hm6cKf7D8auc63dAWhcJat643OFPZJM,27668
|
147
160
|
unique_toolkit/short_term_memory/__init__.py,sha256=2mI3AUrffgH7Yt-xS57EGqnHf7jnn6xquoKEhJqk3Wg,185
|
148
161
|
unique_toolkit/short_term_memory/constants.py,sha256=698CL6-wjup2MvU19RxSmQk3gX7aqW_OOpZB7sbz_Xg,34
|
@@ -152,7 +165,7 @@ unique_toolkit/short_term_memory/service.py,sha256=5PeVBu1ZCAfyDb2HLVvlmqSbyzBBu
|
|
152
165
|
unique_toolkit/smart_rules/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
153
166
|
unique_toolkit/smart_rules/compile.py,sha256=Ozhh70qCn2yOzRWr9d8WmJeTo7AQurwd3tStgBMPFLA,1246
|
154
167
|
unique_toolkit/test_utilities/events.py,sha256=_mwV2bs5iLjxS1ynDCjaIq-gjjKhXYCK-iy3dRfvO3g,6410
|
155
|
-
unique_toolkit-1.
|
156
|
-
unique_toolkit-1.
|
157
|
-
unique_toolkit-1.
|
158
|
-
unique_toolkit-1.
|
168
|
+
unique_toolkit-1.16.1.dist-info/LICENSE,sha256=GlN8wHNdh53xwOPg44URnwag6TEolCjoq3YD_KrWgss,193
|
169
|
+
unique_toolkit-1.16.1.dist-info/METADATA,sha256=bgeCiumOrLKNtqV3xQoDz7Yktvb0z_W8EJnv6pTUVc4,37517
|
170
|
+
unique_toolkit-1.16.1.dist-info/WHEEL,sha256=sP946D7jFCHeNz5Iq4fL4Lu-PrWrFsgfLXbbkciIZwg,88
|
171
|
+
unique_toolkit-1.16.1.dist-info/RECORD,,
|
File without changes
|
File without changes
|