openai-agents 0.2.5__py3-none-any.whl → 0.2.6__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- agents/extensions/models/litellm_model.py +12 -6
- agents/lifecycle.py +2 -2
- agents/model_settings.py +8 -2
- agents/models/chatcmpl_converter.py +14 -11
- agents/models/openai_chatcompletions.py +1 -0
- agents/models/openai_responses.py +11 -4
- agents/realtime/agent.py +6 -0
- agents/realtime/session.py +11 -1
- {openai_agents-0.2.5.dist-info → openai_agents-0.2.6.dist-info}/METADATA +2 -2
- {openai_agents-0.2.5.dist-info → openai_agents-0.2.6.dist-info}/RECORD +12 -12
- {openai_agents-0.2.5.dist-info → openai_agents-0.2.6.dist-info}/WHEEL +0 -0
- {openai_agents-0.2.5.dist-info → openai_agents-0.2.6.dist-info}/licenses/LICENSE +0 -0
|
@@ -18,13 +18,17 @@ except ImportError as _e:
|
|
|
18
18
|
) from _e
|
|
19
19
|
|
|
20
20
|
from openai import NOT_GIVEN, AsyncStream, NotGiven
|
|
21
|
-
from openai.types.chat import
|
|
21
|
+
from openai.types.chat import (
|
|
22
|
+
ChatCompletionChunk,
|
|
23
|
+
ChatCompletionMessageFunctionToolCall,
|
|
24
|
+
)
|
|
22
25
|
from openai.types.chat.chat_completion_message import (
|
|
23
26
|
Annotation,
|
|
24
27
|
AnnotationURLCitation,
|
|
25
28
|
ChatCompletionMessage,
|
|
26
29
|
)
|
|
27
|
-
from openai.types.chat.
|
|
30
|
+
from openai.types.chat.chat_completion_message_function_tool_call import Function
|
|
31
|
+
from openai.types.chat.chat_completion_message_tool_call import ChatCompletionMessageToolCall
|
|
28
32
|
from openai.types.responses import Response
|
|
29
33
|
|
|
30
34
|
from ... import _debug
|
|
@@ -321,6 +325,7 @@ class LitellmModel(Model):
|
|
|
321
325
|
stream=stream,
|
|
322
326
|
stream_options=stream_options,
|
|
323
327
|
reasoning_effort=reasoning_effort,
|
|
328
|
+
top_logprobs=model_settings.top_logprobs,
|
|
324
329
|
extra_headers={**HEADERS, **(model_settings.extra_headers or {})},
|
|
325
330
|
api_key=self.api_key,
|
|
326
331
|
base_url=self.base_url,
|
|
@@ -361,7 +366,7 @@ class LitellmConverter:
|
|
|
361
366
|
if message.role != "assistant":
|
|
362
367
|
raise ModelBehaviorError(f"Unsupported role: {message.role}")
|
|
363
368
|
|
|
364
|
-
tool_calls = (
|
|
369
|
+
tool_calls: list[ChatCompletionMessageToolCall] | None = (
|
|
365
370
|
[LitellmConverter.convert_tool_call_to_openai(tool) for tool in message.tool_calls]
|
|
366
371
|
if message.tool_calls
|
|
367
372
|
else None
|
|
@@ -412,11 +417,12 @@ class LitellmConverter:
|
|
|
412
417
|
@classmethod
|
|
413
418
|
def convert_tool_call_to_openai(
|
|
414
419
|
cls, tool_call: litellm.types.utils.ChatCompletionMessageToolCall
|
|
415
|
-
) ->
|
|
416
|
-
return
|
|
420
|
+
) -> ChatCompletionMessageFunctionToolCall:
|
|
421
|
+
return ChatCompletionMessageFunctionToolCall(
|
|
417
422
|
id=tool_call.id,
|
|
418
423
|
type="function",
|
|
419
424
|
function=Function(
|
|
420
|
-
name=tool_call.function.name or "",
|
|
425
|
+
name=tool_call.function.name or "",
|
|
426
|
+
arguments=tool_call.function.arguments,
|
|
421
427
|
),
|
|
422
428
|
)
|
agents/lifecycle.py
CHANGED
|
@@ -42,7 +42,7 @@ class RunHooksBase(Generic[TContext, TAgent]):
|
|
|
42
42
|
agent: TAgent,
|
|
43
43
|
tool: Tool,
|
|
44
44
|
) -> None:
|
|
45
|
-
"""Called
|
|
45
|
+
"""Called concurrently with tool invocation."""
|
|
46
46
|
pass
|
|
47
47
|
|
|
48
48
|
async def on_tool_end(
|
|
@@ -93,7 +93,7 @@ class AgentHooksBase(Generic[TContext, TAgent]):
|
|
|
93
93
|
agent: TAgent,
|
|
94
94
|
tool: Tool,
|
|
95
95
|
) -> None:
|
|
96
|
-
"""Called
|
|
96
|
+
"""Called concurrently with tool invocation."""
|
|
97
97
|
pass
|
|
98
98
|
|
|
99
99
|
async def on_tool_end(
|
agents/model_settings.py
CHANGED
|
@@ -55,6 +55,7 @@ Headers: TypeAlias = Mapping[str, Union[str, Omit]]
|
|
|
55
55
|
ToolChoice: TypeAlias = Union[Literal["auto", "required", "none"], str, MCPToolChoice, None]
|
|
56
56
|
|
|
57
57
|
|
|
58
|
+
|
|
58
59
|
@dataclass
|
|
59
60
|
class ModelSettings:
|
|
60
61
|
"""Settings to use when calling an LLM.
|
|
@@ -106,16 +107,21 @@ class ModelSettings:
|
|
|
106
107
|
|
|
107
108
|
store: bool | None = None
|
|
108
109
|
"""Whether to store the generated model response for later retrieval.
|
|
109
|
-
|
|
110
|
+
For Responses API: automatically enabled when not specified.
|
|
111
|
+
For Chat Completions API: disabled when not specified."""
|
|
110
112
|
|
|
111
113
|
include_usage: bool | None = None
|
|
112
114
|
"""Whether to include usage chunk.
|
|
113
|
-
|
|
115
|
+
Only available for Chat Completions API."""
|
|
114
116
|
|
|
115
117
|
response_include: list[ResponseIncludable] | None = None
|
|
116
118
|
"""Additional output data to include in the model response.
|
|
117
119
|
[include parameter](https://platform.openai.com/docs/api-reference/responses/create#responses-create-include)"""
|
|
118
120
|
|
|
121
|
+
top_logprobs: int | None = None
|
|
122
|
+
"""Number of top tokens to return logprobs for. Setting this will
|
|
123
|
+
automatically include ``"message.output_text.logprobs"`` in the response."""
|
|
124
|
+
|
|
119
125
|
extra_query: Query | None = None
|
|
120
126
|
"""Additional query fields to provide with the request.
|
|
121
127
|
Defaults to None if not provided."""
|
|
@@ -12,8 +12,8 @@ from openai.types.chat import (
|
|
|
12
12
|
ChatCompletionContentPartTextParam,
|
|
13
13
|
ChatCompletionDeveloperMessageParam,
|
|
14
14
|
ChatCompletionMessage,
|
|
15
|
+
ChatCompletionMessageFunctionToolCallParam,
|
|
15
16
|
ChatCompletionMessageParam,
|
|
16
|
-
ChatCompletionMessageToolCallParam,
|
|
17
17
|
ChatCompletionSystemMessageParam,
|
|
18
18
|
ChatCompletionToolChoiceOptionParam,
|
|
19
19
|
ChatCompletionToolMessageParam,
|
|
@@ -126,15 +126,18 @@ class Converter:
|
|
|
126
126
|
|
|
127
127
|
if message.tool_calls:
|
|
128
128
|
for tool_call in message.tool_calls:
|
|
129
|
-
|
|
130
|
-
|
|
131
|
-
|
|
132
|
-
|
|
133
|
-
|
|
134
|
-
|
|
135
|
-
|
|
129
|
+
if tool_call.type == "function":
|
|
130
|
+
items.append(
|
|
131
|
+
ResponseFunctionToolCall(
|
|
132
|
+
id=FAKE_RESPONSES_ID,
|
|
133
|
+
call_id=tool_call.id,
|
|
134
|
+
arguments=tool_call.function.arguments,
|
|
135
|
+
name=tool_call.function.name,
|
|
136
|
+
type="function_call",
|
|
137
|
+
)
|
|
136
138
|
)
|
|
137
|
-
|
|
139
|
+
elif tool_call.type == "custom":
|
|
140
|
+
pass
|
|
138
141
|
|
|
139
142
|
return items
|
|
140
143
|
|
|
@@ -420,7 +423,7 @@ class Converter:
|
|
|
420
423
|
elif file_search := cls.maybe_file_search_call(item):
|
|
421
424
|
asst = ensure_assistant_message()
|
|
422
425
|
tool_calls = list(asst.get("tool_calls", []))
|
|
423
|
-
new_tool_call =
|
|
426
|
+
new_tool_call = ChatCompletionMessageFunctionToolCallParam(
|
|
424
427
|
id=file_search["id"],
|
|
425
428
|
type="function",
|
|
426
429
|
function={
|
|
@@ -440,7 +443,7 @@ class Converter:
|
|
|
440
443
|
asst = ensure_assistant_message()
|
|
441
444
|
tool_calls = list(asst.get("tool_calls", []))
|
|
442
445
|
arguments = func_call["arguments"] if func_call["arguments"] else "{}"
|
|
443
|
-
new_tool_call =
|
|
446
|
+
new_tool_call = ChatCompletionMessageFunctionToolCallParam(
|
|
444
447
|
id=func_call["call_id"],
|
|
445
448
|
type="function",
|
|
446
449
|
function={
|
|
@@ -287,6 +287,7 @@ class OpenAIChatCompletionsModel(Model):
|
|
|
287
287
|
stream_options=self._non_null_or_not_given(stream_options),
|
|
288
288
|
store=self._non_null_or_not_given(store),
|
|
289
289
|
reasoning_effort=self._non_null_or_not_given(reasoning_effort),
|
|
290
|
+
top_logprobs=self._non_null_or_not_given(model_settings.top_logprobs),
|
|
290
291
|
extra_headers={**HEADERS, **(model_settings.extra_headers or {})},
|
|
291
292
|
extra_query=model_settings.extra_query,
|
|
292
293
|
extra_body=model_settings.extra_body,
|
|
@@ -3,7 +3,7 @@ from __future__ import annotations
|
|
|
3
3
|
import json
|
|
4
4
|
from collections.abc import AsyncIterator
|
|
5
5
|
from dataclasses import dataclass
|
|
6
|
-
from typing import TYPE_CHECKING, Any, Literal, overload
|
|
6
|
+
from typing import TYPE_CHECKING, Any, Literal, cast, overload
|
|
7
7
|
|
|
8
8
|
from openai import NOT_GIVEN, APIStatusError, AsyncOpenAI, AsyncStream, NotGiven
|
|
9
9
|
from openai.types import ChatModel
|
|
@@ -247,9 +247,12 @@ class OpenAIResponsesModel(Model):
|
|
|
247
247
|
converted_tools = Converter.convert_tools(tools, handoffs)
|
|
248
248
|
response_format = Converter.get_response_format(output_schema)
|
|
249
249
|
|
|
250
|
-
|
|
250
|
+
include_set: set[str] = set(converted_tools.includes)
|
|
251
251
|
if model_settings.response_include is not None:
|
|
252
|
-
|
|
252
|
+
include_set.update(model_settings.response_include)
|
|
253
|
+
if model_settings.top_logprobs is not None:
|
|
254
|
+
include_set.add("message.output_text.logprobs")
|
|
255
|
+
include = cast(list[ResponseIncludable], list(include_set))
|
|
253
256
|
|
|
254
257
|
if _debug.DONT_LOG_MODEL_DATA:
|
|
255
258
|
logger.debug("Calling LLM")
|
|
@@ -264,6 +267,10 @@ class OpenAIResponsesModel(Model):
|
|
|
264
267
|
f"Previous response id: {previous_response_id}\n"
|
|
265
268
|
)
|
|
266
269
|
|
|
270
|
+
extra_args = dict(model_settings.extra_args or {})
|
|
271
|
+
if model_settings.top_logprobs is not None:
|
|
272
|
+
extra_args["top_logprobs"] = model_settings.top_logprobs
|
|
273
|
+
|
|
267
274
|
return await self._client.responses.create(
|
|
268
275
|
previous_response_id=self._non_null_or_not_given(previous_response_id),
|
|
269
276
|
instructions=self._non_null_or_not_given(system_instructions),
|
|
@@ -286,7 +293,7 @@ class OpenAIResponsesModel(Model):
|
|
|
286
293
|
store=self._non_null_or_not_given(model_settings.store),
|
|
287
294
|
reasoning=self._non_null_or_not_given(model_settings.reasoning),
|
|
288
295
|
metadata=self._non_null_or_not_given(model_settings.metadata),
|
|
289
|
-
**
|
|
296
|
+
**extra_args,
|
|
290
297
|
)
|
|
291
298
|
|
|
292
299
|
def _get_client(self) -> AsyncOpenAI:
|
agents/realtime/agent.py
CHANGED
|
@@ -7,6 +7,7 @@ from dataclasses import dataclass, field
|
|
|
7
7
|
from typing import Any, Callable, Generic, cast
|
|
8
8
|
|
|
9
9
|
from ..agent import AgentBase
|
|
10
|
+
from ..guardrail import OutputGuardrail
|
|
10
11
|
from ..handoffs import Handoff
|
|
11
12
|
from ..lifecycle import AgentHooksBase, RunHooksBase
|
|
12
13
|
from ..logger import logger
|
|
@@ -62,6 +63,11 @@ class RealtimeAgent(AgentBase, Generic[TContext]):
|
|
|
62
63
|
modularity.
|
|
63
64
|
"""
|
|
64
65
|
|
|
66
|
+
output_guardrails: list[OutputGuardrail[TContext]] = field(default_factory=list)
|
|
67
|
+
"""A list of checks that run on the final output of the agent, after generating a response.
|
|
68
|
+
Runs only if the agent produces a final output.
|
|
69
|
+
"""
|
|
70
|
+
|
|
65
71
|
hooks: RealtimeAgentHooks | None = None
|
|
66
72
|
"""A class that receives callbacks on various lifecycle events for this agent.
|
|
67
73
|
"""
|
agents/realtime/session.py
CHANGED
|
@@ -444,7 +444,17 @@ class RealtimeSession(RealtimeModelListener):
|
|
|
444
444
|
|
|
445
445
|
async def _run_output_guardrails(self, text: str) -> bool:
|
|
446
446
|
"""Run output guardrails on the given text. Returns True if any guardrail was triggered."""
|
|
447
|
-
|
|
447
|
+
combined_guardrails = self._current_agent.output_guardrails + self._run_config.get(
|
|
448
|
+
"output_guardrails", []
|
|
449
|
+
)
|
|
450
|
+
seen_ids: set[int] = set()
|
|
451
|
+
output_guardrails = []
|
|
452
|
+
for guardrail in combined_guardrails:
|
|
453
|
+
guardrail_id = id(guardrail)
|
|
454
|
+
if guardrail_id not in seen_ids:
|
|
455
|
+
output_guardrails.append(guardrail)
|
|
456
|
+
seen_ids.add(guardrail_id)
|
|
457
|
+
|
|
448
458
|
if not output_guardrails or self._interrupted_by_guardrail:
|
|
449
459
|
return False
|
|
450
460
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: openai-agents
|
|
3
|
-
Version: 0.2.
|
|
3
|
+
Version: 0.2.6
|
|
4
4
|
Summary: OpenAI Agents SDK
|
|
5
5
|
Project-URL: Homepage, https://openai.github.io/openai-agents-python/
|
|
6
6
|
Project-URL: Repository, https://github.com/openai/openai-agents-python
|
|
@@ -21,7 +21,7 @@ Classifier: Typing :: Typed
|
|
|
21
21
|
Requires-Python: >=3.9
|
|
22
22
|
Requires-Dist: griffe<2,>=1.5.6
|
|
23
23
|
Requires-Dist: mcp<2,>=1.11.0; python_version >= '3.10'
|
|
24
|
-
Requires-Dist: openai<2,>=1.
|
|
24
|
+
Requires-Dist: openai<2,>=1.99.6
|
|
25
25
|
Requires-Dist: pydantic<3,>=2.10
|
|
26
26
|
Requires-Dist: requests<3,>=2.0
|
|
27
27
|
Requires-Dist: types-requests<3,>=2.0
|
|
@@ -10,9 +10,9 @@ agents/function_schema.py,sha256=yZ3PEOmfy836Me_W4QlItMeFq2j4BtpuI2FmQswbIcQ,135
|
|
|
10
10
|
agents/guardrail.py,sha256=7P-kd9rKPhgB8rtI31MCV5ho4ZrEaNCQxHvE8IK3EOk,9582
|
|
11
11
|
agents/handoffs.py,sha256=31-rQ-iMWlWNd93ivgTTSMGkqlariXrNfWI_udMWt7s,11409
|
|
12
12
|
agents/items.py,sha256=ntrJ-HuqSMC8HtIwS9pcqHYXtiQ2TJB6lHR-bcvNn4c,9848
|
|
13
|
-
agents/lifecycle.py,sha256=
|
|
13
|
+
agents/lifecycle.py,sha256=sJwESHBHbml7rSYH360-P6x1bLyENcQWm4bT4rQcbuo,3129
|
|
14
14
|
agents/logger.py,sha256=p_ef7vWKpBev5FFybPJjhrCCQizK08Yy1A2EDO1SNNg,60
|
|
15
|
-
agents/model_settings.py,sha256=
|
|
15
|
+
agents/model_settings.py,sha256=SKCrfV5A7u0zaY8fh2PZRe08W5sEhArHC3YHpEfeip0,6357
|
|
16
16
|
agents/prompts.py,sha256=Ss5y_7s2HFcRAOAKu4WTxQszs5ybI8TfbxgEYdnj9sg,2231
|
|
17
17
|
agents/py.typed,sha256=AbpHGcgLb-kRsJGnwFEktk7uzpZOCcBY74-YBdrKVGs,1
|
|
18
18
|
agents/repl.py,sha256=FKZlkGfw6QxItTkjFkCAQwXuV_pn69DIamGd3PiKQFk,2361
|
|
@@ -30,7 +30,7 @@ agents/extensions/handoff_filters.py,sha256=Bzkjb1SmIHoibgO26oesNO2Qdx2avfDGkHrS
|
|
|
30
30
|
agents/extensions/handoff_prompt.py,sha256=oGWN0uNh3Z1L7E-Ev2up8W084fFrDNOsLDy7P6bcmic,1006
|
|
31
31
|
agents/extensions/visualization.py,sha256=sf9D_C-HMwkbWdZccTZvvMPRy_NSiwbm48tRJlESQBI,5144
|
|
32
32
|
agents/extensions/models/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
33
|
-
agents/extensions/models/litellm_model.py,sha256=
|
|
33
|
+
agents/extensions/models/litellm_model.py,sha256=4m6MVYaa-pJzXuBNRZGv0vw2R73R32B0EAZ1kXanVVw,15692
|
|
34
34
|
agents/extensions/models/litellm_provider.py,sha256=wTm00Anq8YoNb9AnyT0JOunDG-HCDm_98ORNy7aNJdw,928
|
|
35
35
|
agents/mcp/__init__.py,sha256=yHmmYlrmEHzUas1inRLKL2iPqbb_-107G3gKe_tyg4I,750
|
|
36
36
|
agents/mcp/server.py,sha256=mTXQL4om5oA2fYevk63SUlwDri-RcUleUH_4hFrA0QM,24266
|
|
@@ -39,20 +39,20 @@ agents/memory/__init__.py,sha256=bo2Rb3PqwSCo9PhBVVJOjvjMM1TfytuDPAFEDADYwwA,84
|
|
|
39
39
|
agents/memory/session.py,sha256=9RQ1I7qGh_9DzsyUd9srSPrxRBlw7jks-67NxYqKvvs,13060
|
|
40
40
|
agents/models/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
41
41
|
agents/models/_openai_shared.py,sha256=4Ngwo2Fv2RXY61Pqck1cYPkSln2tDnb8Ai-ao4QG-iE,836
|
|
42
|
-
agents/models/chatcmpl_converter.py,sha256=
|
|
42
|
+
agents/models/chatcmpl_converter.py,sha256=fdA-4_O7GabTCFZJOrtI6TdxFvjS4Bn4vf2RwVC9yNA,20012
|
|
43
43
|
agents/models/chatcmpl_helpers.py,sha256=eIWySobaH7I0AQijAz5i-_rtsXrSvmEHD567s_8Zw1o,1318
|
|
44
44
|
agents/models/chatcmpl_stream_handler.py,sha256=XUoMnNEcSqK6IRMI6GPH8CwMCXi6NhbfHfpCY3SXJOM,24124
|
|
45
45
|
agents/models/fake_id.py,sha256=lbXjUUSMeAQ8eFx4V5QLUnBClHE6adJlYYav55RlG5w,268
|
|
46
46
|
agents/models/interface.py,sha256=TpY_GEk3LLMozCcYAEcC-Y_VRpI3pwE7A7ZM317mk7M,3839
|
|
47
47
|
agents/models/multi_provider.py,sha256=aiDbls5G4YomPfN6qH1pGlj41WS5jlDp2T82zm6qcnM,5578
|
|
48
|
-
agents/models/openai_chatcompletions.py,sha256=
|
|
48
|
+
agents/models/openai_chatcompletions.py,sha256=N_8U_rKRhB1pgMju29bOok1QFWF_EL7JoatlKzy7hLY,13102
|
|
49
49
|
agents/models/openai_provider.py,sha256=NMxTNaoTa329GrA7jj51LC02pb_e2eFh-PCvWADJrkY,3478
|
|
50
|
-
agents/models/openai_responses.py,sha256=
|
|
50
|
+
agents/models/openai_responses.py,sha256=pBAHIwz_kq561bmzqMwz6L4dFd_R4V5C7R21xLBM__o,17048
|
|
51
51
|
agents/realtime/README.md,sha256=5YCYXH5ULmlWoWo1PE9TlbHjeYgjnp-xY8ZssSFY2Vk,126
|
|
52
52
|
agents/realtime/__init__.py,sha256=7qvzK8QJuHRnPHxDgDj21v8-lnSN4Uurg9znwJv_Tqg,4923
|
|
53
53
|
agents/realtime/_default_tracker.py,sha256=4OMxBvD1MnZmMn6JZYKL42uWhVzvK6NdDLDfPP54d78,1765
|
|
54
54
|
agents/realtime/_util.py,sha256=uawurhWKi3_twNFcZ5Yn1mVvv0RKl4IoyCSag8hGxrE,313
|
|
55
|
-
agents/realtime/agent.py,sha256=
|
|
55
|
+
agents/realtime/agent.py,sha256=yZDgycnLFtJcfl7UHak5GEyL2vdBGxegfqEiuuzGPEk,4027
|
|
56
56
|
agents/realtime/config.py,sha256=FMLT2BdxjOCHmBnvd35sZk68U4jEXypngMRAPkm-irk,5828
|
|
57
57
|
agents/realtime/events.py,sha256=YnyXmkc2rkIAcCDoW5yxylMYeXeaq_QTlyRR5u5VsaM,5534
|
|
58
58
|
agents/realtime/handoffs.py,sha256=avLFix5kEutel57IRcddssGiVHzGptOzWL9OqPaLVh8,6702
|
|
@@ -62,7 +62,7 @@ agents/realtime/model_events.py,sha256=X7UrUU_g4u5gWaf2mUesJJ-Ik1Z1QE0Z-ZP7kDmX1
|
|
|
62
62
|
agents/realtime/model_inputs.py,sha256=OW2bn3wD5_pXLunDUf35jhG2q_bTKbC_D7Qu-83aOEA,2243
|
|
63
63
|
agents/realtime/openai_realtime.py,sha256=vgzgklFcRpB9ZfsDda7DtXlBn3NF6bZdysta1DwQhrM,30120
|
|
64
64
|
agents/realtime/runner.py,sha256=KfU7utmc9QFH2htIKN2IN9H-5EnB0qN9ezmvlRTnOm4,2511
|
|
65
|
-
agents/realtime/session.py,sha256=
|
|
65
|
+
agents/realtime/session.py,sha256=yMHFhqhBKDHzlK-k6JTuqXKggMPW3dPt-aavqDoKsec,23375
|
|
66
66
|
agents/tracing/__init__.py,sha256=5HO_6na5S6EwICgwl50OMtxiIIosUrqalhvldlYvSVc,2991
|
|
67
67
|
agents/tracing/create.py,sha256=xpJ4ZRnGyUDPKoVVkA_8hmdhtwOKGhSkwRco2AQIhAo,18003
|
|
68
68
|
agents/tracing/logger.py,sha256=J4KUDRSGa7x5UVfUwWe-gbKwoaq8AeETRqkPt3QvtGg,68
|
|
@@ -97,7 +97,7 @@ agents/voice/models/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSu
|
|
|
97
97
|
agents/voice/models/openai_model_provider.py,sha256=Khn0uT-VhsEbe7_OhBMGFQzXNwL80gcWZyTHl3CaBII,3587
|
|
98
98
|
agents/voice/models/openai_stt.py,sha256=LcVDS7f1pmbm--PWX-IaV9uLg9uv5_L3vSCbVnTJeGs,16864
|
|
99
99
|
agents/voice/models/openai_tts.py,sha256=4KoLQuFDHKu5a1VTJlu9Nj3MHwMlrn9wfT_liJDJ2dw,1477
|
|
100
|
-
openai_agents-0.2.
|
|
101
|
-
openai_agents-0.2.
|
|
102
|
-
openai_agents-0.2.
|
|
103
|
-
openai_agents-0.2.
|
|
100
|
+
openai_agents-0.2.6.dist-info/METADATA,sha256=E_Fnl2A-qaNEFT07zAH1lx7zIj-XVh7Wli5P5NhfjR0,12104
|
|
101
|
+
openai_agents-0.2.6.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
|
|
102
|
+
openai_agents-0.2.6.dist-info/licenses/LICENSE,sha256=E994EspT7Krhy0qGiES7WYNzBHrh1YDk3r--8d1baRU,1063
|
|
103
|
+
openai_agents-0.2.6.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|