llama-index-llms-openai 0.3.15__tar.gz → 0.3.17__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {llama_index_llms_openai-0.3.15 → llama_index_llms_openai-0.3.17}/PKG-INFO +1 -1
- {llama_index_llms_openai-0.3.15 → llama_index_llms_openai-0.3.17}/llama_index/llms/openai/base.py +16 -0
- {llama_index_llms_openai-0.3.15 → llama_index_llms_openai-0.3.17}/pyproject.toml +1 -1
- {llama_index_llms_openai-0.3.15 → llama_index_llms_openai-0.3.17}/README.md +0 -0
- {llama_index_llms_openai-0.3.15 → llama_index_llms_openai-0.3.17}/llama_index/llms/openai/__init__.py +0 -0
- {llama_index_llms_openai-0.3.15 → llama_index_llms_openai-0.3.17}/llama_index/llms/openai/py.typed +0 -0
- {llama_index_llms_openai-0.3.15 → llama_index_llms_openai-0.3.17}/llama_index/llms/openai/utils.py +0 -0
{llama_index_llms_openai-0.3.15 → llama_index_llms_openai-0.3.17}/llama_index/llms/openai/base.py
RENAMED
|
@@ -7,6 +7,7 @@ from typing import (
|
|
|
7
7
|
Dict,
|
|
8
8
|
Generator,
|
|
9
9
|
List,
|
|
10
|
+
Literal,
|
|
10
11
|
Optional,
|
|
11
12
|
Protocol,
|
|
12
13
|
Sequence,
|
|
@@ -217,6 +218,10 @@ class OpenAI(FunctionCallingLLM):
|
|
|
217
218
|
default=False,
|
|
218
219
|
description="Whether to use strict mode for invoking tools/using schemas.",
|
|
219
220
|
)
|
|
221
|
+
reasoning_effort: Literal["low", "medium", "high"] = Field(
|
|
222
|
+
default="medium",
|
|
223
|
+
description="The effort to use for reasoning models.",
|
|
224
|
+
)
|
|
220
225
|
|
|
221
226
|
_client: Optional[SyncOpenAI] = PrivateAttr()
|
|
222
227
|
_aclient: Optional[AsyncOpenAI] = PrivateAttr()
|
|
@@ -248,6 +253,7 @@ class OpenAI(FunctionCallingLLM):
|
|
|
248
253
|
pydantic_program_mode: PydanticProgramMode = PydanticProgramMode.DEFAULT,
|
|
249
254
|
output_parser: Optional[BaseOutputParser] = None,
|
|
250
255
|
strict: bool = False,
|
|
256
|
+
reasoning_effort: Literal["low", "medium", "high"] = "medium",
|
|
251
257
|
**kwargs: Any,
|
|
252
258
|
) -> None:
|
|
253
259
|
additional_kwargs = additional_kwargs or {}
|
|
@@ -281,6 +287,7 @@ class OpenAI(FunctionCallingLLM):
|
|
|
281
287
|
pydantic_program_mode=pydantic_program_mode,
|
|
282
288
|
output_parser=output_parser,
|
|
283
289
|
strict=strict,
|
|
290
|
+
reasoning_effort=reasoning_effort,
|
|
284
291
|
**kwargs,
|
|
285
292
|
)
|
|
286
293
|
|
|
@@ -417,6 +424,15 @@ class OpenAI(FunctionCallingLLM):
|
|
|
417
424
|
all_kwargs = {**base_kwargs, **self.additional_kwargs}
|
|
418
425
|
if "stream" not in all_kwargs and "stream_options" in all_kwargs:
|
|
419
426
|
del all_kwargs["stream_options"]
|
|
427
|
+
if self.model in O1_MODELS and base_kwargs.get("max_tokens") is not None:
|
|
428
|
+
# O1 models use max_completion_tokens instead of max_tokens
|
|
429
|
+
all_kwargs["max_completion_tokens"] = all_kwargs.get(
|
|
430
|
+
"max_completion_tokens", all_kwargs["max_tokens"]
|
|
431
|
+
)
|
|
432
|
+
all_kwargs.pop("max_tokens", None)
|
|
433
|
+
if self.model in O1_MODELS:
|
|
434
|
+
# O1 models support reasoning_effort of low, medium, high
|
|
435
|
+
all_kwargs["reasoning_effort"] = self.reasoning_effort
|
|
420
436
|
|
|
421
437
|
return all_kwargs
|
|
422
438
|
|
|
File without changes
|
|
File without changes
|
{llama_index_llms_openai-0.3.15 → llama_index_llms_openai-0.3.17}/llama_index/llms/openai/py.typed
RENAMED
|
File without changes
|
{llama_index_llms_openai-0.3.15 → llama_index_llms_openai-0.3.17}/llama_index/llms/openai/utils.py
RENAMED
|
File without changes
|