llama-index-llms-openai 0.3.30__tar.gz → 0.3.31__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {llama_index_llms_openai-0.3.30 → llama_index_llms_openai-0.3.31}/PKG-INFO +1 -1
- {llama_index_llms_openai-0.3.30 → llama_index_llms_openai-0.3.31}/llama_index/llms/openai/responses.py +8 -1
- {llama_index_llms_openai-0.3.30 → llama_index_llms_openai-0.3.31}/llama_index/llms/openai/utils.py +5 -1
- {llama_index_llms_openai-0.3.30 → llama_index_llms_openai-0.3.31}/pyproject.toml +2 -1
- {llama_index_llms_openai-0.3.30 → llama_index_llms_openai-0.3.31}/LICENSE +0 -0
- {llama_index_llms_openai-0.3.30 → llama_index_llms_openai-0.3.31}/README.md +0 -0
- {llama_index_llms_openai-0.3.30 → llama_index_llms_openai-0.3.31}/llama_index/llms/openai/__init__.py +0 -0
- {llama_index_llms_openai-0.3.30 → llama_index_llms_openai-0.3.31}/llama_index/llms/openai/base.py +0 -0
- {llama_index_llms_openai-0.3.30 → llama_index_llms_openai-0.3.31}/llama_index/llms/openai/py.typed +0 -0
|
@@ -247,6 +247,10 @@ class OpenAIResponses(FunctionCallingLLM):
|
|
|
247
247
|
default=None,
|
|
248
248
|
description="The effort to use for reasoning models.",
|
|
249
249
|
)
|
|
250
|
+
context_window: Optional[int] = Field(
|
|
251
|
+
default=None,
|
|
252
|
+
description="The context window override for the model.",
|
|
253
|
+
)
|
|
250
254
|
|
|
251
255
|
_client: SyncOpenAI = PrivateAttr()
|
|
252
256
|
_aclient: AsyncOpenAI = PrivateAttr()
|
|
@@ -281,6 +285,7 @@ class OpenAIResponses(FunctionCallingLLM):
|
|
|
281
285
|
async_http_client: Optional[httpx.AsyncClient] = None,
|
|
282
286
|
openai_client: Optional[SyncOpenAI] = None,
|
|
283
287
|
async_openai_client: Optional[AsyncOpenAI] = None,
|
|
288
|
+
context_window: Optional[int] = None,
|
|
284
289
|
**kwargs: Any,
|
|
285
290
|
) -> None:
|
|
286
291
|
additional_kwargs = additional_kwargs or {}
|
|
@@ -316,6 +321,7 @@ class OpenAIResponses(FunctionCallingLLM):
|
|
|
316
321
|
default_headers=default_headers,
|
|
317
322
|
call_metadata=call_metadata,
|
|
318
323
|
strict=strict,
|
|
324
|
+
context_window=context_window,
|
|
319
325
|
**kwargs,
|
|
320
326
|
)
|
|
321
327
|
|
|
@@ -339,7 +345,8 @@ class OpenAIResponses(FunctionCallingLLM):
|
|
|
339
345
|
@property
|
|
340
346
|
def metadata(self) -> LLMMetadata:
|
|
341
347
|
return LLMMetadata(
|
|
342
|
-
context_window=
|
|
348
|
+
context_window=self.context_window
|
|
349
|
+
or openai_modelname_to_contextsize(self._get_model_name()),
|
|
343
350
|
num_output=self.max_output_tokens or -1,
|
|
344
351
|
is_chat_model=True,
|
|
345
352
|
is_function_calling_model=is_function_calling_model(
|
{llama_index_llms_openai-0.3.30 → llama_index_llms_openai-0.3.31}/llama_index/llms/openai/utils.py
RENAMED
|
@@ -257,6 +257,10 @@ def is_chat_model(model: str) -> bool:
|
|
|
257
257
|
|
|
258
258
|
|
|
259
259
|
def is_function_calling_model(model: str) -> bool:
|
|
260
|
+
# default to True for models that are not in the ALL_AVAILABLE_MODELS dict
|
|
261
|
+
if model not in ALL_AVAILABLE_MODELS:
|
|
262
|
+
return True
|
|
263
|
+
|
|
260
264
|
# checking whether the model is fine-tuned or not.
|
|
261
265
|
# fine-tuned model names these days look like:
|
|
262
266
|
# ft:gpt-3.5-turbo:acemeco:suffix:abc123
|
|
@@ -510,7 +514,7 @@ def to_openai_message_dicts(
|
|
|
510
514
|
message_dicts = to_openai_responses_message_dict(
|
|
511
515
|
message,
|
|
512
516
|
drop_none=drop_none,
|
|
513
|
-
model=
|
|
517
|
+
model="o3-mini", # hardcode to ensure developer messages are used
|
|
514
518
|
)
|
|
515
519
|
if isinstance(message_dicts, list):
|
|
516
520
|
final_message_dicts.extend(message_dicts)
|
|
@@ -28,7 +28,7 @@ exclude = ["**/BUILD"]
|
|
|
28
28
|
license = "MIT"
|
|
29
29
|
name = "llama-index-llms-openai"
|
|
30
30
|
readme = "README.md"
|
|
31
|
-
version = "0.3.
|
|
31
|
+
version = "0.3.31"
|
|
32
32
|
|
|
33
33
|
[tool.poetry.dependencies]
|
|
34
34
|
python = ">=3.9,<4.0"
|
|
@@ -42,6 +42,7 @@ mypy = "0.991"
|
|
|
42
42
|
pre-commit = "3.2.0"
|
|
43
43
|
pylint = "2.15.10"
|
|
44
44
|
pytest = "7.2.1"
|
|
45
|
+
pytest-asyncio = "0.21.0"
|
|
45
46
|
pytest-mock = "3.11.1"
|
|
46
47
|
ruff = "0.0.292"
|
|
47
48
|
tree-sitter-languages = "^1.8.0"
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{llama_index_llms_openai-0.3.30 → llama_index_llms_openai-0.3.31}/llama_index/llms/openai/base.py
RENAMED
|
File without changes
|
{llama_index_llms_openai-0.3.30 → llama_index_llms_openai-0.3.31}/llama_index/llms/openai/py.typed
RENAMED
|
File without changes
|