llama-index-llms-openai 0.3.30__py3-none-any.whl → 0.3.31__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -247,6 +247,10 @@ class OpenAIResponses(FunctionCallingLLM):
247
247
  default=None,
248
248
  description="The effort to use for reasoning models.",
249
249
  )
250
+ context_window: Optional[int] = Field(
251
+ default=None,
252
+ description="The context window override for the model.",
253
+ )
250
254
 
251
255
  _client: SyncOpenAI = PrivateAttr()
252
256
  _aclient: AsyncOpenAI = PrivateAttr()
@@ -281,6 +285,7 @@ class OpenAIResponses(FunctionCallingLLM):
281
285
  async_http_client: Optional[httpx.AsyncClient] = None,
282
286
  openai_client: Optional[SyncOpenAI] = None,
283
287
  async_openai_client: Optional[AsyncOpenAI] = None,
288
+ context_window: Optional[int] = None,
284
289
  **kwargs: Any,
285
290
  ) -> None:
286
291
  additional_kwargs = additional_kwargs or {}
@@ -316,6 +321,7 @@ class OpenAIResponses(FunctionCallingLLM):
316
321
  default_headers=default_headers,
317
322
  call_metadata=call_metadata,
318
323
  strict=strict,
324
+ context_window=context_window,
319
325
  **kwargs,
320
326
  )
321
327
 
@@ -339,7 +345,8 @@ class OpenAIResponses(FunctionCallingLLM):
339
345
  @property
340
346
  def metadata(self) -> LLMMetadata:
341
347
  return LLMMetadata(
342
- context_window=openai_modelname_to_contextsize(self._get_model_name()),
348
+ context_window=self.context_window
349
+ or openai_modelname_to_contextsize(self._get_model_name()),
343
350
  num_output=self.max_output_tokens or -1,
344
351
  is_chat_model=True,
345
352
  is_function_calling_model=is_function_calling_model(
@@ -257,6 +257,10 @@ def is_chat_model(model: str) -> bool:
257
257
 
258
258
 
259
259
  def is_function_calling_model(model: str) -> bool:
260
+ # default to True for models that are not in the ALL_AVAILABLE_MODELS dict
261
+ if model not in ALL_AVAILABLE_MODELS:
262
+ return True
263
+
260
264
  # checking whether the model is fine-tuned or not.
261
265
  # fine-tuned model names these days look like:
262
266
  # ft:gpt-3.5-turbo:acemeco:suffix:abc123
@@ -510,7 +514,7 @@ def to_openai_message_dicts(
510
514
  message_dicts = to_openai_responses_message_dict(
511
515
  message,
512
516
  drop_none=drop_none,
513
- model=model,
517
+ model="o3-mini", # hardcode to ensure developer messages are used
514
518
  )
515
519
  if isinstance(message_dicts, list):
516
520
  final_message_dicts.extend(message_dicts)
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: llama-index-llms-openai
3
- Version: 0.3.30
3
+ Version: 0.3.31
4
4
  Summary: llama-index llms openai integration
5
5
  License: MIT
6
6
  Author: llama-index
@@ -0,0 +1,9 @@
1
+ llama_index/llms/openai/__init__.py,sha256=8nmgixeXifQ4eVSgtCic54WxXqrrpXQPL4rhACWCSFs,229
2
+ llama_index/llms/openai/base.py,sha256=z4k8aFFt7av2CevyOFnV8LSauhHjZyaqAxLukbR54sM,38680
3
+ llama_index/llms/openai/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
4
+ llama_index/llms/openai/responses.py,sha256=0HOC7Q6zQFAYRLSWAGHnkHRm-7ZRsqhVKi35BTE18B4,34726
5
+ llama_index/llms/openai/utils.py,sha256=IMmXLQ47YXDoL01Bj1HDZdw1RpZHNV2MmoRRrqUGKPw,26213
6
+ llama_index_llms_openai-0.3.31.dist-info/LICENSE,sha256=JPQLUZD9rKvCTdu192Nk0V5PAwklIg6jANii3UmTyMs,1065
7
+ llama_index_llms_openai-0.3.31.dist-info/METADATA,sha256=hx7htf-tBX-uQ68jZx65WQxM49W61M7BYDvV1v9mKdc,3322
8
+ llama_index_llms_openai-0.3.31.dist-info/WHEEL,sha256=sP946D7jFCHeNz5Iq4fL4Lu-PrWrFsgfLXbbkciIZwg,88
9
+ llama_index_llms_openai-0.3.31.dist-info/RECORD,,
@@ -1,9 +0,0 @@
1
- llama_index/llms/openai/__init__.py,sha256=8nmgixeXifQ4eVSgtCic54WxXqrrpXQPL4rhACWCSFs,229
2
- llama_index/llms/openai/base.py,sha256=z4k8aFFt7av2CevyOFnV8LSauhHjZyaqAxLukbR54sM,38680
3
- llama_index/llms/openai/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
4
- llama_index/llms/openai/responses.py,sha256=TikqnpW-UQmgjmMYGznsBx7eEo5prNIaxE81RO1ZZjE,34465
5
- llama_index/llms/openai/utils.py,sha256=qp9qpXY7HbUnUsVDx6TgK98feibzTRi-bLdq_F3S0fo,26017
6
- llama_index_llms_openai-0.3.30.dist-info/LICENSE,sha256=JPQLUZD9rKvCTdu192Nk0V5PAwklIg6jANii3UmTyMs,1065
7
- llama_index_llms_openai-0.3.30.dist-info/METADATA,sha256=C615l-GASFs0vMxeo4YnoYgCQ5LpOeHdm-9fI7Oul9o,3322
8
- llama_index_llms_openai-0.3.30.dist-info/WHEEL,sha256=sP946D7jFCHeNz5Iq4fL4Lu-PrWrFsgfLXbbkciIZwg,88
9
- llama_index_llms_openai-0.3.30.dist-info/RECORD,,