llama-index-llms-openai 0.3.16__tar.gz → 0.3.17__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: llama-index-llms-openai
3
- Version: 0.3.16
3
+ Version: 0.3.17
4
4
  Summary: llama-index llms openai integration
5
5
  License: MIT
6
6
  Author: llama-index
@@ -424,7 +424,7 @@ class OpenAI(FunctionCallingLLM):
424
424
  all_kwargs = {**base_kwargs, **self.additional_kwargs}
425
425
  if "stream" not in all_kwargs and "stream_options" in all_kwargs:
426
426
  del all_kwargs["stream_options"]
427
- if self.model in O1_MODELS and base_kwargs["max_tokens"] is not None:
427
+ if self.model in O1_MODELS and base_kwargs.get("max_tokens") is not None:
428
428
  # O1 models use max_completion_tokens instead of max_tokens
429
429
  all_kwargs["max_completion_tokens"] = all_kwargs.get(
430
430
  "max_completion_tokens", all_kwargs["max_tokens"]
@@ -29,7 +29,7 @@ exclude = ["**/BUILD"]
29
29
  license = "MIT"
30
30
  name = "llama-index-llms-openai"
31
31
  readme = "README.md"
32
- version = "0.3.16"
32
+ version = "0.3.17"
33
33
 
34
34
  [tool.poetry.dependencies]
35
35
  python = ">=3.9,<4.0"