llama-index-llms-openai 0.5.2__tar.gz → 0.5.4__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/PKG-INFO +1 -1
- {llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/llama_index/llms/openai/base.py +3 -9
- {llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/llama_index/llms/openai/utils.py +1 -0
- {llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/pyproject.toml +1 -1
- {llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/.gitignore +0 -0
- {llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/LICENSE +0 -0
- {llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/README.md +0 -0
- {llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/llama_index/llms/openai/__init__.py +0 -0
- {llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/llama_index/llms/openai/py.typed +0 -0
- {llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/llama_index/llms/openai/responses.py +0 -0
{llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/llama_index/llms/openai/base.py
RENAMED
|
@@ -77,7 +77,7 @@ from llama_index.llms.openai.utils import (
|
|
|
77
77
|
update_tool_calls,
|
|
78
78
|
is_json_schema_supported,
|
|
79
79
|
)
|
|
80
|
-
from openai import AsyncOpenAI
|
|
80
|
+
from openai import AsyncOpenAI
|
|
81
81
|
from openai import OpenAI as SyncOpenAI
|
|
82
82
|
from openai.types.chat.chat_completion_chunk import (
|
|
83
83
|
ChatCompletionChunk,
|
|
@@ -532,10 +532,7 @@ class OpenAI(FunctionCallingLLM):
|
|
|
532
532
|
if len(response.choices) > 0:
|
|
533
533
|
delta = response.choices[0].delta
|
|
534
534
|
else:
|
|
535
|
-
|
|
536
|
-
continue
|
|
537
|
-
else:
|
|
538
|
-
delta = ChoiceDelta()
|
|
535
|
+
delta = ChoiceDelta()
|
|
539
536
|
|
|
540
537
|
if delta is None:
|
|
541
538
|
continue
|
|
@@ -801,10 +798,7 @@ class OpenAI(FunctionCallingLLM):
|
|
|
801
798
|
continue
|
|
802
799
|
delta = response.choices[0].delta
|
|
803
800
|
else:
|
|
804
|
-
|
|
805
|
-
continue
|
|
806
|
-
else:
|
|
807
|
-
delta = ChoiceDelta()
|
|
801
|
+
delta = ChoiceDelta()
|
|
808
802
|
first_chat_chunk = False
|
|
809
803
|
|
|
810
804
|
if delta is None:
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/llama_index/llms/openai/__init__.py
RENAMED
|
File without changes
|
{llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/llama_index/llms/openai/py.typed
RENAMED
|
File without changes
|
{llama_index_llms_openai-0.5.2 → llama_index_llms_openai-0.5.4}/llama_index/llms/openai/responses.py
RENAMED
|
File without changes
|