gllm-inference-binary 0.5.19__cp313-cp313-macosx_13_0_x86_64.whl → 0.5.20__cp313-cp313-macosx_13_0_x86_64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of gllm-inference-binary might be problematic. Click here for more details.
- gllm_inference/builder/build_lm_invoker.pyi +1 -1
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi +6 -5
- gllm_inference/schema/model_id.pyi +1 -1
- gllm_inference.cpython-313-darwin.so +0 -0
- {gllm_inference_binary-0.5.19.dist-info → gllm_inference_binary-0.5.20.dist-info}/METADATA +1 -1
- {gllm_inference_binary-0.5.19.dist-info → gllm_inference_binary-0.5.20.dist-info}/RECORD +7 -7
- {gllm_inference_binary-0.5.19.dist-info → gllm_inference_binary-0.5.20.dist-info}/WHEEL +0 -0
|
@@ -107,7 +107,7 @@ def build_lm_invoker(model_id: str | ModelId, credentials: str | dict[str, Any]
|
|
|
107
107
|
# Using Azure OpenAI
|
|
108
108
|
```python
|
|
109
109
|
lm_invoker = build_lm_invoker(
|
|
110
|
-
model_id="azure-openai/https://my-resource.openai.azure.com:my-deployment",
|
|
110
|
+
model_id="azure-openai/https://my-resource.openai.azure.com/openai/v1:my-deployment",
|
|
111
111
|
credentials="azure-api-key"
|
|
112
112
|
)
|
|
113
113
|
```
|
|
@@ -1,13 +1,15 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.schema.tool import Tool as Tool
|
|
3
3
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
4
|
-
from gllm_inference.constants import
|
|
4
|
+
from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
5
5
|
from gllm_inference.lm_invoker.openai_lm_invoker import OpenAILMInvoker as OpenAILMInvoker, ReasoningEffort as ReasoningEffort, ReasoningSummary as ReasoningSummary
|
|
6
6
|
from gllm_inference.lm_invoker.schema.openai import Key as Key
|
|
7
7
|
from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema
|
|
8
8
|
from langchain_core.tools import Tool as LangChainTool
|
|
9
9
|
from typing import Any
|
|
10
10
|
|
|
11
|
+
URL_SUFFIX: str
|
|
12
|
+
|
|
11
13
|
class AzureOpenAILMInvoker(OpenAILMInvoker):
|
|
12
14
|
'''A language model invoker to interact with Azure OpenAI language models.
|
|
13
15
|
|
|
@@ -33,7 +35,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
|
|
|
33
35
|
The `AzureOpenAILMInvoker` can be used as follows:
|
|
34
36
|
```python
|
|
35
37
|
lm_invoker = AzureOpenAILMInvoker(
|
|
36
|
-
azure_endpoint="https://<your-azure-openai-endpoint>.openai.azure.com/",
|
|
38
|
+
azure_endpoint="https://<your-azure-openai-endpoint>.openai.azure.com/openai/v1",
|
|
37
39
|
azure_deployment="<your-azure-openai-deployment>",
|
|
38
40
|
)
|
|
39
41
|
result = await lm_invoker.invoke("Hi there!")
|
|
@@ -220,7 +222,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
|
|
|
220
222
|
Defaults to an empty list.
|
|
221
223
|
'''
|
|
222
224
|
client: Incomplete
|
|
223
|
-
def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str =
|
|
225
|
+
def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None) -> None:
|
|
224
226
|
"""Initializes a new instance of the AzureOpenAILMInvoker class.
|
|
225
227
|
|
|
226
228
|
Args:
|
|
@@ -228,8 +230,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
|
|
|
228
230
|
azure_deployment (str): The deployment name of the Azure OpenAI service.
|
|
229
231
|
api_key (str | None, optional): The API key for authenticating with Azure OpenAI. Defaults to None, in
|
|
230
232
|
which case the `AZURE_OPENAI_API_KEY` environment variable will be used.
|
|
231
|
-
api_version (str, optional):
|
|
232
|
-
`DEFAULT_AZURE_OPENAI_API_VERSION`.
|
|
233
|
+
api_version (str | None, optional): Deprecated parameter to be removed in v0.6. Defaults to None.
|
|
233
234
|
model_kwargs (dict[str, Any] | None, optional): Additional model parameters. Defaults to None.
|
|
234
235
|
default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
|
|
235
236
|
Defaults to None.
|
|
@@ -57,7 +57,7 @@ class ModelId(BaseModel):
|
|
|
57
57
|
|
|
58
58
|
# Using Azure OpenAI
|
|
59
59
|
```python
|
|
60
|
-
model_id = ModelId.from_string("azure-openai/https://my-resource.openai.azure.com:my-deployment")
|
|
60
|
+
model_id = ModelId.from_string("azure-openai/https://my-resource.openai.azure.com/openai/v1:my-deployment")
|
|
61
61
|
```
|
|
62
62
|
|
|
63
63
|
# Using OpenAI compatible endpoints (e.g. Groq)
|
|
Binary file
|
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
2
2
|
gllm_inference/builder/__init__.pyi,sha256=usz2lvfwO4Yk-ZGKXbCWG1cEr3nlQXxMNDNC-2yc1NM,500
|
|
3
3
|
gllm_inference/builder/build_em_invoker.pyi,sha256=7JTt8XpfqLQdv5cltIyi1r6Sf8MAUKgornHn5z57raA,5873
|
|
4
|
-
gllm_inference/builder/build_lm_invoker.pyi,sha256=
|
|
4
|
+
gllm_inference/builder/build_lm_invoker.pyi,sha256=9-M0CJWhd0MVUYHHLD7QCfeW835JBaZ8_XQzUBJJdAc,6886
|
|
5
5
|
gllm_inference/builder/build_lm_request_processor.pyi,sha256=33Gi3onftl-V2e_mkJios5zmXRKSoAVPX3UK7YBExjk,4491
|
|
6
6
|
gllm_inference/builder/build_output_parser.pyi,sha256=_Lrq-bh1oPsb_Nwkkr_zyEUwIOMysRFZkvEtEM29LZM,936
|
|
7
7
|
gllm_inference/catalog/__init__.pyi,sha256=JBkPGTyiiZ30GECzJBW-mW8LekWyY2qyzal3eW7ynaM,287
|
|
@@ -34,7 +34,7 @@ gllm_inference/exceptions/error_parser.pyi,sha256=4RkVfS2Fl9kjz_h2bK9eoAeI-Y-VkH
|
|
|
34
34
|
gllm_inference/exceptions/exceptions.pyi,sha256=5YRackwVNvyOJjOtiVszqu8q87s8ioXTa-XwaYmeiC4,4643
|
|
35
35
|
gllm_inference/lm_invoker/__init__.pyi,sha256=NmQSqObPjevEP1KbbrNnaz4GMh175EVPERZ19vK5Emc,1202
|
|
36
36
|
gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=MsF3OmDo0L9aEHuTJYTgsoDILi2B_IgKtPpDcDMduWc,14925
|
|
37
|
-
gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=
|
|
37
|
+
gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=AoEC8GdPW2LAyiCfH7CoSGHRTlVUcteVx0WTIwkBljI,14527
|
|
38
38
|
gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=IuLxgCThOSBHx1AXqnhL6yVu5_JV6hAeGBWWm5P1JCo,12423
|
|
39
39
|
gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=QS84w3WpD3Oyl5HdxrucsadCmsHE8gn6Ewl3l01DCgI,9203
|
|
40
40
|
gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=aPOlaw2rexUDhbMFaXnuKqOT7lqeKxjfeToe9LjwEUw,16787
|
|
@@ -85,7 +85,7 @@ gllm_inference/schema/config.pyi,sha256=rAL_UeXyQeXVk1P2kqd8vFWOMwmKenfpQLtvMP74
|
|
|
85
85
|
gllm_inference/schema/enums.pyi,sha256=w5Bq3m-Ixl4yAd4801APhw9fjCiuqttWuUXWvSWSEEs,717
|
|
86
86
|
gllm_inference/schema/lm_output.pyi,sha256=GafJV0KeD-VSwWkwG1oz-uruXrQ7KDZTuoojPCBRpg8,1956
|
|
87
87
|
gllm_inference/schema/message.pyi,sha256=VP9YppKj2mo1esl9cy6qQO9m2mMHUjTmfGDdyUor880,2220
|
|
88
|
-
gllm_inference/schema/model_id.pyi,sha256=
|
|
88
|
+
gllm_inference/schema/model_id.pyi,sha256=qrr0x4qkd6cGIbc4XATWJb0uckKhd1sAdR_xT7vGIXI,5491
|
|
89
89
|
gllm_inference/schema/reasoning.pyi,sha256=SlTuiDw87GdnAn-I6YOPIJRhEBiwQljM46JohG05guQ,562
|
|
90
90
|
gllm_inference/schema/token_usage.pyi,sha256=1GTQVORV0dBNmD_jix8aVaUqxMKFF04KpLP7y2urqbk,2950
|
|
91
91
|
gllm_inference/schema/tool_call.pyi,sha256=zQaVxCnkVxOfOEhBidqohU85gb4PRwnwBiygKaunamk,389
|
|
@@ -96,8 +96,8 @@ gllm_inference/utils/io_utils.pyi,sha256=7kUTacHAVRYoemFUOjCH7-Qmw-YsQGd6rGYxjf_
|
|
|
96
96
|
gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
|
|
97
97
|
gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
|
|
98
98
|
gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
|
|
99
|
-
gllm_inference.cpython-313-darwin.so,sha256=
|
|
99
|
+
gllm_inference.cpython-313-darwin.so,sha256=Jxr6R2qWo4_MezMbK1Sj4kV44dbhtfmEpgWj7kt2CLw,4191016
|
|
100
100
|
gllm_inference.pyi,sha256=uxl1voKdn19LurAHKEZLWbq9ryPO4UkJ1Nk1MM8IL34,3636
|
|
101
|
-
gllm_inference_binary-0.5.
|
|
102
|
-
gllm_inference_binary-0.5.
|
|
103
|
-
gllm_inference_binary-0.5.
|
|
101
|
+
gllm_inference_binary-0.5.20.dist-info/METADATA,sha256=Uq49Bd6PIOb23ezYyti27LutHIe-fqCtcpC9rbv2PNE,4608
|
|
102
|
+
gllm_inference_binary-0.5.20.dist-info/WHEEL,sha256=PCOZcL_jcbAVhuFR5ylE4Mr-7HPGHAcfJk9OpuMh7RQ,107
|
|
103
|
+
gllm_inference_binary-0.5.20.dist-info/RECORD,,
|
|
File without changes
|