gllm-inference-binary 0.5.19__cp312-cp312-manylinux_2_31_x86_64.whl → 0.5.20__cp312-cp312-manylinux_2_31_x86_64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -107,7 +107,7 @@ def build_lm_invoker(model_id: str | ModelId, credentials: str | dict[str, Any]
107
107
  # Using Azure OpenAI
108
108
  ```python
109
109
  lm_invoker = build_lm_invoker(
110
- model_id="azure-openai/https://my-resource.openai.azure.com:my-deployment",
110
+ model_id="azure-openai/https://my-resource.openai.azure.com/openai/v1:my-deployment",
111
111
  credentials="azure-api-key"
112
112
  )
113
113
  ```
@@ -1,13 +1,15 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.schema.tool import Tool as Tool
3
3
  from gllm_core.utils.retry import RetryConfig as RetryConfig
4
- from gllm_inference.constants import DEFAULT_AZURE_OPENAI_API_VERSION as DEFAULT_AZURE_OPENAI_API_VERSION, INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
4
+ from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
5
5
  from gllm_inference.lm_invoker.openai_lm_invoker import OpenAILMInvoker as OpenAILMInvoker, ReasoningEffort as ReasoningEffort, ReasoningSummary as ReasoningSummary
6
6
  from gllm_inference.lm_invoker.schema.openai import Key as Key
7
7
  from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema
8
8
  from langchain_core.tools import Tool as LangChainTool
9
9
  from typing import Any
10
10
 
11
+ URL_SUFFIX: str
12
+
11
13
  class AzureOpenAILMInvoker(OpenAILMInvoker):
12
14
  '''A language model invoker to interact with Azure OpenAI language models.
13
15
 
@@ -33,7 +35,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
33
35
  The `AzureOpenAILMInvoker` can be used as follows:
34
36
  ```python
35
37
  lm_invoker = AzureOpenAILMInvoker(
36
- azure_endpoint="https://<your-azure-openai-endpoint>.openai.azure.com/",
38
+ azure_endpoint="https://<your-azure-openai-endpoint>.openai.azure.com/openai/v1",
37
39
  azure_deployment="<your-azure-openai-deployment>",
38
40
  )
39
41
  result = await lm_invoker.invoke("Hi there!")
@@ -220,7 +222,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
220
222
  Defaults to an empty list.
221
223
  '''
222
224
  client: Incomplete
223
- def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None) -> None:
225
+ def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None) -> None:
224
226
  """Initializes a new instance of the AzureOpenAILMInvoker class.
225
227
 
226
228
  Args:
@@ -228,8 +230,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
228
230
  azure_deployment (str): The deployment name of the Azure OpenAI service.
229
231
  api_key (str | None, optional): The API key for authenticating with Azure OpenAI. Defaults to None, in
230
232
  which case the `AZURE_OPENAI_API_KEY` environment variable will be used.
231
- api_version (str, optional): The API version of the Azure OpenAI service. Defaults to
232
- `DEFAULT_AZURE_OPENAI_API_VERSION`.
233
+ api_version (str | None, optional): Deprecated parameter to be removed in v0.6. Defaults to None.
233
234
  model_kwargs (dict[str, Any] | None, optional): Additional model parameters. Defaults to None.
234
235
  default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
235
236
  Defaults to None.
@@ -57,7 +57,7 @@ class ModelId(BaseModel):
57
57
 
58
58
  # Using Azure OpenAI
59
59
  ```python
60
- model_id = ModelId.from_string("azure-openai/https://my-resource.openai.azure.com:my-deployment")
60
+ model_id = ModelId.from_string("azure-openai/https://my-resource.openai.azure.com/openai/v1:my-deployment")
61
61
  ```
62
62
 
63
63
  # Using OpenAI compatible endpoints (e.g. Groq)
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: gllm-inference-binary
3
- Version: 0.5.19
3
+ Version: 0.5.20
4
4
  Summary: A library containing components related to model inferences in Gen AI applications.
5
5
  Author: Henry Wicaksono
6
6
  Author-email: henry.wicaksono@gdplabs.id
@@ -1,7 +1,7 @@
1
1
  gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
2
2
  gllm_inference/builder/__init__.pyi,sha256=usz2lvfwO4Yk-ZGKXbCWG1cEr3nlQXxMNDNC-2yc1NM,500
3
3
  gllm_inference/builder/build_em_invoker.pyi,sha256=7JTt8XpfqLQdv5cltIyi1r6Sf8MAUKgornHn5z57raA,5873
4
- gllm_inference/builder/build_lm_invoker.pyi,sha256=igJdLnWiY5QcdT4EClvgxFGSVZjARpq5hz2FYcBHWEQ,6876
4
+ gllm_inference/builder/build_lm_invoker.pyi,sha256=9-M0CJWhd0MVUYHHLD7QCfeW835JBaZ8_XQzUBJJdAc,6886
5
5
  gllm_inference/builder/build_lm_request_processor.pyi,sha256=33Gi3onftl-V2e_mkJios5zmXRKSoAVPX3UK7YBExjk,4491
6
6
  gllm_inference/builder/build_output_parser.pyi,sha256=_Lrq-bh1oPsb_Nwkkr_zyEUwIOMysRFZkvEtEM29LZM,936
7
7
  gllm_inference/catalog/__init__.pyi,sha256=JBkPGTyiiZ30GECzJBW-mW8LekWyY2qyzal3eW7ynaM,287
@@ -34,7 +34,7 @@ gllm_inference/exceptions/error_parser.pyi,sha256=4RkVfS2Fl9kjz_h2bK9eoAeI-Y-VkH
34
34
  gllm_inference/exceptions/exceptions.pyi,sha256=5YRackwVNvyOJjOtiVszqu8q87s8ioXTa-XwaYmeiC4,4643
35
35
  gllm_inference/lm_invoker/__init__.pyi,sha256=NmQSqObPjevEP1KbbrNnaz4GMh175EVPERZ19vK5Emc,1202
36
36
  gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=MsF3OmDo0L9aEHuTJYTgsoDILi2B_IgKtPpDcDMduWc,14925
37
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=EV_yrj6mnV_rCDEqSZaIikfw76_rXXyDlC-w_y-m7K0,14603
37
+ gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=AoEC8GdPW2LAyiCfH7CoSGHRTlVUcteVx0WTIwkBljI,14527
38
38
  gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=IuLxgCThOSBHx1AXqnhL6yVu5_JV6hAeGBWWm5P1JCo,12423
39
39
  gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=QS84w3WpD3Oyl5HdxrucsadCmsHE8gn6Ewl3l01DCgI,9203
40
40
  gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=aPOlaw2rexUDhbMFaXnuKqOT7lqeKxjfeToe9LjwEUw,16787
@@ -85,7 +85,7 @@ gllm_inference/schema/config.pyi,sha256=rAL_UeXyQeXVk1P2kqd8vFWOMwmKenfpQLtvMP74
85
85
  gllm_inference/schema/enums.pyi,sha256=w5Bq3m-Ixl4yAd4801APhw9fjCiuqttWuUXWvSWSEEs,717
86
86
  gllm_inference/schema/lm_output.pyi,sha256=GafJV0KeD-VSwWkwG1oz-uruXrQ7KDZTuoojPCBRpg8,1956
87
87
  gllm_inference/schema/message.pyi,sha256=VP9YppKj2mo1esl9cy6qQO9m2mMHUjTmfGDdyUor880,2220
88
- gllm_inference/schema/model_id.pyi,sha256=0UJS7M91hPlzWuZI3CMGZm9ewYrTxmLUMtIhHqJOg0Q,5481
88
+ gllm_inference/schema/model_id.pyi,sha256=qrr0x4qkd6cGIbc4XATWJb0uckKhd1sAdR_xT7vGIXI,5491
89
89
  gllm_inference/schema/reasoning.pyi,sha256=SlTuiDw87GdnAn-I6YOPIJRhEBiwQljM46JohG05guQ,562
90
90
  gllm_inference/schema/token_usage.pyi,sha256=1GTQVORV0dBNmD_jix8aVaUqxMKFF04KpLP7y2urqbk,2950
91
91
  gllm_inference/schema/tool_call.pyi,sha256=zQaVxCnkVxOfOEhBidqohU85gb4PRwnwBiygKaunamk,389
@@ -96,8 +96,8 @@ gllm_inference/utils/io_utils.pyi,sha256=7kUTacHAVRYoemFUOjCH7-Qmw-YsQGd6rGYxjf_
96
96
  gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
97
97
  gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
98
98
  gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
99
- gllm_inference.cpython-312-x86_64-linux-gnu.so,sha256=r_B-gLt9eHTaHWVjZACjKweElve13eOpQo8aOmRlVLA,4383944
99
+ gllm_inference.cpython-312-x86_64-linux-gnu.so,sha256=BGrm24-rWbhByl5sarWz1w76yWSbS7ItVtsebevHISw,4388040
100
100
  gllm_inference.pyi,sha256=lTVixRzlC12Joi4kW_vxnux0rLHAUB_3j7RMFOwLK-M,3616
101
- gllm_inference_binary-0.5.19.dist-info/METADATA,sha256=RYPC2mk8-uXt5cFOZh7VdB6ccfFPjZ7AjvSsCoCfZQI,4608
102
- gllm_inference_binary-0.5.19.dist-info/WHEEL,sha256=mNY4pwQL4AOAoPmLYEQs2SSpMIbATFeiJFktRD5iKkY,110
103
- gllm_inference_binary-0.5.19.dist-info/RECORD,,
101
+ gllm_inference_binary-0.5.20.dist-info/METADATA,sha256=Uq49Bd6PIOb23ezYyti27LutHIe-fqCtcpC9rbv2PNE,4608
102
+ gllm_inference_binary-0.5.20.dist-info/WHEEL,sha256=mNY4pwQL4AOAoPmLYEQs2SSpMIbATFeiJFktRD5iKkY,110
103
+ gllm_inference_binary-0.5.20.dist-info/RECORD,,