gllm-inference-binary 0.5.18__cp311-cp311-macosx_13_0_x86_64.whl → 0.5.20__cp311-cp311-macosx_13_0_x86_64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of gllm-inference-binary might be problematic. Click here for more details.

@@ -107,7 +107,7 @@ def build_lm_invoker(model_id: str | ModelId, credentials: str | dict[str, Any]
107
107
  # Using Azure OpenAI
108
108
  ```python
109
109
  lm_invoker = build_lm_invoker(
110
- model_id="azure-openai/https://my-resource.openai.azure.com:my-deployment",
110
+ model_id="azure-openai/https://my-resource.openai.azure.com/openai/v1:my-deployment",
111
111
  credentials="azure-api-key"
112
112
  )
113
113
  ```
@@ -1,6 +1,5 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.utils.retry import RetryConfig as RetryConfig
3
- from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
4
3
  from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
5
4
  from gllm_inference.em_invoker.schema.twelvelabs import InputType as InputType, Key as Key, OutputType as OutputType
6
5
  from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EMContent as EMContent, ModelId as ModelId, ModelProvider as ModelProvider, TruncationConfig as TruncationConfig, Vector as Vector
@@ -1,13 +1,15 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.schema.tool import Tool as Tool
3
3
  from gllm_core.utils.retry import RetryConfig as RetryConfig
4
- from gllm_inference.constants import DEFAULT_AZURE_OPENAI_API_VERSION as DEFAULT_AZURE_OPENAI_API_VERSION, INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
4
+ from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
5
5
  from gllm_inference.lm_invoker.openai_lm_invoker import OpenAILMInvoker as OpenAILMInvoker, ReasoningEffort as ReasoningEffort, ReasoningSummary as ReasoningSummary
6
6
  from gllm_inference.lm_invoker.schema.openai import Key as Key
7
7
  from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema
8
8
  from langchain_core.tools import Tool as LangChainTool
9
9
  from typing import Any
10
10
 
11
+ URL_SUFFIX: str
12
+
11
13
  class AzureOpenAILMInvoker(OpenAILMInvoker):
12
14
  '''A language model invoker to interact with Azure OpenAI language models.
13
15
 
@@ -33,7 +35,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
33
35
  The `AzureOpenAILMInvoker` can be used as follows:
34
36
  ```python
35
37
  lm_invoker = AzureOpenAILMInvoker(
36
- azure_endpoint="https://<your-azure-openai-endpoint>.openai.azure.com/",
38
+ azure_endpoint="https://<your-azure-openai-endpoint>.openai.azure.com/openai/v1",
37
39
  azure_deployment="<your-azure-openai-deployment>",
38
40
  )
39
41
  result = await lm_invoker.invoke("Hi there!")
@@ -220,7 +222,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
220
222
  Defaults to an empty list.
221
223
  '''
222
224
  client: Incomplete
223
- def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None) -> None:
225
+ def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None) -> None:
224
226
  """Initializes a new instance of the AzureOpenAILMInvoker class.
225
227
 
226
228
  Args:
@@ -228,8 +230,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
228
230
  azure_deployment (str): The deployment name of the Azure OpenAI service.
229
231
  api_key (str | None, optional): The API key for authenticating with Azure OpenAI. Defaults to None, in
230
232
  which case the `AZURE_OPENAI_API_KEY` environment variable will be used.
231
- api_version (str, optional): The API version of the Azure OpenAI service. Defaults to
232
- `DEFAULT_AZURE_OPENAI_API_VERSION`.
233
+ api_version (str | None, optional): Deprecated parameter to be removed in v0.6. Defaults to None.
233
234
  model_kwargs (dict[str, Any] | None, optional): Additional model parameters. Defaults to None.
234
235
  default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
235
236
  Defaults to None.
@@ -57,7 +57,7 @@ class ModelId(BaseModel):
57
57
 
58
58
  # Using Azure OpenAI
59
59
  ```python
60
- model_id = ModelId.from_string("azure-openai/https://my-resource.openai.azure.com:my-deployment")
60
+ model_id = ModelId.from_string("azure-openai/https://my-resource.openai.azure.com/openai/v1:my-deployment")
61
61
  ```
62
62
 
63
63
  # Using OpenAI compatible endpoints (e.g. Groq)
Binary file
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: gllm-inference-binary
3
- Version: 0.5.18
3
+ Version: 0.5.20
4
4
  Summary: A library containing components related to model inferences in Gen AI applications.
5
5
  Author: Henry Wicaksono
6
6
  Author-email: henry.wicaksono@gdplabs.id
@@ -1,7 +1,7 @@
1
1
  gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
2
2
  gllm_inference/builder/__init__.pyi,sha256=usz2lvfwO4Yk-ZGKXbCWG1cEr3nlQXxMNDNC-2yc1NM,500
3
3
  gllm_inference/builder/build_em_invoker.pyi,sha256=7JTt8XpfqLQdv5cltIyi1r6Sf8MAUKgornHn5z57raA,5873
4
- gllm_inference/builder/build_lm_invoker.pyi,sha256=igJdLnWiY5QcdT4EClvgxFGSVZjARpq5hz2FYcBHWEQ,6876
4
+ gllm_inference/builder/build_lm_invoker.pyi,sha256=9-M0CJWhd0MVUYHHLD7QCfeW835JBaZ8_XQzUBJJdAc,6886
5
5
  gllm_inference/builder/build_lm_request_processor.pyi,sha256=33Gi3onftl-V2e_mkJios5zmXRKSoAVPX3UK7YBExjk,4491
6
6
  gllm_inference/builder/build_output_parser.pyi,sha256=_Lrq-bh1oPsb_Nwkkr_zyEUwIOMysRFZkvEtEM29LZM,936
7
7
  gllm_inference/catalog/__init__.pyi,sha256=JBkPGTyiiZ30GECzJBW-mW8LekWyY2qyzal3eW7ynaM,287
@@ -27,14 +27,14 @@ gllm_inference/em_invoker/schema/openai.pyi,sha256=Q_dsEcodkOXYXPdrkOkW0LnuLhfeq
27
27
  gllm_inference/em_invoker/schema/openai_compatible.pyi,sha256=gmvGtsWoOMBelke_tZjC6dKimFBW9f4Vrgv0Ig0OM9Q,150
28
28
  gllm_inference/em_invoker/schema/twelvelabs.pyi,sha256=F6wKHgG01bYskJpKoheBSpRpHUfFpteKn9sj9n5YfcU,372
29
29
  gllm_inference/em_invoker/schema/voyage.pyi,sha256=HVpor0fqNy-IwapCICfsgFmqf1FJXCOMIxS2vOXhHd8,289
30
- gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=RYY4td4N3cPSy-yXTEUKwYLz0AH8mLxi2j5Bfr9PS0g,5447
30
+ gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=9F37VuLYTH5xsPSS_uzhrsPJug0QifX_qh2GwG0jSTU,5345
31
31
  gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=zJZqMvvFKu3sHdrNM773UYjfHVlnwE2w2BmvdFcHzV0,5515
32
32
  gllm_inference/exceptions/__init__.pyi,sha256=v9uxjW5DssIn7n_bKqT7L83CeqFET2Z45GFOvi78UuE,977
33
33
  gllm_inference/exceptions/error_parser.pyi,sha256=4RkVfS2Fl9kjz_h2bK9eoAeI-Y-VkHcUqXWj68BsYig,2393
34
34
  gllm_inference/exceptions/exceptions.pyi,sha256=5YRackwVNvyOJjOtiVszqu8q87s8ioXTa-XwaYmeiC4,4643
35
35
  gllm_inference/lm_invoker/__init__.pyi,sha256=NmQSqObPjevEP1KbbrNnaz4GMh175EVPERZ19vK5Emc,1202
36
36
  gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=MsF3OmDo0L9aEHuTJYTgsoDILi2B_IgKtPpDcDMduWc,14925
37
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=EV_yrj6mnV_rCDEqSZaIikfw76_rXXyDlC-w_y-m7K0,14603
37
+ gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=AoEC8GdPW2LAyiCfH7CoSGHRTlVUcteVx0WTIwkBljI,14527
38
38
  gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=IuLxgCThOSBHx1AXqnhL6yVu5_JV6hAeGBWWm5P1JCo,12423
39
39
  gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=QS84w3WpD3Oyl5HdxrucsadCmsHE8gn6Ewl3l01DCgI,9203
40
40
  gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=aPOlaw2rexUDhbMFaXnuKqOT7lqeKxjfeToe9LjwEUw,16787
@@ -85,7 +85,7 @@ gllm_inference/schema/config.pyi,sha256=rAL_UeXyQeXVk1P2kqd8vFWOMwmKenfpQLtvMP74
85
85
  gllm_inference/schema/enums.pyi,sha256=w5Bq3m-Ixl4yAd4801APhw9fjCiuqttWuUXWvSWSEEs,717
86
86
  gllm_inference/schema/lm_output.pyi,sha256=GafJV0KeD-VSwWkwG1oz-uruXrQ7KDZTuoojPCBRpg8,1956
87
87
  gllm_inference/schema/message.pyi,sha256=VP9YppKj2mo1esl9cy6qQO9m2mMHUjTmfGDdyUor880,2220
88
- gllm_inference/schema/model_id.pyi,sha256=0UJS7M91hPlzWuZI3CMGZm9ewYrTxmLUMtIhHqJOg0Q,5481
88
+ gllm_inference/schema/model_id.pyi,sha256=qrr0x4qkd6cGIbc4XATWJb0uckKhd1sAdR_xT7vGIXI,5491
89
89
  gllm_inference/schema/reasoning.pyi,sha256=SlTuiDw87GdnAn-I6YOPIJRhEBiwQljM46JohG05guQ,562
90
90
  gllm_inference/schema/token_usage.pyi,sha256=1GTQVORV0dBNmD_jix8aVaUqxMKFF04KpLP7y2urqbk,2950
91
91
  gllm_inference/schema/tool_call.pyi,sha256=zQaVxCnkVxOfOEhBidqohU85gb4PRwnwBiygKaunamk,389
@@ -96,8 +96,8 @@ gllm_inference/utils/io_utils.pyi,sha256=7kUTacHAVRYoemFUOjCH7-Qmw-YsQGd6rGYxjf_
96
96
  gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
97
97
  gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
98
98
  gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
99
- gllm_inference.cpython-311-darwin.so,sha256=cAJinH-x194WwhgzWdn8K5-D6NYWyAYCXEDxAgJLRH0,4235808
99
+ gllm_inference.cpython-311-darwin.so,sha256=4kpDVeQ3MgeYGTSnTFQhMTwp0pZ5nghcURmhP5inPzE,4236040
100
100
  gllm_inference.pyi,sha256=lTVixRzlC12Joi4kW_vxnux0rLHAUB_3j7RMFOwLK-M,3616
101
- gllm_inference_binary-0.5.18.dist-info/METADATA,sha256=Heo4b0XS1gQgOtbWPzIYPTic4-esBmsLMJ2LO68SP3U,4608
102
- gllm_inference_binary-0.5.18.dist-info/WHEEL,sha256=r3EiIdyNg8wC0u2K9wWWq7Elb6S4XGGmkyBqljSOtNU,107
103
- gllm_inference_binary-0.5.18.dist-info/RECORD,,
101
+ gllm_inference_binary-0.5.20.dist-info/METADATA,sha256=Uq49Bd6PIOb23ezYyti27LutHIe-fqCtcpC9rbv2PNE,4608
102
+ gllm_inference_binary-0.5.20.dist-info/WHEEL,sha256=r3EiIdyNg8wC0u2K9wWWq7Elb6S4XGGmkyBqljSOtNU,107
103
+ gllm_inference_binary-0.5.20.dist-info/RECORD,,