gllm-inference-binary 0.5.1.post1__cp312-cp312-manylinux_2_31_x86_64.whl → 0.5.2__cp312-cp312-manylinux_2_31_x86_64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of gllm-inference-binary might be problematic. Click here for more details.

@@ -3,5 +3,6 @@ from _typeshed import Incomplete
3
3
  DEFAULT_AZURE_OPENAI_API_VERSION: str
4
4
  DOCUMENT_MIME_TYPES: Incomplete
5
5
  GOOGLE_SCOPES: Incomplete
6
+ INVOKER_DEFAULT_RETRY_CONFIG: Incomplete
6
7
  HEX_REPR_LENGTH: int
7
8
  HTTP_STATUS_CODE_PATTERNS: Incomplete
@@ -1,6 +1,6 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.utils.retry import RetryConfig as RetryConfig
3
- from gllm_inference.constants import DEFAULT_AZURE_OPENAI_API_VERSION as DEFAULT_AZURE_OPENAI_API_VERSION
3
+ from gllm_inference.constants import DEFAULT_AZURE_OPENAI_API_VERSION as DEFAULT_AZURE_OPENAI_API_VERSION, INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
4
4
  from gllm_inference.em_invoker.openai_em_invoker import OpenAIEMInvoker as OpenAIEMInvoker
5
5
  from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider
6
6
  from typing import Any
@@ -67,7 +67,7 @@ class AzureOpenAIEMInvoker(OpenAIEMInvoker):
67
67
  ```
68
68
  '''
69
69
  client: Incomplete
70
- def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None) -> None:
70
+ def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig = ...) -> None:
71
71
  """Initializes a new instance of the AzureOpenAIEMInvoker class.
72
72
 
73
73
  Args:
@@ -80,6 +80,6 @@ class AzureOpenAIEMInvoker(OpenAIEMInvoker):
80
80
  model_kwargs (dict[str, Any] | None, optional): Additional model parameters. Defaults to None.
81
81
  default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
82
82
  Defaults to None.
83
- retry_config (RetryConfig | None, optional): The retry configuration for the embedding model.
84
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
83
+ retry_config (RetryConfig, optional): The retry configuration for the embedding model.
84
+ Defaults to a default config with no retry and 30.0 seconds timeout.
85
85
  """
@@ -1,8 +1,8 @@
1
1
  import abc
2
2
  from _typeshed import Incomplete
3
3
  from abc import ABC
4
- from gllm_core.utils.retry import RetryConfig
5
- from gllm_inference.constants import DOCUMENT_MIME_TYPES as DOCUMENT_MIME_TYPES
4
+ from gllm_core.utils.retry import RetryConfig as RetryConfig
5
+ from gllm_inference.constants import DOCUMENT_MIME_TYPES as DOCUMENT_MIME_TYPES, INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
6
6
  from gllm_inference.exceptions import parse_error_message as parse_error_message
7
7
  from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EMContent as EMContent, ModelId as ModelId, Vector as Vector
8
8
  from typing import Any
@@ -21,7 +21,7 @@ class BaseEMInvoker(ABC, metaclass=abc.ABCMeta):
21
21
  """
22
22
  default_hyperparameters: Incomplete
23
23
  retry_config: Incomplete
24
- def __init__(self, model_id: ModelId, default_hyperparameters: dict[str, Any] | None = None, supported_attachments: set[str] | None = None, retry_config: RetryConfig | None = None) -> None:
24
+ def __init__(self, model_id: ModelId, default_hyperparameters: dict[str, Any] | None = None, supported_attachments: set[str] | None = None, retry_config: RetryConfig = ...) -> None:
25
25
  """Initializes a new instance of the BaseEMInvoker class.
26
26
 
27
27
  Args:
@@ -30,8 +30,8 @@ class BaseEMInvoker(ABC, metaclass=abc.ABCMeta):
30
30
  embedding model. Defaults to None, in which case an empty dictionary is used.
31
31
  supported_attachments (set[str] | None, optional): A set of supported attachment types. Defaults to None,
32
32
  in which case an empty set is used (indicating that no attachments are supported).
33
- retry_config (RetryConfig | None, optional): The retry configuration for the embedding model.
34
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
33
+ retry_config (RetryConfig, optional): The retry configuration for the embedding model.
34
+ Defaults to a default config with no retry and 30.0 seconds timeout.
35
35
  """
36
36
  @property
37
37
  def model_id(self) -> str:
@@ -1,6 +1,6 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.utils.retry import RetryConfig as RetryConfig
3
- from gllm_inference.constants import GOOGLE_SCOPES as GOOGLE_SCOPES
3
+ from gllm_inference.constants import GOOGLE_SCOPES as GOOGLE_SCOPES, INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
4
4
  from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
5
5
  from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider, Vector as Vector
6
6
  from typing import Any
@@ -97,7 +97,7 @@ class GoogleEMInvoker(BaseEMInvoker):
97
97
  ```
98
98
  '''
99
99
  client_params: Incomplete
100
- def __init__(self, model_name: str, api_key: str | None = None, credentials_path: str | None = None, project_id: str | None = None, location: str = 'us-central1', model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None) -> None:
100
+ def __init__(self, model_name: str, api_key: str | None = None, credentials_path: str | None = None, project_id: str | None = None, location: str = 'us-central1', model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig = ...) -> None:
101
101
  '''Initializes a new instance of the GoogleEMInvoker class.
102
102
 
103
103
  Args:
@@ -114,8 +114,8 @@ class GoogleEMInvoker(BaseEMInvoker):
114
114
  Defaults to None.
115
115
  default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
116
116
  Defaults to None.
117
- retry_config (RetryConfig | None, optional): The retry configuration for the embedding model.
118
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
117
+ retry_config (RetryConfig, optional): The retry configuration for the embedding model.
118
+ Defaults to a default config with no retry and 30.0 seconds timeout.
119
119
 
120
120
  Note:
121
121
  If neither `api_key` nor `credentials_path` is provided, Google Gen AI will be used by default.
@@ -1,5 +1,6 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.utils.retry import RetryConfig as RetryConfig
3
+ from gllm_inference.constants import INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
3
4
  from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
4
5
  from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider, Vector as Vector
5
6
  from gllm_inference.utils import load_langchain_model as load_langchain_model, parse_model_data as parse_model_data
@@ -19,7 +20,7 @@ class LangChainEMInvoker(BaseEMInvoker):
19
20
  retry_config (RetryConfig): The retry configuration for the embedding model.
20
21
  """
21
22
  model: Incomplete
22
- def __init__(self, model: Embeddings | None = None, model_class_path: str | None = None, model_name: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None) -> None:
23
+ def __init__(self, model: Embeddings | None = None, model_class_path: str | None = None, model_name: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig = ...) -> None:
23
24
  '''Initializes a new instance of the LangChainEMInvoker class.
24
25
 
25
26
  Args:
@@ -34,6 +35,6 @@ class LangChainEMInvoker(BaseEMInvoker):
34
35
  `model_class_path` is provided. Defaults to None.
35
36
  default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
36
37
  Defaults to None.
37
- retry_config (RetryConfig | None, optional): The retry configuration for the embedding model.
38
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
38
+ retry_config (RetryConfig, optional): The retry configuration for the embedding model.
39
+ Defaults to a default config with no retry and 30.0 seconds timeout.
39
40
  '''
@@ -1,5 +1,6 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.utils.retry import RetryConfig as RetryConfig
3
+ from gllm_inference.constants import INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
3
4
  from gllm_inference.em_invoker.openai_em_invoker import OpenAIEMInvoker as OpenAIEMInvoker
4
5
  from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider
5
6
  from typing import Any
@@ -74,7 +75,7 @@ class OpenAICompatibleEMInvoker(OpenAIEMInvoker):
74
75
  ```
75
76
  '''
76
77
  client: Incomplete
77
- def __init__(self, model_name: str, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None) -> None:
78
+ def __init__(self, model_name: str, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig = ...) -> None:
78
79
  """Initializes a new instance of the OpenAICompatibleEMInvoker class.
79
80
 
80
81
  Args:
@@ -85,6 +86,6 @@ class OpenAICompatibleEMInvoker(OpenAIEMInvoker):
85
86
  model_kwargs (dict[str, Any] | None, optional): Additional model parameters. Defaults to None.
86
87
  default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
87
88
  Defaults to None.
88
- retry_config (RetryConfig | None, optional): The retry configuration for the embedding model.
89
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
89
+ retry_config (RetryConfig, optional): The retry configuration for the embedding model.
90
+ Defaults to a default config with no retry and 30.0 seconds timeout.
90
91
  """
@@ -1,5 +1,6 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.utils.retry import RetryConfig as RetryConfig
3
+ from gllm_inference.constants import INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
3
4
  from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
4
5
  from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider, Vector as Vector
5
6
  from typing import Any
@@ -68,7 +69,7 @@ class OpenAIEMInvoker(BaseEMInvoker):
68
69
  ```
69
70
  '''
70
71
  client: Incomplete
71
- def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None) -> None:
72
+ def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig = ...) -> None:
72
73
  """Initializes a new instance of the OpenAIEMInvoker class.
73
74
 
74
75
  Args:
@@ -79,6 +80,6 @@ class OpenAIEMInvoker(BaseEMInvoker):
79
80
  Defaults to None.
80
81
  default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
81
82
  Defaults to None.
82
- retry_config (RetryConfig | None, optional): The retry configuration for the embedding model.
83
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
83
+ retry_config (RetryConfig, optional): The retry configuration for the embedding model.
84
+ Defaults to a default config with no retry and 30.0 seconds timeout.
84
85
  """
@@ -1,5 +1,6 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.utils.retry import RetryConfig as RetryConfig
3
+ from gllm_inference.constants import INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
3
4
  from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
4
5
  from gllm_inference.em_invoker.schema.twelvelabs import InputType as InputType, Key as Key, OutputType as OutputType
5
6
  from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EMContent as EMContent, ModelId as ModelId, ModelProvider as ModelProvider, Vector as Vector
@@ -82,7 +83,7 @@ class TwelveLabsEMInvoker(BaseEMInvoker):
82
83
  ```
83
84
  '''
84
85
  client: Incomplete
85
- def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None) -> None:
86
+ def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig = ...) -> None:
86
87
  """Initializes a new instance of the TwelveLabsEMInvoker class.
87
88
 
88
89
  Args:
@@ -93,6 +94,6 @@ class TwelveLabsEMInvoker(BaseEMInvoker):
93
94
  Defaults to None.
94
95
  default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
95
96
  Defaults to None.
96
- retry_config (RetryConfig | None, optional): The retry configuration for the embedding model.
97
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
97
+ retry_config (RetryConfig, optional): The retry configuration for the embedding model.
98
+ Defaults to a default config with no retry and 30.0 seconds timeout.
98
99
  """
@@ -1,5 +1,6 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.utils.retry import RetryConfig as RetryConfig
3
+ from gllm_inference.constants import INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
3
4
  from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
4
5
  from gllm_inference.em_invoker.schema.voyage import InputType as InputType, Key as Key
5
6
  from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EMContent as EMContent, ModelId as ModelId, ModelProvider as ModelProvider, Vector as Vector
@@ -84,7 +85,7 @@ class VoyageEMInvoker(BaseEMInvoker):
84
85
  ```
85
86
  '''
86
87
  client: Incomplete
87
- def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None) -> None:
88
+ def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig = ...) -> None:
88
89
  """Initializes a new instance of the VoyageEMInvoker class.
89
90
 
90
91
  Args:
@@ -95,6 +96,6 @@ class VoyageEMInvoker(BaseEMInvoker):
95
96
  Defaults to None.
96
97
  default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
97
98
  Defaults to None.
98
- retry_config (RetryConfig | None, optional): The retry configuration for the embedding model.
99
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
99
+ retry_config (RetryConfig, optional): The retry configuration for the embedding model.
100
+ Defaults to a default config with no retry and 30.0 seconds timeout.
100
101
  """
@@ -1,6 +1,7 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.event import EventEmitter as EventEmitter
3
3
  from gllm_core.utils.retry import RetryConfig as RetryConfig
4
+ from gllm_inference.constants import INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
4
5
  from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
5
6
  from gllm_inference.lm_invoker.schema.anthropic import InputType as InputType, Key as Key, OutputType as OutputType
6
7
  from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EmitDataType as EmitDataType, LMOutput as LMOutput, Message as Message, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
@@ -221,7 +222,7 @@ class AnthropicLMInvoker(BaseLMInvoker):
221
222
  client: Incomplete
222
223
  thinking: Incomplete
223
224
  thinking_budget: Incomplete
224
- def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, thinking: bool = False, thinking_budget: int = ...) -> None:
225
+ def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig = ..., thinking: bool = False, thinking_budget: int = ...) -> None:
225
226
  """Initializes the AnthropicLmInvoker instance.
226
227
 
227
228
  Args:
@@ -236,8 +237,8 @@ class AnthropicLMInvoker(BaseLMInvoker):
236
237
  output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
237
238
  dictionary. Defaults to None.
238
239
  output_analytics (bool, optional): Whether to output the invocation analytics. Defaults to False.
239
- retry_config (RetryConfig | None, optional): The retry configuration for the language model.
240
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
240
+ retry_config (RetryConfig, optional): The retry configuration for the language model.
241
+ Defaults to a default config with no retry and 30.0 seconds timeout.
241
242
  thinking (bool, optional): Whether to enable thinking. Only allowed for thinking models. Defaults to False.
242
243
  thinking_budget (int, optional): The tokens allocated for the thinking process. Must be greater than or
243
244
  equal to 1024. Only allowed for thinking models. Defaults to DEFAULT_THINKING_BUDGET.
@@ -1,6 +1,6 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.utils.retry import RetryConfig as RetryConfig
3
- from gllm_inference.constants import DEFAULT_AZURE_OPENAI_API_VERSION as DEFAULT_AZURE_OPENAI_API_VERSION
3
+ from gllm_inference.constants import DEFAULT_AZURE_OPENAI_API_VERSION as DEFAULT_AZURE_OPENAI_API_VERSION, INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
4
4
  from gllm_inference.lm_invoker.openai_lm_invoker import OpenAILMInvoker as OpenAILMInvoker, ReasoningEffort as ReasoningEffort, ReasoningSummary as ReasoningSummary
5
5
  from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema
6
6
  from langchain_core.tools import Tool as Tool
@@ -218,7 +218,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
218
218
  Defaults to an empty list.
219
219
  '''
220
220
  client: Incomplete
221
- def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None) -> None:
221
+ def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig = ..., reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None) -> None:
222
222
  """Initializes a new instance of the AzureOpenAILMInvoker class.
223
223
 
224
224
  Args:
@@ -237,8 +237,8 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
237
237
  output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
238
238
  dictionary. Defaults to None.
239
239
  output_analytics (bool, optional): Whether to output the invocation analytics. Defaults to False.
240
- retry_config (RetryConfig | None, optional): The retry configuration for the language model.
241
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
240
+ retry_config (RetryConfig, optional): The retry configuration for the language model.
241
+ Defaults to a default config with no retry and 30.0 seconds timeout.
242
242
  reasoning_effort (ReasoningEffort | None, optional): The reasoning effort for reasoning models. Not allowed
243
243
  for non-reasoning models. If None, the model will perform medium reasoning effort. Defaults to None.
244
244
  reasoning_summary (ReasoningSummary | None, optional): The reasoning summary level for reasoning models.
@@ -1,6 +1,7 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.event import EventEmitter as EventEmitter
3
3
  from gllm_core.utils.retry import RetryConfig as RetryConfig
4
+ from gllm_inference.constants import INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
4
5
  from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
5
6
  from gllm_inference.lm_invoker.schema.bedrock import InputType as InputType, Key as Key, OutputType as OutputType
6
7
  from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, LMOutput as LMOutput, Message as Message, ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
@@ -179,7 +180,7 @@ class BedrockLMInvoker(BaseLMInvoker):
179
180
  '''
180
181
  session: Incomplete
181
182
  client_kwargs: Incomplete
182
- def __init__(self, model_name: str, access_key_id: str | None = None, secret_access_key: str | None = None, region_name: str = 'us-east-1', model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None) -> None:
183
+ def __init__(self, model_name: str, access_key_id: str | None = None, secret_access_key: str | None = None, region_name: str = 'us-east-1', model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig = ...) -> None:
183
184
  '''Initializes the BedrockLMInvoker instance.
184
185
 
185
186
  Args:
@@ -197,8 +198,8 @@ class BedrockLMInvoker(BaseLMInvoker):
197
198
  output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
198
199
  dictionary. Defaults to None.
199
200
  output_analytics (bool, optional): Whether to output the invocation analytics. Defaults to False.
200
- retry_config (RetryConfig | None, optional): The retry configuration for the language model.
201
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
201
+ retry_config (RetryConfig, optional): The retry configuration for the language model.
202
+ Defaults to a default config with no retry and 30.0 seconds timeout.
202
203
 
203
204
  Raises:
204
205
  ValueError: If `response_schema` is provided, but `tools` are also provided.
@@ -1,7 +1,7 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.event import EventEmitter as EventEmitter
3
3
  from gllm_core.utils.retry import RetryConfig as RetryConfig
4
- from gllm_inference.constants import DOCUMENT_MIME_TYPES as DOCUMENT_MIME_TYPES
4
+ from gllm_inference.constants import DOCUMENT_MIME_TYPES as DOCUMENT_MIME_TYPES, INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
5
5
  from gllm_inference.lm_invoker.openai_compatible_lm_invoker import OpenAICompatibleLMInvoker as OpenAICompatibleLMInvoker
6
6
  from gllm_inference.lm_invoker.schema.datasaur import InputType as InputType, Key as Key
7
7
  from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, LMOutput as LMOutput, Message as Message, ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema, ToolCall as ToolCall, ToolResult as ToolResult
@@ -121,7 +121,7 @@ class DatasaurLMInvoker(OpenAICompatibleLMInvoker):
121
121
  '''
122
122
  client: Incomplete
123
123
  citations: Incomplete
124
- def __init__(self, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, citations: bool = False) -> None:
124
+ def __init__(self, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, output_analytics: bool = False, retry_config: RetryConfig = ..., citations: bool = False) -> None:
125
125
  """Initializes a new instance of the DatasaurLMInvoker class.
126
126
 
127
127
  Args:
@@ -132,8 +132,8 @@ class DatasaurLMInvoker(OpenAICompatibleLMInvoker):
132
132
  default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
133
133
  Defaults to None.
134
134
  output_analytics (bool, optional): Whether to output the invocation analytics. Defaults to False.
135
- retry_config (RetryConfig | None, optional): The retry configuration for the language model.
136
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
135
+ retry_config (RetryConfig, optional): The retry configuration for the language model.
136
+ Defaults to a default config with no retry and 30.0 seconds timeout.
137
137
  citations (bool, optional): Whether to output the citations. Defaults to False.
138
138
 
139
139
  Raises:
@@ -1,7 +1,7 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.event import EventEmitter as EventEmitter
3
3
  from gllm_core.utils.retry import RetryConfig as RetryConfig
4
- from gllm_inference.constants import GOOGLE_SCOPES as GOOGLE_SCOPES
4
+ from gllm_inference.constants import GOOGLE_SCOPES as GOOGLE_SCOPES, INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
5
5
  from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
6
6
  from gllm_inference.lm_invoker.schema.google import InputType as InputType, Key as Key
7
7
  from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EmitDataType as EmitDataType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
@@ -250,7 +250,7 @@ class GoogleLMInvoker(BaseLMInvoker):
250
250
  client_params: Incomplete
251
251
  thinking: Incomplete
252
252
  thinking_budget: Incomplete
253
- def __init__(self, model_name: str, api_key: str | None = None, credentials_path: str | None = None, project_id: str | None = None, location: str = 'us-central1', model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, thinking: bool | None = None, thinking_budget: int = ...) -> None:
253
+ def __init__(self, model_name: str, api_key: str | None = None, credentials_path: str | None = None, project_id: str | None = None, location: str = 'us-central1', model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig = ..., thinking: bool | None = None, thinking_budget: int = ...) -> None:
254
254
  '''Initializes a new instance of the GoogleLMInvoker class.
255
255
 
256
256
  Args:
@@ -273,8 +273,8 @@ class GoogleLMInvoker(BaseLMInvoker):
273
273
  output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
274
274
  dictionary. Defaults to None.
275
275
  output_analytics (bool, optional): Whether to output the invocation analytics. Defaults to False.
276
- retry_config (RetryConfig | None, optional): The retry configuration for the language model.
277
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
276
+ retry_config (RetryConfig, optional): The retry configuration for the language model.
277
+ Defaults to a default config with no retry and 30.0 seconds timeout.
278
278
  thinking (bool | None, optional): Whether to enable thinking. Only allowed for thinking models.
279
279
  Defaults to True for Gemini 2.5 Pro models and False for other models.
280
280
  thinking_budget (int, optional): The tokens allowed for thinking process. Only allowed for thinking models.
@@ -1,6 +1,7 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.event import EventEmitter as EventEmitter
3
3
  from gllm_core.utils.retry import RetryConfig as RetryConfig
4
+ from gllm_inference.constants import INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
4
5
  from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
5
6
  from gllm_inference.lm_invoker.schema.langchain import InputType as InputType, Key as Key
6
7
  from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
@@ -203,7 +204,7 @@ class LangChainLMInvoker(BaseLMInvoker):
203
204
  Defaults to an empty list.
204
205
  '''
205
206
  model: Incomplete
206
- def __init__(self, model: BaseChatModel | None = None, model_class_path: str | None = None, model_name: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None) -> None:
207
+ def __init__(self, model: BaseChatModel | None = None, model_class_path: str | None = None, model_name: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig = ...) -> None:
207
208
  '''Initializes a new instance of the LangChainLMInvoker class.
208
209
 
209
210
  Args:
@@ -223,8 +224,8 @@ class LangChainLMInvoker(BaseLMInvoker):
223
224
  output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
224
225
  dictionary. Defaults to None.
225
226
  output_analytics (bool, optional): Whether to output the invocation analytics. Defaults to False.
226
- retry_config (RetryConfig | None, optional): The retry configuration for the language model.
227
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
227
+ retry_config (RetryConfig, optional): The retry configuration for the language model.
228
+ Defaults to a default config with no retry and 30.0 seconds timeout.
228
229
 
229
230
  Raises:
230
231
  ValueError: If `response_schema` is provided, but `tools` are also provided.
@@ -1,6 +1,7 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.event import EventEmitter as EventEmitter
3
3
  from gllm_core.utils.retry import RetryConfig as RetryConfig
4
+ from gllm_inference.constants import INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
4
5
  from gllm_inference.lm_invoker.openai_compatible_lm_invoker import OpenAICompatibleLMInvoker as OpenAICompatibleLMInvoker
5
6
  from gllm_inference.lm_invoker.openai_lm_invoker import ReasoningEffort as ReasoningEffort
6
7
  from gllm_inference.schema import AttachmentType as AttachmentType, LMOutput as LMOutput, ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema
@@ -227,7 +228,7 @@ class LiteLLMLMInvoker(OpenAICompatibleLMInvoker):
227
228
  Defaults to an empty list.
228
229
  '''
229
230
  completion: Incomplete
230
- def __init__(self, model_id: str, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None) -> None:
231
+ def __init__(self, model_id: str, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig = ..., reasoning_effort: ReasoningEffort | None = None) -> None:
231
232
  """Initializes a new instance of the LiteLLMLMInvoker class.
232
233
 
233
234
  Args:
@@ -239,8 +240,8 @@ class LiteLLMLMInvoker(OpenAICompatibleLMInvoker):
239
240
  output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
240
241
  dictionary. Defaults to None.
241
242
  output_analytics (bool, optional): Whether to output the invocation analytics. Defaults to False.
242
- retry_config (RetryConfig | None, optional): The retry configuration for the language model.
243
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
243
+ retry_config (RetryConfig, optional): The retry configuration for the language model.
244
+ Defaults to a default config with no retry and 30.0 seconds timeout.
244
245
  reasoning_effort (ReasoningEffort | None, optional): The reasoning effort for reasoning models.
245
246
  Defaults to None.
246
247
  """
@@ -3,7 +3,7 @@ from _typeshed import Incomplete
3
3
  from abc import ABC
4
4
  from gllm_core.event import EventEmitter as EventEmitter
5
5
  from gllm_core.utils.retry import RetryConfig
6
- from gllm_inference.constants import DOCUMENT_MIME_TYPES as DOCUMENT_MIME_TYPES
6
+ from gllm_inference.constants import DOCUMENT_MIME_TYPES as DOCUMENT_MIME_TYPES, INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
7
7
  from gllm_inference.exceptions import parse_error_message as parse_error_message
8
8
  from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EmitDataType as EmitDataType, LMOutput as LMOutput, Message as Message, MessageContent as MessageContent, MessageRole as MessageRole, ModelId as ModelId, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ToolCall as ToolCall, ToolResult as ToolResult
9
9
  from langchain_core.tools import Tool as Tool
@@ -46,7 +46,7 @@ class BaseLMInvoker(ABC, metaclass=abc.ABCMeta):
46
46
  response_schema: Incomplete
47
47
  output_analytics: Incomplete
48
48
  retry_config: Incomplete
49
- def __init__(self, model_id: ModelId, default_hyperparameters: dict[str, Any] | None = None, supported_attachments: set[str] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None) -> None:
49
+ def __init__(self, model_id: ModelId, default_hyperparameters: dict[str, Any] | None = None, supported_attachments: set[str] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig = ...) -> None:
50
50
  """Initializes a new instance of the BaseLMInvoker class.
51
51
 
52
52
  Args:
@@ -61,8 +61,8 @@ class BaseLMInvoker(ABC, metaclass=abc.ABCMeta):
61
61
  output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
62
62
  dictionary. Defaults to None.
63
63
  output_analytics (bool, optional): Whether to output the invocation analytics. Defaults to False.
64
- retry_config (RetryConfig | None, optional): The retry configuration for the language model.
65
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
64
+ retry_config (RetryConfig, optional): The retry configuration for the language model.
65
+ Defaults to a default config with no retry and 30.0 seconds timeout.
66
66
  """
67
67
  @property
68
68
  def model_id(self) -> str:
@@ -1,6 +1,7 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.event import EventEmitter as EventEmitter
3
3
  from gllm_core.utils.retry import RetryConfig as RetryConfig
4
+ from gllm_inference.constants import INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
4
5
  from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
5
6
  from gllm_inference.lm_invoker.schema.openai_compatible import InputType as InputType, Key as Key, ReasoningEffort as ReasoningEffort
6
7
  from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EmitDataType as EmitDataType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
@@ -230,14 +231,15 @@ class OpenAICompatibleLMInvoker(BaseLMInvoker):
230
231
  Defaults to an empty list.
231
232
  '''
232
233
  client: Incomplete
233
- def __init__(self, model_name: str, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None) -> None:
234
- """Initializes a new instance of the OpenAICompatibleLMInvoker class.
234
+ def __init__(self, model_name: str, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig = ..., reasoning_effort: ReasoningEffort | None = None) -> None:
235
+ '''Initializes a new instance of the OpenAICompatibleLMInvoker class.
235
236
 
236
237
  Args:
237
238
  model_name (str): The name of the language model hosted on the OpenAI compatible endpoint.
238
239
  base_url (str): The base URL for the OpenAI compatible endpoint.
239
240
  api_key (str | None, optional): The API key for authenticating with the OpenAI compatible endpoint.
240
241
  Defaults to None, in which case the `OPENAI_API_KEY` environment variable will be used.
242
+ If the endpoint does not require an API key, a dummy value can be passed (e.g. "<empty>").
241
243
  model_kwargs (dict[str, Any] | None, optional): Additional model parameters. Defaults to None.
242
244
  default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
243
245
  Defaults to None.
@@ -247,10 +249,10 @@ class OpenAICompatibleLMInvoker(BaseLMInvoker):
247
249
  output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
248
250
  dictionary. Defaults to None.
249
251
  output_analytics (bool, optional): Whether to output the invocation analytics. Defaults to False.
250
- retry_config (RetryConfig | None, optional): The retry configuration for the language model.
251
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
252
+ retry_config (RetryConfig, optional): The retry configuration for the language model.
253
+ Defaults to a default config with no retry and 30.0 seconds timeout.
252
254
  reasoning_effort (str | None, optional): The reasoning effort for the language model. Defaults to None.
253
- """
255
+ '''
254
256
  def set_response_schema(self, response_schema: ResponseSchema | None) -> None:
255
257
  """Sets the response schema for the language model hosted on the OpenAI compatible endpoint.
256
258
 
@@ -1,6 +1,7 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.event import EventEmitter as EventEmitter
3
3
  from gllm_core.utils.retry import RetryConfig as RetryConfig
4
+ from gllm_inference.constants import INVOKER_DEFAULT_RETRY_CONFIG as INVOKER_DEFAULT_RETRY_CONFIG
4
5
  from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
5
6
  from gllm_inference.lm_invoker.schema.openai import InputType as InputType, Key as Key, OutputType as OutputType, ReasoningEffort as ReasoningEffort, ReasoningSummary as ReasoningSummary
6
7
  from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, CodeExecResult as CodeExecResult, EmitDataType as EmitDataType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
@@ -318,7 +319,7 @@ class OpenAILMInvoker(BaseLMInvoker):
318
319
  enabled and the language model decides to execute any codes. Defaults to an empty list.
319
320
  '''
320
321
  client: Incomplete
321
- def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None, code_interpreter: bool = False, web_search: bool = False) -> None:
322
+ def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig = ..., reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None, code_interpreter: bool = False, web_search: bool = False) -> None:
322
323
  """Initializes a new instance of the OpenAILMInvoker class.
323
324
 
324
325
  Args:
@@ -334,8 +335,8 @@ class OpenAILMInvoker(BaseLMInvoker):
334
335
  output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
335
336
  dictionary. Defaults to None.
336
337
  output_analytics (bool, optional): Whether to output the invocation analytics. Defaults to False.
337
- retry_config (RetryConfig | None, optional): The retry configuration for the language model.
338
- Defaults to None, in which case a default config with no retry and 30.0 seconds timeout is used.
338
+ retry_config (RetryConfig, optional): The retry configuration for the language model.
339
+ Defaults to a default config with no retry and 30.0 seconds timeout.
339
340
  reasoning_effort (ReasoningEffort | None, optional): The reasoning effort for reasoning models. Not allowed
340
341
  for non-reasoning models. If None, the model will perform medium reasoning effort. Defaults to None.
341
342
  reasoning_summary (ReasoningSummary | None, optional): The reasoning summary level for reasoning models.
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: gllm-inference-binary
3
- Version: 0.5.1.post1
3
+ Version: 0.5.2
4
4
  Summary: A library containing components related to model inferences in Gen AI applications.
5
5
  Author: Henry Wicaksono
6
6
  Author-email: henry.wicaksono@gdplabs.id
@@ -8,35 +8,35 @@ gllm_inference/catalog/__init__.pyi,sha256=JBkPGTyiiZ30GECzJBW-mW8LekWyY2qyzal3e
8
8
  gllm_inference/catalog/catalog.pyi,sha256=a4RNG1lKv51GxQpOqh47tz-PAROMPaeP2o5XNLBSZaU,4790
9
9
  gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=wjzufPEqey-byBU3hPWwEawT9c182WwjzSWOJ2bnqIs,4599
10
10
  gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=o4JSzISjlStBataofZ2MB2_t3wnGTkkFJ3Dm_NSm5qo,3159
11
- gllm_inference/constants.pyi,sha256=RKZ5ILmObzNbIie90ggB8L1JxuuLL-mzvCN9srWkh_0,189
11
+ gllm_inference/constants.pyi,sha256=mK_YuuN1MU0Kk4aD-4kDhRZnwPiPFy2l34Mg2ijoir8,230
12
12
  gllm_inference/em_invoker/__init__.pyi,sha256=XESsrYo1PZeeHe7AMRyuzKoV7XDD5oN89ZTH01zRf4k,873
13
- gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=bBhwKg44mjWd1LW_xGB7Qafs1SOOiJcz3-lw0IQCBYg,4468
14
- gllm_inference/em_invoker/em_invoker.pyi,sha256=IIi3q374cv0uCRRSMjbUt6KtHB9TJ4NRfwE8pZoKL0g,4284
15
- gllm_inference/em_invoker/google_em_invoker.pyi,sha256=hiVFqCnmK7yRprbCZGkgbVuoP8F6mMhF5xUm3wRyCek,6062
13
+ gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=TDE5g8PM0dJUTH9yBQ8uIM2Wd9NKKwIoVkbRwJhV-HE,4487
14
+ gllm_inference/em_invoker/em_invoker.pyi,sha256=ffEwrFpdfV6vlwjuop4iPQm_dgcojLIMHXAhHacRlWM,4318
15
+ gllm_inference/em_invoker/google_em_invoker.pyi,sha256=dPuc68qk1hdjswCCAoPNJ6ZoeW3oX1VYRW2LhWAp8IY,6081
16
16
  gllm_inference/em_invoker/langchain/__init__.pyi,sha256=VYGKE5OgU0my1RlhgzkU_A7-GLGnUDDnNFuctuRwILE,148
17
17
  gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi,sha256=6nASLqi0FXCpqyYPl7kM3g7hAW-xS5ZwsS3GFudns98,2347
18
- gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=LrwbDtXjA8HuZk7kdyKhFH1MpCzF1Gl50-rllcTs_lE,2572
19
- gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=Ef91Kk1bSzJsdEYlfs7dFq74UJ8JdybC89BSdg2ZRLY,4797
20
- gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=nUjPUf5DWd8WIdObCtw1Q00q7dTWVfjwod5TynGxtMA,4065
18
+ gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=bRrywoGXIyZYIye85WhSbt0Pz5Qft1FWcp8hog2ikfk,2627
19
+ gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=w1vhe5PCHgOs16qKhdomSPnD513WbIpbtzv_4yEVJTw,4852
20
+ gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=JXpVxNY26jrx3sraI1bGUQqut4zoRNd_ZmKzsT8VxQA,4120
21
21
  gllm_inference/em_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
22
22
  gllm_inference/em_invoker/schema/twelvelabs.pyi,sha256=y9nfqvuzs0o1jlZ5bwNmJkMSmvHAwBElWNMU0wJNc64,334
23
23
  gllm_inference/em_invoker/schema/voyage.pyi,sha256=7d9hv8uCJ4NboRaqK77g5ekb0X78pPW__sz2EYeKJKw,219
24
- gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=a4-0vUVo8kOeBzrEfJ_0r5vnyT9DWaTYtEp386Bp2S4,4959
25
- gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=0xI7MCV080LD9zIJkt0TVedzCT_xtdP-8uDSX7vaHmA,5027
24
+ gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=MtC-JC0QHxs00WQyU4M0buft3JalcuW68ZzE9JuEBR0,5014
25
+ gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=tW79qfg-O_8HuH1OQUzIuWTQ70l9rJQE34mDQzSpDcg,5082
26
26
  gllm_inference/exceptions/__init__.pyi,sha256=v9uxjW5DssIn7n_bKqT7L83CeqFET2Z45GFOvi78UuE,977
27
27
  gllm_inference/exceptions/error_parser.pyi,sha256=4RkVfS2Fl9kjz_h2bK9eoAeI-Y-VkHcUqXWj68BsYig,2393
28
28
  gllm_inference/exceptions/exceptions.pyi,sha256=5YRackwVNvyOJjOtiVszqu8q87s8ioXTa-XwaYmeiC4,4643
29
29
  gllm_inference/lm_invoker/__init__.pyi,sha256=8oUFgavpItvjRUqsLqLSHqQVIPyTKCgNQ_euf58H2zY,1104
30
- gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=1FxJ8AM03rsIHyw6hUe71rXCTtG4wj2d9QGwX0pnR5s,14658
31
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=7gfoX-cZOnxvemA8t8jwqNIptTw9z3NEhoAisDByOtU,14402
32
- gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=yFY9WAGN1vrBSwM_Es-SGmukSDxP9tbTKtWdUJMU8Ko,12295
33
- gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=VXnqot4ijggXHxW9dNBFQqQnw-bXtMRUnDi66sFJX4c,9044
34
- gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=pTInkRYZooYcUQMqwkvqv5y6kV-X8lSkNm1mwggULc4,16388
35
- gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=E4KeYVJucg0R3J26Z48I-1uCzndpdOu7W7S6wNvML1Q,13013
36
- gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=SkHELX0z_XNXYejQKdJMRjleaJpprJ6ZlPmv6WBSy3U,13032
37
- gllm_inference/lm_invoker/lm_invoker.pyi,sha256=81eGTt6dV7K06Oc3dK8jDDjrxpi832CTp3OFkLBx7VQ,7553
38
- gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=B1OYYEDGfh1BWl7cZ-BPBYQRF7Aqa5ZCU29fsjwgtZw,14647
39
- gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=PvyvdCL-br4AusbnNxdntxHjgVwIhOxB-T8kwivGMAs,19320
30
+ gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=J_Y63fX1oQlDFcruD7QZuADovDj21lv-6uiWlRTaoto,14713
31
+ gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=AUgRctJrMfh9QxnAygDplajcd8F7zavz44Qq6QurHi0,14421
32
+ gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=YZYK_zLRlmx97M14h0C8P1icuvWi718tgt-s0chXnaw,12350
33
+ gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=a3jbg8cAxx7hnV1iRYX13LkH8zU4HImnUjsQhYHHMWE,9063
34
+ gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=-0cKbJaJ1egEnP5KcGemW8SAQUTeeYqv1tHDpbCQfc0,16407
35
+ gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=35vw9SDgE1ucgG3ucePPALQpKhdWYP1jQ1T6_mIfv08,13068
36
+ gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=RrCBEzxb3aQVYf3iQnj1POna4dT-3-042la0iKOUSYA,13087
37
+ gllm_inference/lm_invoker/lm_invoker.pyi,sha256=Vj-FkS4wcAHfAdGTVSwIeNvW9j-SvUnRTXx7bhT1a7c,7572
38
+ gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=fllMyWvAebhua2j1MRU3-ie4B8TVUJZ3pPFKtYE5R4k,14809
39
+ gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=E2gIVn4k_6WvI-7VWvc6BkKItCY7A7HpQhIo2XPQMKw,19375
40
40
  gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
41
41
  gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=Y7NAjB7H0Wmiwv3YGZuG83K1rOCjSDB6hoZ6LjndWss,971
42
42
  gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=FOUMZkBi6KRa__mYoy3FNJ5sP0EC6rgLuhoijnwelIg,927
@@ -86,8 +86,8 @@ gllm_inference/utils/__init__.pyi,sha256=npmBmmlBv7cPHMg1hdL3S2_RelD6vk_LhCsGELh
86
86
  gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
87
87
  gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
88
88
  gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
89
- gllm_inference.cpython-312-x86_64-linux-gnu.so,sha256=UnkTDp5BdugW2MKfK6sZVzUcKvvwM27oixcbYWRb7Y0,3908072
89
+ gllm_inference.cpython-312-x86_64-linux-gnu.so,sha256=gG5xsEvzds899Qe-ZzzLbpgJVVbs98RU9UnsdwhHEgk,3924552
90
90
  gllm_inference.pyi,sha256=VEcscbPCJ-6lXU4jV3YYXwwumk9kWxpCAsS84ssKG6o,3295
91
- gllm_inference_binary-0.5.1.post1.dist-info/METADATA,sha256=8Me6oGk1deXkobBF1mrU6dBO6_jG6ejVl8H_qhdh23o,4537
92
- gllm_inference_binary-0.5.1.post1.dist-info/WHEEL,sha256=mNY4pwQL4AOAoPmLYEQs2SSpMIbATFeiJFktRD5iKkY,110
93
- gllm_inference_binary-0.5.1.post1.dist-info/RECORD,,
91
+ gllm_inference_binary-0.5.2.dist-info/METADATA,sha256=g7UfIlC7oMewzWaJ8cI7SKlNYdEKL11XARSYJGpyEbM,4531
92
+ gllm_inference_binary-0.5.2.dist-info/WHEEL,sha256=mNY4pwQL4AOAoPmLYEQs2SSpMIbATFeiJFktRD5iKkY,110
93
+ gllm_inference_binary-0.5.2.dist-info/RECORD,,