gllm-inference-binary 0.5.37__cp311-cp311-manylinux_2_31_x86_64.whl → 0.5.38__cp311-cp311-manylinux_2_31_x86_64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of gllm-inference-binary might be problematic. Click here for more details.

@@ -13,7 +13,7 @@ class AzureOpenAIEMInvoker(OpenAIEMInvoker):
13
13
  model_id (str): The model ID of the embedding model.
14
14
  model_provider (str): The provider of the embedding model.
15
15
  model_name (str): The name of the Azure OpenAI embedding model deployment.
16
- client (AsyncAzureOpenAI): The client for the Azure OpenAI API.
16
+ client_kwargs (dict[str, Any]): The keyword arguments for the Azure OpenAI client.
17
17
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the embedding model.
18
18
  retry_config (RetryConfig): The retry configuration for the embedding model.
19
19
  truncation_config (TruncationConfig | None): The truncation configuration for the embedding model.
@@ -68,7 +68,7 @@ class AzureOpenAIEMInvoker(OpenAIEMInvoker):
68
68
  em_invoker = AzureOpenAIEMInvoker(..., retry_config=retry_config)
69
69
  ```
70
70
  '''
71
- client: Incomplete
71
+ client_kwargs: Incomplete
72
72
  def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
73
73
  """Initializes a new instance of the AzureOpenAIEMInvoker class.
74
74
 
@@ -15,14 +15,14 @@ class OpenAICompatibleEMInvoker(OpenAIEMInvoker):
15
15
  model_id (str): The model ID of the embedding model.
16
16
  model_provider (str): The provider of the embedding model.
17
17
  model_name (str): The name of the embedding model.
18
- client (AsyncOpenAI): The OpenAI client instance.
18
+ client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
19
19
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the embedding model.
20
20
  retry_config (RetryConfig): The retry configuration for the embedding model.
21
21
  truncation_config (TruncationConfig | None): The truncation configuration for the embedding model.
22
22
 
23
23
  This class is deprecated and will be removed in v0.6. Please use the `OpenAIEMInvoker` class instead.
24
24
  """
25
- client: Incomplete
25
+ client_kwargs: Incomplete
26
26
  def __init__(self, model_name: str, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
27
27
  """Initializes a new instance of the OpenAICompatibleEMInvoker class.
28
28
 
@@ -16,7 +16,7 @@ class OpenAIEMInvoker(BaseEMInvoker):
16
16
  model_id (str): The model ID of the embedding model.
17
17
  model_provider (str): The provider of the embedding model.
18
18
  model_name (str): The name of the embedding model.
19
- client (AsyncOpenAI): The client for the OpenAI API.
19
+ client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
20
20
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the embedding model.
21
21
  retry_config (RetryConfig): The retry configuration for the embedding model.
22
22
  truncation_config (TruncationConfig | None): The truncation configuration for the embedding model.
@@ -96,7 +96,7 @@ class OpenAIEMInvoker(BaseEMInvoker):
96
96
  em_invoker = OpenAIEMInvoker(..., retry_config=retry_config)
97
97
  ```
98
98
  '''
99
- client: Incomplete
99
+ client_kwargs: Incomplete
100
100
  def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
101
101
  '''Initializes a new instance of the OpenAIEMInvoker class.
102
102
 
@@ -15,7 +15,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
15
15
  model_id (str): The model ID of the language model.
16
16
  model_provider (str): The provider of the language model.
17
17
  model_name (str): The name of the Azure OpenAI language model deployment.
18
- client (AsyncAzureOpenAI): The Azure OpenAI client instance.
18
+ client_kwargs (dict[str, Any]): The keyword arguments for the Azure OpenAI client.
19
19
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
20
20
  tools (list[Tool]): The list of tools provided to the model to enable tool calling.
21
21
  response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
@@ -226,7 +226,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
226
226
  Defaults to an empty list.
227
227
  2.10. mcp_calls (list[MCPCall]): The MCP calls. Currently not supported. Defaults to an empty list.
228
228
  '''
229
- client: Incomplete
229
+ client_kwargs: Incomplete
230
230
  def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None) -> None:
231
231
  """Initializes a new instance of the AzureOpenAILMInvoker class.
232
232
 
@@ -18,7 +18,7 @@ class DatasaurLMInvoker(OpenAICompatibleLMInvoker):
18
18
  model_id (str): The model ID of the language model.
19
19
  model_provider (str): The provider of the language model.
20
20
  model_name (str): The name of the language model.
21
- client (AsyncOpenAI): The OpenAI client instance.
21
+ client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
22
22
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
23
23
  tools (list[Any]): The list of tools provided to the model to enable tool calling. Currently not supported.
24
24
  response_schema (ResponseSchema | None): The schema of the response. Currently not supported.
@@ -121,7 +121,7 @@ class DatasaurLMInvoker(OpenAICompatibleLMInvoker):
121
121
  Defaults to an empty list.
122
122
  2.10. mcp_calls (list[MCPCall]): The MCP calls. Currently not supported. Defaults to an empty list.
123
123
  '''
124
- client: Incomplete
124
+ client_kwargs: Incomplete
125
125
  citations: Incomplete
126
126
  def __init__(self, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, citations: bool = False) -> None:
127
127
  """Initializes a new instance of the DatasaurLMInvoker class.
@@ -23,7 +23,7 @@ class OpenAIChatCompletionsLMInvoker(BaseLMInvoker):
23
23
  model_id (str): The model ID of the language model.
24
24
  model_provider (str): The provider of the language model.
25
25
  model_name (str): The name of the language model.
26
- client (AsyncOpenAI): The OpenAI client instance.
26
+ client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
27
27
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
28
28
  tools (list[Tool]): The list of tools provided to the model to enable tool calling.
29
29
  response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
@@ -243,7 +243,7 @@ class OpenAIChatCompletionsLMInvoker(BaseLMInvoker):
243
243
  Defaults to an empty list.
244
244
  2.10. mcp_calls (list[MCPCall]): The MCP calls. Currently not supported. Defaults to an empty list.
245
245
  '''
246
- client: Incomplete
246
+ client_kwargs: Incomplete
247
247
  def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None) -> None:
248
248
  '''Initializes a new instance of the OpenAIChatCompletionsLMInvoker class.
249
249
 
@@ -15,7 +15,7 @@ class OpenAICompatibleLMInvoker(OpenAIChatCompletionsLMInvoker):
15
15
  model_id (str): The model ID of the language model.
16
16
  model_provider (str): The provider of the language model.
17
17
  model_name (str): The name of the language model.
18
- client (AsyncOpenAI): The OpenAI client instance.
18
+ client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
19
19
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
20
20
  tools (list[Tool]): The list of tools provided to the model to enable tool calling.
21
21
  response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
@@ -25,7 +25,7 @@ class OpenAILMInvoker(BaseLMInvoker):
25
25
  model_id (str): The model ID of the language model.
26
26
  model_provider (str): The provider of the language model.
27
27
  model_name (str): The name of the language model.
28
- client (AsyncOpenAI): The OpenAI client instance.
28
+ client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
29
29
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
30
30
  tools (list[Tool]): The list of tools provided to the model to enable tool calling.
31
31
  response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
@@ -391,7 +391,7 @@ class OpenAILMInvoker(BaseLMInvoker):
391
391
  2.10. mcp_calls (list[MCPCall]): The MCP calls, if the MCP servers are provided and the language model
392
392
  decides to invoke MCP tools. Defaults to an empty list.
393
393
  '''
394
- client: Incomplete
394
+ client_kwargs: Incomplete
395
395
  def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None, mcp_servers: list[MCPServer] | None = None, code_interpreter: bool = False, web_search: bool = False) -> None:
396
396
  '''Initializes a new instance of the OpenAILMInvoker class.
397
397
 
@@ -1,5 +1,7 @@
1
1
  class Key:
2
2
  """Defines valid keys in Datasaur."""
3
+ API_KEY: str
4
+ BASE_URL: str
3
5
  CONTEXTS: str
4
6
  MAX_RETRIES: str
5
7
  NAME: str
@@ -3,8 +3,10 @@ from enum import StrEnum
3
3
  class Key:
4
4
  """Defines valid keys in OpenAI."""
5
5
  ALLOWED_TOOLS: str
6
+ API_KEY: str
6
7
  ARGS: str
7
8
  ARGUMENTS: str
9
+ BASE_URL: str
8
10
  CALL_ID: str
9
11
  CONTAINER: str
10
12
  CONTENT: str
@@ -2,7 +2,9 @@ from enum import StrEnum
2
2
 
3
3
  class Key:
4
4
  """Defines valid keys in OpenAI Chat Completions."""
5
+ API_KEY: str
5
6
  ARGUMENTS: str
7
+ BASE_URL: str
6
8
  CONTENT: str
7
9
  CHOICES: str
8
10
  DATA: str
gllm_inference.pyi CHANGED
@@ -44,7 +44,6 @@ import gllm_core.utils.imports
44
44
  import gllm_inference.schema.ModelId
45
45
  import gllm_inference.schema.ModelProvider
46
46
  import gllm_inference.schema.TruncationConfig
47
- import openai
48
47
  import asyncio
49
48
  import enum
50
49
  import gllm_inference.exceptions.BaseInvokerError
@@ -70,6 +69,7 @@ import gllm_inference.exceptions.InvokerRuntimeError
70
69
  import gllm_inference.exceptions.build_debug_info
71
70
  import gllm_inference.utils.load_langchain_model
72
71
  import gllm_inference.utils.parse_model_data
72
+ import openai
73
73
  import io
74
74
  import httpx
75
75
  import twelvelabs
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.2
2
2
  Name: gllm-inference-binary
3
- Version: 0.5.37
3
+ Version: 0.5.38
4
4
  Summary: A library containing components related to model inferences in Gen AI applications.
5
5
  Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, Resti Febrina <resti.febrina@gdplabs.id>
6
6
  Requires-Python: <3.14,>=3.11
@@ -1,5 +1,5 @@
1
- gllm_inference.cpython-311-x86_64-linux-gnu.so,sha256=UxGVABj0FeTj6Xf919Ya1tLaPn1qvy5dx4zxdCBzlio,4897232
2
- gllm_inference.pyi,sha256=TS9pXKtqkj_3NepYCDtGH4yp9caJ9gMBJFXL7t8P9YQ,4730
1
+ gllm_inference.cpython-311-x86_64-linux-gnu.so,sha256=6pAP4J3vu5vRU7g5zCuLjYpYgF_dHMqDtH1R8lypYGY,4909520
2
+ gllm_inference.pyi,sha256=esCvmd7d4XUM_kXqLmlXmHAm_wrcKJuNu30Yx1UCPQo,4730
3
3
  gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
4
4
  gllm_inference/constants.pyi,sha256=uCf0Rc8GkKwUKeMC5wx5WG_x0iV_piKhjZQl1gNffSQ,291
5
5
  gllm_inference/builder/__init__.pyi,sha256=usz2lvfwO4Yk-ZGKXbCWG1cEr3nlQXxMNDNC-2yc1NM,500
@@ -12,13 +12,13 @@ gllm_inference/catalog/catalog.pyi,sha256=a4RNG1lKv51GxQpOqh47tz-PAROMPaeP2o5XNL
12
12
  gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=WW1j8jWujnatF0c9rCk94CyobtFe3gOky2vjoGCV6nw,5424
13
13
  gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=OU8k_4HbqjZEzHZlzSM3uzGQZJmM2uGD76Csqom0CEQ,3197
14
14
  gllm_inference/em_invoker/__init__.pyi,sha256=83QVCkMjS2-jMKdAvmZska4LuJ-un755lAxjuVSLZ9o,987
15
- gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=Tp92kRqUmB2FpqdnWdJXDGZ_ibOzgFeZAEey4coaD5E,4933
15
+ gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=gICS8iadpItNaND3wzDIuQyYyHxz71QKXOlNboN8UEw,4959
16
16
  gllm_inference/em_invoker/bedrock_em_invoker.pyi,sha256=77omAUXnGU_62KPC5sKOnWTwdqsoEXX38IM1JScd6K0,5723
17
17
  gllm_inference/em_invoker/em_invoker.pyi,sha256=KGjLiAWGIA3ziV50zMwSzx6lTVUbYspQCl4LFxqbDlY,5101
18
18
  gllm_inference/em_invoker/google_em_invoker.pyi,sha256=VVpizDo47kznqRk07t6e-Lp_K6Ojfn5KDkIKoqn2boE,6822
19
19
  gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=BhZjkYZoxQXPJjP0PgA8W0xrtwiqDkW5E6NpGit0h1E,3498
20
- gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=WMJTEYdkRh5Mro8hcnZ1TIrL6rXk7bDYmsKzjGU8wwA,2848
21
- gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=m7zhI75nXXa9OwZSSnt59Aoy-V2KuApUI2EESm1Puqk,6162
20
+ gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=z4_jEuaprna6AJF2wXH-JgmynpVFXeb2HQZr-0ATmMw,2882
21
+ gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=FxQNV1xGEnHEWUak95OWVAwFPx65z0fphLLQ7R1qJLA,6193
22
22
  gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=9F37VuLYTH5xsPSS_uzhrsPJug0QifX_qh2GwG0jSTU,5345
23
23
  gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=zJZqMvvFKu3sHdrNM773UYjfHVlnwE2w2BmvdFcHzV0,5515
24
24
  gllm_inference/em_invoker/langchain/__init__.pyi,sha256=VYGKE5OgU0my1RlhgzkU_A7-GLGnUDDnNFuctuRwILE,148
@@ -37,27 +37,27 @@ gllm_inference/exceptions/exceptions.pyi,sha256=Bv996qLa_vju0Qjf4GewMxdkq8CV9LRZ
37
37
  gllm_inference/exceptions/provider_error_map.pyi,sha256=P1WnhWkM103FW6hqMfNZBOmYSWOmsJtll3VQV8DGb8E,1210
38
38
  gllm_inference/lm_invoker/__init__.pyi,sha256=IGF3h8Z7Yr0bLrkDMRTDVPNBU6Y3liJabW3acjstJDY,1374
39
39
  gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=inf_fuiy89t9CPiwCs_gsF2k7G7qZ9hMKEKYsGXLtNQ,17237
40
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=7zF8pj3LbuOwu1QArPX0ra6IrqUq5AkkGbC_wXiGhlA,15027
40
+ gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=Ex91dB8-2lEo2ZKHUJghNGehAfGstQqcGzIKQ75jSUk,15056
41
41
  gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=uZ9wpzOKSOvgu1ICMLqEXcrOE3RIbUmqHmgtuwBekPg,12802
42
- gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=J_tfnIgVDr-zQ7YE5_TKMyZyA336ly04g1l-ZKnr1As,9315
42
+ gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=T4p2abiP4Vv7psroJ7_rtcSqgzBkFp_B0HGFeIZW3H8,9349
43
43
  gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=u4egvGHRG30IotgAN6r_jrjDWFujDxjTdyoIlNLXjYA,17166
44
44
  gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=Bxi9hT4-sIZTz-VhAlxd5JyjAjQZQbUq2xtGhUL6oIs,13705
45
45
  gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=5dEgqLvMB04W30PO2MEUJ4h7rEMmM7ETdQ7LOlLpeXg,13238
46
46
  gllm_inference/lm_invoker/lm_invoker.pyi,sha256=IDqmtIe9T47mrqJzb0e7dynH-9gbwqhmcb7EeVzPFos,8167
47
- gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=94gLqMhuPUGkC_gwV1-GXnQHCxDrGgHf3wQ8lNqbqdc,15738
48
- gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=PFOqWhLGHFcoZROW-yreldVaZjHKTzPhuZY2ELM6_SY,3837
49
- gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=BtjSKEPc75xDJIU2AXY9AZL8600rYRsmx-C7_ZjTwfQ,23807
47
+ gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=Z3eAshWJzD8k1Phyo78GUWV8LkdOYOvZqo5ZFWYdJqE,15772
48
+ gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=by0LR_YqP6-uL3VMRq94t_ykO03wGz6Vv71UpWavQtk,3864
49
+ gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=vBo4XEZCmZBiScA6-n5aALJzIU90vsIPRah2p9neEzY,23841
50
50
  gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=KtCGKADVbrLfvkFOPHoeLNjlA2TKRf8F_kQHekSvQQI,15689
51
51
  gllm_inference/lm_invoker/batch/__init__.pyi,sha256=W4W-_yfk7lL20alREJai6GnwuQvdlKRfwQCX4mQK4XI,127
52
52
  gllm_inference/lm_invoker/batch/batch_operations.pyi,sha256=Oo7hoyPSfPZdy1mXvSdvtRndvq-XTIbPIjEoGvJj5C0,5372
53
53
  gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
54
54
  gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=6lreMyHKRfZzX5NBYKnQf1Z6RzXBjTvqZj2VbMeaTLQ,1098
55
55
  gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=FJLY-ZkkLUYDV48pfsLatnot4ev_xxz9xAayLK28CpU,1027
56
- gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=aA4DhTXIezwLvFzphR24a5ueVln2FCBIloP9Hbt3iz4,230
56
+ gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=WSuwOqL1j2ZioCZFC-gbB7vTRIZHQ3sU40c3ool5L6c,265
57
57
  gllm_inference/lm_invoker/schema/google.pyi,sha256=AIsNgq0ZZuicHmx4bL7z6q-946T05nWts3HUeA8hhHQ,505
58
58
  gllm_inference/lm_invoker/schema/langchain.pyi,sha256=rZcIxuvABI4pKfyVvkRBRqfJJogZ67EFPydpubHt49c,429
59
- gllm_inference/lm_invoker/schema/openai.pyi,sha256=oju4itbH6mm-yMCqX3m-448XJra4cg6oHHq7abYGM_g,2187
60
- gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=RyULzoGmIHuPwIwlSCg6dFdAYrHhZ0OttNxP2TCMWY8,1164
59
+ gllm_inference/lm_invoker/schema/openai.pyi,sha256=J_rT5Z3rx0hLIae-me1ENeemOESpavcRmYI5pgpkhhk,2222
60
+ gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=8byBRZ4xyTidIQJsZqiSjp5t1X875Obe-aEbT0yYfuA,1199
61
61
  gllm_inference/lm_invoker/schema/xai.pyi,sha256=cWnbJmDtllqRH3NXpQbiXgkNBcUXr8ksDSDywcgJebE,632
62
62
  gllm_inference/model/__init__.pyi,sha256=qClHIgljqhPPCKlGTKmHsWgYb4_hADybxtC2q1U8a5Q,593
63
63
  gllm_inference/model/em/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
@@ -117,7 +117,7 @@ gllm_inference/utils/io_utils.pyi,sha256=7kUTacHAVRYoemFUOjCH7-Qmw-YsQGd6rGYxjf_
117
117
  gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
118
118
  gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
119
119
  gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
120
- gllm_inference_binary-0.5.37.dist-info/METADATA,sha256=EZZMljrGByJU9D5YPcM5O9XBdgEBcq-h_haG61XsTTU,5636
121
- gllm_inference_binary-0.5.37.dist-info/WHEEL,sha256=WMelAR6z66VnlU3tu68fV-jM5qbG8iPyeTqaBcpU3pI,108
122
- gllm_inference_binary-0.5.37.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
123
- gllm_inference_binary-0.5.37.dist-info/RECORD,,
120
+ gllm_inference_binary-0.5.38.dist-info/METADATA,sha256=FU3f7oxGCu59ZA1CbA4zZi2in2Bt21J5hJtEJ8zbtUI,5636
121
+ gllm_inference_binary-0.5.38.dist-info/WHEEL,sha256=WMelAR6z66VnlU3tu68fV-jM5qbG8iPyeTqaBcpU3pI,108
122
+ gllm_inference_binary-0.5.38.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
123
+ gllm_inference_binary-0.5.38.dist-info/RECORD,,