gllm-inference-binary 0.5.37__cp311-cp311-manylinux_2_31_x86_64.whl → 0.5.38__cp311-cp311-manylinux_2_31_x86_64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of gllm-inference-binary might be problematic. Click here for more details.
- gllm_inference/em_invoker/azure_openai_em_invoker.pyi +2 -2
- gllm_inference/em_invoker/openai_compatible_em_invoker.pyi +2 -2
- gllm_inference/em_invoker/openai_em_invoker.pyi +2 -2
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi +2 -2
- gllm_inference/lm_invoker/datasaur_lm_invoker.pyi +2 -2
- gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi +2 -2
- gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi +1 -1
- gllm_inference/lm_invoker/openai_lm_invoker.pyi +2 -2
- gllm_inference/lm_invoker/schema/datasaur.pyi +2 -0
- gllm_inference/lm_invoker/schema/openai.pyi +2 -0
- gllm_inference/lm_invoker/schema/openai_chat_completions.pyi +2 -0
- gllm_inference.cpython-311-x86_64-linux-gnu.so +0 -0
- gllm_inference.pyi +1 -1
- {gllm_inference_binary-0.5.37.dist-info → gllm_inference_binary-0.5.38.dist-info}/METADATA +1 -1
- {gllm_inference_binary-0.5.37.dist-info → gllm_inference_binary-0.5.38.dist-info}/RECORD +17 -17
- {gllm_inference_binary-0.5.37.dist-info → gllm_inference_binary-0.5.38.dist-info}/WHEEL +0 -0
- {gllm_inference_binary-0.5.37.dist-info → gllm_inference_binary-0.5.38.dist-info}/top_level.txt +0 -0
|
@@ -13,7 +13,7 @@ class AzureOpenAIEMInvoker(OpenAIEMInvoker):
|
|
|
13
13
|
model_id (str): The model ID of the embedding model.
|
|
14
14
|
model_provider (str): The provider of the embedding model.
|
|
15
15
|
model_name (str): The name of the Azure OpenAI embedding model deployment.
|
|
16
|
-
|
|
16
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the Azure OpenAI client.
|
|
17
17
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the embedding model.
|
|
18
18
|
retry_config (RetryConfig): The retry configuration for the embedding model.
|
|
19
19
|
truncation_config (TruncationConfig | None): The truncation configuration for the embedding model.
|
|
@@ -68,7 +68,7 @@ class AzureOpenAIEMInvoker(OpenAIEMInvoker):
|
|
|
68
68
|
em_invoker = AzureOpenAIEMInvoker(..., retry_config=retry_config)
|
|
69
69
|
```
|
|
70
70
|
'''
|
|
71
|
-
|
|
71
|
+
client_kwargs: Incomplete
|
|
72
72
|
def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
|
|
73
73
|
"""Initializes a new instance of the AzureOpenAIEMInvoker class.
|
|
74
74
|
|
|
@@ -15,14 +15,14 @@ class OpenAICompatibleEMInvoker(OpenAIEMInvoker):
|
|
|
15
15
|
model_id (str): The model ID of the embedding model.
|
|
16
16
|
model_provider (str): The provider of the embedding model.
|
|
17
17
|
model_name (str): The name of the embedding model.
|
|
18
|
-
|
|
18
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
|
|
19
19
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the embedding model.
|
|
20
20
|
retry_config (RetryConfig): The retry configuration for the embedding model.
|
|
21
21
|
truncation_config (TruncationConfig | None): The truncation configuration for the embedding model.
|
|
22
22
|
|
|
23
23
|
This class is deprecated and will be removed in v0.6. Please use the `OpenAIEMInvoker` class instead.
|
|
24
24
|
"""
|
|
25
|
-
|
|
25
|
+
client_kwargs: Incomplete
|
|
26
26
|
def __init__(self, model_name: str, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
|
|
27
27
|
"""Initializes a new instance of the OpenAICompatibleEMInvoker class.
|
|
28
28
|
|
|
@@ -16,7 +16,7 @@ class OpenAIEMInvoker(BaseEMInvoker):
|
|
|
16
16
|
model_id (str): The model ID of the embedding model.
|
|
17
17
|
model_provider (str): The provider of the embedding model.
|
|
18
18
|
model_name (str): The name of the embedding model.
|
|
19
|
-
|
|
19
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
|
|
20
20
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the embedding model.
|
|
21
21
|
retry_config (RetryConfig): The retry configuration for the embedding model.
|
|
22
22
|
truncation_config (TruncationConfig | None): The truncation configuration for the embedding model.
|
|
@@ -96,7 +96,7 @@ class OpenAIEMInvoker(BaseEMInvoker):
|
|
|
96
96
|
em_invoker = OpenAIEMInvoker(..., retry_config=retry_config)
|
|
97
97
|
```
|
|
98
98
|
'''
|
|
99
|
-
|
|
99
|
+
client_kwargs: Incomplete
|
|
100
100
|
def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
|
|
101
101
|
'''Initializes a new instance of the OpenAIEMInvoker class.
|
|
102
102
|
|
|
@@ -15,7 +15,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
|
|
|
15
15
|
model_id (str): The model ID of the language model.
|
|
16
16
|
model_provider (str): The provider of the language model.
|
|
17
17
|
model_name (str): The name of the Azure OpenAI language model deployment.
|
|
18
|
-
|
|
18
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the Azure OpenAI client.
|
|
19
19
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
|
|
20
20
|
tools (list[Tool]): The list of tools provided to the model to enable tool calling.
|
|
21
21
|
response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
|
|
@@ -226,7 +226,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
|
|
|
226
226
|
Defaults to an empty list.
|
|
227
227
|
2.10. mcp_calls (list[MCPCall]): The MCP calls. Currently not supported. Defaults to an empty list.
|
|
228
228
|
'''
|
|
229
|
-
|
|
229
|
+
client_kwargs: Incomplete
|
|
230
230
|
def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None) -> None:
|
|
231
231
|
"""Initializes a new instance of the AzureOpenAILMInvoker class.
|
|
232
232
|
|
|
@@ -18,7 +18,7 @@ class DatasaurLMInvoker(OpenAICompatibleLMInvoker):
|
|
|
18
18
|
model_id (str): The model ID of the language model.
|
|
19
19
|
model_provider (str): The provider of the language model.
|
|
20
20
|
model_name (str): The name of the language model.
|
|
21
|
-
|
|
21
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
|
|
22
22
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
|
|
23
23
|
tools (list[Any]): The list of tools provided to the model to enable tool calling. Currently not supported.
|
|
24
24
|
response_schema (ResponseSchema | None): The schema of the response. Currently not supported.
|
|
@@ -121,7 +121,7 @@ class DatasaurLMInvoker(OpenAICompatibleLMInvoker):
|
|
|
121
121
|
Defaults to an empty list.
|
|
122
122
|
2.10. mcp_calls (list[MCPCall]): The MCP calls. Currently not supported. Defaults to an empty list.
|
|
123
123
|
'''
|
|
124
|
-
|
|
124
|
+
client_kwargs: Incomplete
|
|
125
125
|
citations: Incomplete
|
|
126
126
|
def __init__(self, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, citations: bool = False) -> None:
|
|
127
127
|
"""Initializes a new instance of the DatasaurLMInvoker class.
|
|
@@ -23,7 +23,7 @@ class OpenAIChatCompletionsLMInvoker(BaseLMInvoker):
|
|
|
23
23
|
model_id (str): The model ID of the language model.
|
|
24
24
|
model_provider (str): The provider of the language model.
|
|
25
25
|
model_name (str): The name of the language model.
|
|
26
|
-
|
|
26
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
|
|
27
27
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
|
|
28
28
|
tools (list[Tool]): The list of tools provided to the model to enable tool calling.
|
|
29
29
|
response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
|
|
@@ -243,7 +243,7 @@ class OpenAIChatCompletionsLMInvoker(BaseLMInvoker):
|
|
|
243
243
|
Defaults to an empty list.
|
|
244
244
|
2.10. mcp_calls (list[MCPCall]): The MCP calls. Currently not supported. Defaults to an empty list.
|
|
245
245
|
'''
|
|
246
|
-
|
|
246
|
+
client_kwargs: Incomplete
|
|
247
247
|
def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None) -> None:
|
|
248
248
|
'''Initializes a new instance of the OpenAIChatCompletionsLMInvoker class.
|
|
249
249
|
|
|
@@ -15,7 +15,7 @@ class OpenAICompatibleLMInvoker(OpenAIChatCompletionsLMInvoker):
|
|
|
15
15
|
model_id (str): The model ID of the language model.
|
|
16
16
|
model_provider (str): The provider of the language model.
|
|
17
17
|
model_name (str): The name of the language model.
|
|
18
|
-
|
|
18
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
|
|
19
19
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
|
|
20
20
|
tools (list[Tool]): The list of tools provided to the model to enable tool calling.
|
|
21
21
|
response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
|
|
@@ -25,7 +25,7 @@ class OpenAILMInvoker(BaseLMInvoker):
|
|
|
25
25
|
model_id (str): The model ID of the language model.
|
|
26
26
|
model_provider (str): The provider of the language model.
|
|
27
27
|
model_name (str): The name of the language model.
|
|
28
|
-
|
|
28
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
|
|
29
29
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
|
|
30
30
|
tools (list[Tool]): The list of tools provided to the model to enable tool calling.
|
|
31
31
|
response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
|
|
@@ -391,7 +391,7 @@ class OpenAILMInvoker(BaseLMInvoker):
|
|
|
391
391
|
2.10. mcp_calls (list[MCPCall]): The MCP calls, if the MCP servers are provided and the language model
|
|
392
392
|
decides to invoke MCP tools. Defaults to an empty list.
|
|
393
393
|
'''
|
|
394
|
-
|
|
394
|
+
client_kwargs: Incomplete
|
|
395
395
|
def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None, mcp_servers: list[MCPServer] | None = None, code_interpreter: bool = False, web_search: bool = False) -> None:
|
|
396
396
|
'''Initializes a new instance of the OpenAILMInvoker class.
|
|
397
397
|
|
|
Binary file
|
gllm_inference.pyi
CHANGED
|
@@ -44,7 +44,6 @@ import gllm_core.utils.imports
|
|
|
44
44
|
import gllm_inference.schema.ModelId
|
|
45
45
|
import gllm_inference.schema.ModelProvider
|
|
46
46
|
import gllm_inference.schema.TruncationConfig
|
|
47
|
-
import openai
|
|
48
47
|
import asyncio
|
|
49
48
|
import enum
|
|
50
49
|
import gllm_inference.exceptions.BaseInvokerError
|
|
@@ -70,6 +69,7 @@ import gllm_inference.exceptions.InvokerRuntimeError
|
|
|
70
69
|
import gllm_inference.exceptions.build_debug_info
|
|
71
70
|
import gllm_inference.utils.load_langchain_model
|
|
72
71
|
import gllm_inference.utils.parse_model_data
|
|
72
|
+
import openai
|
|
73
73
|
import io
|
|
74
74
|
import httpx
|
|
75
75
|
import twelvelabs
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.2
|
|
2
2
|
Name: gllm-inference-binary
|
|
3
|
-
Version: 0.5.
|
|
3
|
+
Version: 0.5.38
|
|
4
4
|
Summary: A library containing components related to model inferences in Gen AI applications.
|
|
5
5
|
Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, Resti Febrina <resti.febrina@gdplabs.id>
|
|
6
6
|
Requires-Python: <3.14,>=3.11
|
|
@@ -1,5 +1,5 @@
|
|
|
1
|
-
gllm_inference.cpython-311-x86_64-linux-gnu.so,sha256=
|
|
2
|
-
gllm_inference.pyi,sha256=
|
|
1
|
+
gllm_inference.cpython-311-x86_64-linux-gnu.so,sha256=6pAP4J3vu5vRU7g5zCuLjYpYgF_dHMqDtH1R8lypYGY,4909520
|
|
2
|
+
gllm_inference.pyi,sha256=esCvmd7d4XUM_kXqLmlXmHAm_wrcKJuNu30Yx1UCPQo,4730
|
|
3
3
|
gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
4
4
|
gllm_inference/constants.pyi,sha256=uCf0Rc8GkKwUKeMC5wx5WG_x0iV_piKhjZQl1gNffSQ,291
|
|
5
5
|
gllm_inference/builder/__init__.pyi,sha256=usz2lvfwO4Yk-ZGKXbCWG1cEr3nlQXxMNDNC-2yc1NM,500
|
|
@@ -12,13 +12,13 @@ gllm_inference/catalog/catalog.pyi,sha256=a4RNG1lKv51GxQpOqh47tz-PAROMPaeP2o5XNL
|
|
|
12
12
|
gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=WW1j8jWujnatF0c9rCk94CyobtFe3gOky2vjoGCV6nw,5424
|
|
13
13
|
gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=OU8k_4HbqjZEzHZlzSM3uzGQZJmM2uGD76Csqom0CEQ,3197
|
|
14
14
|
gllm_inference/em_invoker/__init__.pyi,sha256=83QVCkMjS2-jMKdAvmZska4LuJ-un755lAxjuVSLZ9o,987
|
|
15
|
-
gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=
|
|
15
|
+
gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=gICS8iadpItNaND3wzDIuQyYyHxz71QKXOlNboN8UEw,4959
|
|
16
16
|
gllm_inference/em_invoker/bedrock_em_invoker.pyi,sha256=77omAUXnGU_62KPC5sKOnWTwdqsoEXX38IM1JScd6K0,5723
|
|
17
17
|
gllm_inference/em_invoker/em_invoker.pyi,sha256=KGjLiAWGIA3ziV50zMwSzx6lTVUbYspQCl4LFxqbDlY,5101
|
|
18
18
|
gllm_inference/em_invoker/google_em_invoker.pyi,sha256=VVpizDo47kznqRk07t6e-Lp_K6Ojfn5KDkIKoqn2boE,6822
|
|
19
19
|
gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=BhZjkYZoxQXPJjP0PgA8W0xrtwiqDkW5E6NpGit0h1E,3498
|
|
20
|
-
gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=
|
|
21
|
-
gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=
|
|
20
|
+
gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=z4_jEuaprna6AJF2wXH-JgmynpVFXeb2HQZr-0ATmMw,2882
|
|
21
|
+
gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=FxQNV1xGEnHEWUak95OWVAwFPx65z0fphLLQ7R1qJLA,6193
|
|
22
22
|
gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=9F37VuLYTH5xsPSS_uzhrsPJug0QifX_qh2GwG0jSTU,5345
|
|
23
23
|
gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=zJZqMvvFKu3sHdrNM773UYjfHVlnwE2w2BmvdFcHzV0,5515
|
|
24
24
|
gllm_inference/em_invoker/langchain/__init__.pyi,sha256=VYGKE5OgU0my1RlhgzkU_A7-GLGnUDDnNFuctuRwILE,148
|
|
@@ -37,27 +37,27 @@ gllm_inference/exceptions/exceptions.pyi,sha256=Bv996qLa_vju0Qjf4GewMxdkq8CV9LRZ
|
|
|
37
37
|
gllm_inference/exceptions/provider_error_map.pyi,sha256=P1WnhWkM103FW6hqMfNZBOmYSWOmsJtll3VQV8DGb8E,1210
|
|
38
38
|
gllm_inference/lm_invoker/__init__.pyi,sha256=IGF3h8Z7Yr0bLrkDMRTDVPNBU6Y3liJabW3acjstJDY,1374
|
|
39
39
|
gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=inf_fuiy89t9CPiwCs_gsF2k7G7qZ9hMKEKYsGXLtNQ,17237
|
|
40
|
-
gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=
|
|
40
|
+
gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=Ex91dB8-2lEo2ZKHUJghNGehAfGstQqcGzIKQ75jSUk,15056
|
|
41
41
|
gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=uZ9wpzOKSOvgu1ICMLqEXcrOE3RIbUmqHmgtuwBekPg,12802
|
|
42
|
-
gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=
|
|
42
|
+
gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=T4p2abiP4Vv7psroJ7_rtcSqgzBkFp_B0HGFeIZW3H8,9349
|
|
43
43
|
gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=u4egvGHRG30IotgAN6r_jrjDWFujDxjTdyoIlNLXjYA,17166
|
|
44
44
|
gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=Bxi9hT4-sIZTz-VhAlxd5JyjAjQZQbUq2xtGhUL6oIs,13705
|
|
45
45
|
gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=5dEgqLvMB04W30PO2MEUJ4h7rEMmM7ETdQ7LOlLpeXg,13238
|
|
46
46
|
gllm_inference/lm_invoker/lm_invoker.pyi,sha256=IDqmtIe9T47mrqJzb0e7dynH-9gbwqhmcb7EeVzPFos,8167
|
|
47
|
-
gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=
|
|
48
|
-
gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=
|
|
49
|
-
gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=
|
|
47
|
+
gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=Z3eAshWJzD8k1Phyo78GUWV8LkdOYOvZqo5ZFWYdJqE,15772
|
|
48
|
+
gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=by0LR_YqP6-uL3VMRq94t_ykO03wGz6Vv71UpWavQtk,3864
|
|
49
|
+
gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=vBo4XEZCmZBiScA6-n5aALJzIU90vsIPRah2p9neEzY,23841
|
|
50
50
|
gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=KtCGKADVbrLfvkFOPHoeLNjlA2TKRf8F_kQHekSvQQI,15689
|
|
51
51
|
gllm_inference/lm_invoker/batch/__init__.pyi,sha256=W4W-_yfk7lL20alREJai6GnwuQvdlKRfwQCX4mQK4XI,127
|
|
52
52
|
gllm_inference/lm_invoker/batch/batch_operations.pyi,sha256=Oo7hoyPSfPZdy1mXvSdvtRndvq-XTIbPIjEoGvJj5C0,5372
|
|
53
53
|
gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
54
54
|
gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=6lreMyHKRfZzX5NBYKnQf1Z6RzXBjTvqZj2VbMeaTLQ,1098
|
|
55
55
|
gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=FJLY-ZkkLUYDV48pfsLatnot4ev_xxz9xAayLK28CpU,1027
|
|
56
|
-
gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=
|
|
56
|
+
gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=WSuwOqL1j2ZioCZFC-gbB7vTRIZHQ3sU40c3ool5L6c,265
|
|
57
57
|
gllm_inference/lm_invoker/schema/google.pyi,sha256=AIsNgq0ZZuicHmx4bL7z6q-946T05nWts3HUeA8hhHQ,505
|
|
58
58
|
gllm_inference/lm_invoker/schema/langchain.pyi,sha256=rZcIxuvABI4pKfyVvkRBRqfJJogZ67EFPydpubHt49c,429
|
|
59
|
-
gllm_inference/lm_invoker/schema/openai.pyi,sha256=
|
|
60
|
-
gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=
|
|
59
|
+
gllm_inference/lm_invoker/schema/openai.pyi,sha256=J_rT5Z3rx0hLIae-me1ENeemOESpavcRmYI5pgpkhhk,2222
|
|
60
|
+
gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=8byBRZ4xyTidIQJsZqiSjp5t1X875Obe-aEbT0yYfuA,1199
|
|
61
61
|
gllm_inference/lm_invoker/schema/xai.pyi,sha256=cWnbJmDtllqRH3NXpQbiXgkNBcUXr8ksDSDywcgJebE,632
|
|
62
62
|
gllm_inference/model/__init__.pyi,sha256=qClHIgljqhPPCKlGTKmHsWgYb4_hADybxtC2q1U8a5Q,593
|
|
63
63
|
gllm_inference/model/em/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
@@ -117,7 +117,7 @@ gllm_inference/utils/io_utils.pyi,sha256=7kUTacHAVRYoemFUOjCH7-Qmw-YsQGd6rGYxjf_
|
|
|
117
117
|
gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
|
|
118
118
|
gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
|
|
119
119
|
gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
|
|
120
|
-
gllm_inference_binary-0.5.
|
|
121
|
-
gllm_inference_binary-0.5.
|
|
122
|
-
gllm_inference_binary-0.5.
|
|
123
|
-
gllm_inference_binary-0.5.
|
|
120
|
+
gllm_inference_binary-0.5.38.dist-info/METADATA,sha256=FU3f7oxGCu59ZA1CbA4zZi2in2Bt21J5hJtEJ8zbtUI,5636
|
|
121
|
+
gllm_inference_binary-0.5.38.dist-info/WHEEL,sha256=WMelAR6z66VnlU3tu68fV-jM5qbG8iPyeTqaBcpU3pI,108
|
|
122
|
+
gllm_inference_binary-0.5.38.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
|
|
123
|
+
gllm_inference_binary-0.5.38.dist-info/RECORD,,
|
|
File without changes
|
{gllm_inference_binary-0.5.37.dist-info → gllm_inference_binary-0.5.38.dist-info}/top_level.txt
RENAMED
|
File without changes
|