gllm-inference-binary 0.5.51__cp311-cp311-win_amd64.whl → 0.5.66__cp311-cp311-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- gllm_inference/builder/build_lm_invoker.pyi +10 -1
- gllm_inference/constants.pyi +0 -1
- gllm_inference/em_invoker/cohere_em_invoker.pyi +1 -2
- gllm_inference/lm_invoker/__init__.pyi +2 -1
- gllm_inference/lm_invoker/anthropic_lm_invoker.pyi +95 -109
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi +92 -109
- gllm_inference/lm_invoker/batch/batch_operations.pyi +2 -1
- gllm_inference/lm_invoker/bedrock_lm_invoker.pyi +51 -65
- gllm_inference/lm_invoker/datasaur_lm_invoker.pyi +36 -36
- gllm_inference/lm_invoker/google_lm_invoker.pyi +195 -110
- gllm_inference/lm_invoker/langchain_lm_invoker.pyi +52 -64
- gllm_inference/lm_invoker/litellm_lm_invoker.pyi +86 -106
- gllm_inference/lm_invoker/lm_invoker.pyi +20 -1
- gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi +87 -107
- gllm_inference/lm_invoker/openai_lm_invoker.pyi +235 -185
- gllm_inference/lm_invoker/portkey_lm_invoker.pyi +104 -68
- gllm_inference/lm_invoker/schema/google.pyi +12 -0
- gllm_inference/lm_invoker/schema/openai.pyi +22 -0
- gllm_inference/lm_invoker/sea_lion_lm_invoker.pyi +48 -0
- gllm_inference/lm_invoker/xai_lm_invoker.pyi +94 -131
- gllm_inference/model/__init__.pyi +2 -1
- gllm_inference/model/lm/sea_lion_lm.pyi +16 -0
- gllm_inference/prompt_builder/prompt_builder.pyi +6 -2
- gllm_inference/schema/__init__.pyi +4 -3
- gllm_inference/schema/attachment.pyi +20 -6
- gllm_inference/schema/enums.pyi +14 -1
- gllm_inference/schema/events.pyi +2 -2
- gllm_inference/schema/formatter.pyi +31 -0
- gllm_inference/schema/lm_output.pyi +245 -23
- gllm_inference/schema/model_id.pyi +1 -1
- gllm_inference/utils/validation.pyi +3 -0
- gllm_inference.cp311-win_amd64.pyd +0 -0
- gllm_inference.pyi +5 -6
- {gllm_inference_binary-0.5.51.dist-info → gllm_inference_binary-0.5.66.dist-info}/METADATA +5 -5
- {gllm_inference_binary-0.5.51.dist-info → gllm_inference_binary-0.5.66.dist-info}/RECORD +37 -34
- {gllm_inference_binary-0.5.51.dist-info → gllm_inference_binary-0.5.66.dist-info}/WHEEL +0 -0
- {gllm_inference_binary-0.5.51.dist-info → gllm_inference_binary-0.5.66.dist-info}/top_level.txt +0 -0
|
@@ -1,6 +1,8 @@
|
|
|
1
|
-
from
|
|
1
|
+
from _typeshed import Incomplete
|
|
2
|
+
from gllm_core.schema import Chunk
|
|
2
3
|
from gllm_inference.schema.attachment import Attachment as Attachment
|
|
3
4
|
from gllm_inference.schema.code_exec_result import CodeExecResult as CodeExecResult
|
|
5
|
+
from gllm_inference.schema.enums import LMOutputType as LMOutputType
|
|
4
6
|
from gllm_inference.schema.mcp import MCPCall as MCPCall
|
|
5
7
|
from gllm_inference.schema.reasoning import Reasoning as Reasoning
|
|
6
8
|
from gllm_inference.schema.token_usage import TokenUsage as TokenUsage
|
|
@@ -8,37 +10,257 @@ from gllm_inference.schema.tool_call import ToolCall as ToolCall
|
|
|
8
10
|
from pydantic import BaseModel
|
|
9
11
|
from typing import Any
|
|
10
12
|
|
|
13
|
+
LMOutputData = str | dict[str, Any] | BaseModel | Attachment | ToolCall | Reasoning | Chunk | CodeExecResult | MCPCall
|
|
14
|
+
logger: Incomplete
|
|
15
|
+
|
|
16
|
+
class LMOutputItem(BaseModel):
|
|
17
|
+
"""Defines the output item of a language model.
|
|
18
|
+
|
|
19
|
+
Attributes:
|
|
20
|
+
type (str): The type of the output item.
|
|
21
|
+
output (LMOutputData): The output data of the output item.
|
|
22
|
+
"""
|
|
23
|
+
type: str
|
|
24
|
+
output: LMOutputData
|
|
25
|
+
|
|
11
26
|
class LMOutput(BaseModel):
|
|
12
27
|
"""Defines the output of a language model.
|
|
13
28
|
|
|
14
29
|
Attributes:
|
|
15
|
-
|
|
16
|
-
attachments (list[Attachment]): The attachments, if the language model decides to output attachments.
|
|
17
|
-
Defaults to an empty list.
|
|
18
|
-
tool_calls (list[ToolCall]): The tool calls, if the language model decides to invoke tools.
|
|
19
|
-
Defaults to an empty list.
|
|
20
|
-
structured_output (dict[str, Any] | BaseModel | None): The structured output, if a response schema is defined
|
|
21
|
-
for the language model. Defaults to None.
|
|
30
|
+
outputs (list[LMOutputItem]): The outputs of the language model in sequential order. Defaults to an empty list.
|
|
22
31
|
token_usage (TokenUsage | None): The token usage analytics, if requested. Defaults to None.
|
|
23
32
|
duration (float | None): The duration of the invocation in seconds, if requested. Defaults to None.
|
|
24
33
|
finish_details (dict[str, Any]): The details about how the generation finished, if requested.
|
|
25
34
|
Defaults to an empty dictionary.
|
|
26
|
-
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
35
|
+
|
|
36
|
+
text (str): The first text response.
|
|
37
|
+
structured_output (dict[str, Any] | BaseModel | None): The first structured output.
|
|
38
|
+
|
|
39
|
+
texts (list[str]): The texts from the outputs.
|
|
40
|
+
structured_outputs (list[dict[str, Any] | BaseModel]): The structured outputs from the outputs.
|
|
41
|
+
attachments (list[Attachment]): The attachments from the outputs.
|
|
42
|
+
tool_calls (list[ToolCall]): The tool calls from the outputs.
|
|
43
|
+
thinkings (list[Reasoning]): The thinkings from the outputs.
|
|
44
|
+
citations (list[Chunk]): The citations from the outputs.
|
|
45
|
+
code_exec_results (list[CodeExecResult]): The code exec results from the outputs.
|
|
46
|
+
mcp_calls (list[MCPCall]): The MCP calls from the outputs.
|
|
47
|
+
|
|
48
|
+
response (str): Deprecated. Replaced by `text`.
|
|
49
|
+
reasoning (list[Reasoning]): Deprecated. Replaced by `thinkings`.
|
|
33
50
|
"""
|
|
34
|
-
|
|
35
|
-
attachments: list[Attachment]
|
|
36
|
-
tool_calls: list[ToolCall]
|
|
37
|
-
structured_output: dict[str, Any] | BaseModel | None
|
|
51
|
+
outputs: list[LMOutputItem]
|
|
38
52
|
token_usage: TokenUsage | None
|
|
39
53
|
duration: float | None
|
|
40
54
|
finish_details: dict[str, Any]
|
|
41
|
-
reasoning: list[Reasoning]
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
|
|
55
|
+
def __init__(self, *, outputs: list[LMOutputItem] | None = None, token_usage: TokenUsage | None = None, duration: float | None = None, finish_details: dict[str, Any] | None = None, response: str = '', structured_output: dict[str, Any] | BaseModel | None = None, tool_calls: list[ToolCall] | None = None, reasoning: list[Reasoning] | None = None, attachments: list[Attachment] | None = None, citations: list[Chunk] | None = None, code_exec_results: list[CodeExecResult] | None = None, mcp_calls: list[MCPCall] | None = None) -> None:
|
|
56
|
+
"""Initialize the LMOutput.
|
|
57
|
+
|
|
58
|
+
This constructor is created for backward compatibility with the legacy method to initialize the LMOutput.
|
|
59
|
+
This constructor will be removed in v0.6.
|
|
60
|
+
|
|
61
|
+
Args:
|
|
62
|
+
outputs (list[LMOutputItem] | None, optional): The output items. Defaults to an empty list.
|
|
63
|
+
token_usage (TokenUsage | None, optional): The token usage analytics. Defaults to None.
|
|
64
|
+
duration (float | None, optional): The duration of the invocation in seconds. Defaults to None.
|
|
65
|
+
finish_details (dict[str, Any] | None, optional): The details about how the generation finished.
|
|
66
|
+
Defaults to an empty dictionary.
|
|
67
|
+
response (str, optional): The first text response. Defaults to an empty string.
|
|
68
|
+
structured_output (dict[str, Any] | BaseModel | None, optional): The first structured output.
|
|
69
|
+
Defaults to None.
|
|
70
|
+
tool_calls (list[ToolCall] | None, optional): The tool calls. Defaults to None.
|
|
71
|
+
reasoning (list[Reasoning] | None, optional): The thinkings. Defaults to None.
|
|
72
|
+
attachments (list[Attachment] | None, optional): The attachments. Defaults to None.
|
|
73
|
+
citations (list[Chunk] | None, optional): The citations. Defaults to None.
|
|
74
|
+
code_exec_results (list[CodeExecResult] | None, optional): The code exec results. Defaults to None.
|
|
75
|
+
mcp_calls (list[MCPCall] | None, optional): The MCP calls. Defaults to None.
|
|
76
|
+
"""
|
|
77
|
+
@property
|
|
78
|
+
def response(self) -> str:
|
|
79
|
+
"""Deprecated property to get the first text response from the LMOutput.
|
|
80
|
+
|
|
81
|
+
Returns:
|
|
82
|
+
str: The first text response from the LMOutput.
|
|
83
|
+
"""
|
|
84
|
+
@response.setter
|
|
85
|
+
def response(self, value: str) -> None:
|
|
86
|
+
"""Deprecated setter to set the first text response to the LMOutput.
|
|
87
|
+
|
|
88
|
+
Args:
|
|
89
|
+
value (str): The first text response to set.
|
|
90
|
+
"""
|
|
91
|
+
@property
|
|
92
|
+
def text(self) -> str:
|
|
93
|
+
"""Get the first text from the LMOutput.
|
|
94
|
+
|
|
95
|
+
Returns:
|
|
96
|
+
str: The first text from the LMOutput.
|
|
97
|
+
"""
|
|
98
|
+
@property
|
|
99
|
+
def structured_output(self) -> dict[str, Any] | BaseModel | None:
|
|
100
|
+
"""Deprecated property to get the first structured output from the LMOutput.
|
|
101
|
+
|
|
102
|
+
Returns:
|
|
103
|
+
dict[str, Any] | BaseModel | None: The first structured output from the LMOutput.
|
|
104
|
+
"""
|
|
105
|
+
@structured_output.setter
|
|
106
|
+
def structured_output(self, value: dict[str, Any] | BaseModel) -> None:
|
|
107
|
+
"""Deprecated setter to set the first structured output to the LMOutput.
|
|
108
|
+
|
|
109
|
+
Args:
|
|
110
|
+
value (dict[str, Any] | BaseModel): The first structured output to set.
|
|
111
|
+
"""
|
|
112
|
+
@property
|
|
113
|
+
def texts(self) -> list[str]:
|
|
114
|
+
"""Get the texts from the LMOutput.
|
|
115
|
+
|
|
116
|
+
Returns:
|
|
117
|
+
list[str]: The texts from the LMOutput.
|
|
118
|
+
"""
|
|
119
|
+
@property
|
|
120
|
+
def structured_outputs(self) -> list[dict[str, Any] | BaseModel]:
|
|
121
|
+
"""Get the structured outputs from the LMOutput.
|
|
122
|
+
|
|
123
|
+
Returns:
|
|
124
|
+
list[dict[str, Any] | BaseModel]: The structured outputs from the LMOutput.
|
|
125
|
+
"""
|
|
126
|
+
@property
|
|
127
|
+
def attachments(self) -> list[Attachment]:
|
|
128
|
+
"""Get the attachments from the LMOutput.
|
|
129
|
+
|
|
130
|
+
Returns:
|
|
131
|
+
list[Attachment]: The attachments from the LMOutput.
|
|
132
|
+
"""
|
|
133
|
+
@attachments.setter
|
|
134
|
+
def attachments(self, value: list[Attachment]) -> None:
|
|
135
|
+
"""Deprecated setter to set the attachments to the LMOutput.
|
|
136
|
+
|
|
137
|
+
Args:
|
|
138
|
+
value (list[Attachment]): The attachments to set.
|
|
139
|
+
"""
|
|
140
|
+
@property
|
|
141
|
+
def tool_calls(self) -> list[ToolCall]:
|
|
142
|
+
"""Get the tool calls from the LMOutput.
|
|
143
|
+
|
|
144
|
+
Returns:
|
|
145
|
+
list[ToolCall]: The tool calls from the LMOutput.
|
|
146
|
+
"""
|
|
147
|
+
@tool_calls.setter
|
|
148
|
+
def tool_calls(self, value: list[ToolCall]) -> None:
|
|
149
|
+
"""Deprecated setter to set the tool calls to the LMOutput.
|
|
150
|
+
|
|
151
|
+
Args:
|
|
152
|
+
value (list[ToolCall]): The tool calls to set.
|
|
153
|
+
"""
|
|
154
|
+
@property
|
|
155
|
+
def reasoning(self) -> list[Reasoning]:
|
|
156
|
+
"""Deprecated property to get the thinkings from the LMOutput.
|
|
157
|
+
|
|
158
|
+
Returns:
|
|
159
|
+
list[Reasoning]: The thinkings from the LMOutput.
|
|
160
|
+
"""
|
|
161
|
+
@reasoning.setter
|
|
162
|
+
def reasoning(self, value: list[Reasoning]) -> None:
|
|
163
|
+
"""Deprecated setter to set the thinkings to the LMOutput.
|
|
164
|
+
|
|
165
|
+
Args:
|
|
166
|
+
value (list[Reasoning]): The thinkings to set.
|
|
167
|
+
"""
|
|
168
|
+
@property
|
|
169
|
+
def thinkings(self) -> list[Reasoning]:
|
|
170
|
+
"""Get the thinkings from the LMOutput.
|
|
171
|
+
|
|
172
|
+
Returns:
|
|
173
|
+
list[Reasoning]: The thinkings from the LMOutput.
|
|
174
|
+
"""
|
|
175
|
+
@property
|
|
176
|
+
def citations(self) -> list[Chunk]:
|
|
177
|
+
"""Get the citations from the LMOutput.
|
|
178
|
+
|
|
179
|
+
Returns:
|
|
180
|
+
list[Chunk]: The citations from the LMOutput.
|
|
181
|
+
"""
|
|
182
|
+
@citations.setter
|
|
183
|
+
def citations(self, value: list[Chunk]) -> None:
|
|
184
|
+
"""Deprecated setter to set the citations to the LMOutput.
|
|
185
|
+
|
|
186
|
+
Args:
|
|
187
|
+
value (list[Chunk]): The citations to set.
|
|
188
|
+
"""
|
|
189
|
+
@property
|
|
190
|
+
def code_exec_results(self) -> list[CodeExecResult]:
|
|
191
|
+
"""Get the code exec results from the LMOutput.
|
|
192
|
+
|
|
193
|
+
Returns:
|
|
194
|
+
list[CodeExecResult]: The code exec results from the LMOutput.
|
|
195
|
+
"""
|
|
196
|
+
@code_exec_results.setter
|
|
197
|
+
def code_exec_results(self, value: list[CodeExecResult]) -> None:
|
|
198
|
+
"""Deprecated setter to set the code exec results to the LMOutput.
|
|
199
|
+
|
|
200
|
+
Args:
|
|
201
|
+
value (list[CodeExecResult]): The code exec results to set.
|
|
202
|
+
"""
|
|
203
|
+
@property
|
|
204
|
+
def mcp_calls(self) -> list[MCPCall]:
|
|
205
|
+
"""Get the MCP calls from the LMOutput.
|
|
206
|
+
|
|
207
|
+
Returns:
|
|
208
|
+
list[MCPCall]: The MCP calls from the LMOutput.
|
|
209
|
+
"""
|
|
210
|
+
@mcp_calls.setter
|
|
211
|
+
def mcp_calls(self, value: list[MCPCall]) -> None:
|
|
212
|
+
"""Deprecated setter to set the MCP calls to the LMOutput.
|
|
213
|
+
|
|
214
|
+
Args:
|
|
215
|
+
value (list[MCPCall]): The MCP calls to set.
|
|
216
|
+
"""
|
|
217
|
+
def add_text(self, text: str | list[str]) -> None:
|
|
218
|
+
"""Add an output or a list of outputs to the LMOutput.
|
|
219
|
+
|
|
220
|
+
Args:
|
|
221
|
+
text (str | list[str]): The text or a list of texts to add.
|
|
222
|
+
"""
|
|
223
|
+
def add_attachment(self, attachment: Attachment | list[Attachment]) -> None:
|
|
224
|
+
"""Add an attachment or a list of attachments to the LMOutput.
|
|
225
|
+
|
|
226
|
+
Args:
|
|
227
|
+
attachment (Attachment | list[Attachment]): The attachment or a list of attachments to add.
|
|
228
|
+
"""
|
|
229
|
+
def add_tool_call(self, tool_call: ToolCall | list[ToolCall]) -> None:
|
|
230
|
+
"""Add a tool call or a list of tool calls to the LMOutput.
|
|
231
|
+
|
|
232
|
+
Args:
|
|
233
|
+
tool_call (ToolCall | list[ToolCall]): The tool call or a list of tool calls to add.
|
|
234
|
+
"""
|
|
235
|
+
def add_structured(self, structured: dict[str, Any] | BaseModel | list[dict[str, Any] | BaseModel]) -> None:
|
|
236
|
+
"""Add a structured output or a list of structured outputs to the LMOutput.
|
|
237
|
+
|
|
238
|
+
Args:
|
|
239
|
+
structured (dict[str, Any] | BaseModel | list[dict[str, Any] | BaseModel]): The structured output
|
|
240
|
+
or a list of structured outputs to add.
|
|
241
|
+
"""
|
|
242
|
+
def add_thinking(self, thinking: Reasoning | list[Reasoning]) -> None:
|
|
243
|
+
"""Add a thinking or a list of thoughts to the LMOutput.
|
|
244
|
+
|
|
245
|
+
Args:
|
|
246
|
+
thinking (Reasoning | list[Reasoning]): The thinking or a list of thoughts to add.
|
|
247
|
+
"""
|
|
248
|
+
def add_citation(self, citation: Chunk | list[Chunk]) -> None:
|
|
249
|
+
"""Add a citation or a list of citations to the LMOutput.
|
|
250
|
+
|
|
251
|
+
Args:
|
|
252
|
+
citation (Chunk | list[Chunk]): The citation or a list of citations to add.
|
|
253
|
+
"""
|
|
254
|
+
def add_code_exec_result(self, code_exec_result: CodeExecResult | list[CodeExecResult]) -> None:
|
|
255
|
+
"""Add a code exec result or a list of code exec results to the LMOutput.
|
|
256
|
+
|
|
257
|
+
Args:
|
|
258
|
+
code_exec_result (CodeExecResult | list[CodeExecResult]): The code exec result or a list of code exec
|
|
259
|
+
results to add.
|
|
260
|
+
"""
|
|
261
|
+
def add_mcp_call(self, mcp_call: MCPCall | list[MCPCall]) -> None:
|
|
262
|
+
"""Add an MCP call or a list of MCP calls to the LMOutput.
|
|
263
|
+
|
|
264
|
+
Args:
|
|
265
|
+
mcp_call (MCPCall | list[MCPCall]): The MCP call or a list of MCP calls to add.
|
|
266
|
+
"""
|
|
@@ -1,6 +1,5 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from enum import StrEnum
|
|
3
|
-
from gllm_inference.utils import validate_string_enum as validate_string_enum
|
|
4
3
|
from pydantic import BaseModel
|
|
5
4
|
|
|
6
5
|
PROVIDER_SEPARATOR: str
|
|
@@ -22,6 +21,7 @@ class ModelProvider(StrEnum):
|
|
|
22
21
|
PORTKEY = 'portkey'
|
|
23
22
|
OPENAI_CHAT_COMPLETIONS = 'openai-chat-completions'
|
|
24
23
|
OPENAI_COMPATIBLE = 'openai-compatible'
|
|
24
|
+
SEA_LION = 'sea-lion'
|
|
25
25
|
TWELVELABS = 'twelvelabs'
|
|
26
26
|
VOYAGE = 'voyage'
|
|
27
27
|
XAI = 'xai'
|
|
Binary file
|
gllm_inference.pyi
CHANGED
|
@@ -33,6 +33,7 @@ import gllm_inference.lm_invoker.OpenAIChatCompletionsLMInvoker
|
|
|
33
33
|
import gllm_inference.lm_invoker.OpenAICompatibleLMInvoker
|
|
34
34
|
import gllm_inference.lm_invoker.OpenAILMInvoker
|
|
35
35
|
import gllm_inference.lm_invoker.PortkeyLMInvoker
|
|
36
|
+
import gllm_inference.lm_invoker.SeaLionLMInvoker
|
|
36
37
|
import gllm_inference.lm_invoker.XAILMInvoker
|
|
37
38
|
import gllm_inference.prompt_builder.PromptBuilder
|
|
38
39
|
import gllm_inference.output_parser.JSONOutputParser
|
|
@@ -57,7 +58,6 @@ import gllm_inference.schema.AttachmentType
|
|
|
57
58
|
import gllm_inference.schema.EMContent
|
|
58
59
|
import gllm_inference.schema.Vector
|
|
59
60
|
import aioboto3
|
|
60
|
-
import gllm_inference.utils.validate_string_enum
|
|
61
61
|
import cohere
|
|
62
62
|
import asyncio.CancelledError
|
|
63
63
|
import gllm_inference.exceptions.convert_to_base_invoker_error
|
|
@@ -68,9 +68,7 @@ import google.genai
|
|
|
68
68
|
import google.genai.types
|
|
69
69
|
import httpx
|
|
70
70
|
import gllm_inference.exceptions.ProviderInternalError
|
|
71
|
-
import
|
|
72
|
-
import concurrent.futures
|
|
73
|
-
import concurrent.futures.ThreadPoolExecutor
|
|
71
|
+
import gllm_core.utils.concurrency
|
|
74
72
|
import langchain_core
|
|
75
73
|
import langchain_core.embeddings
|
|
76
74
|
import gllm_inference.exceptions.InvokerRuntimeError
|
|
@@ -85,6 +83,7 @@ import voyageai
|
|
|
85
83
|
import voyageai.client_async
|
|
86
84
|
import http
|
|
87
85
|
import http.HTTPStatus
|
|
86
|
+
import __future__
|
|
88
87
|
import uuid
|
|
89
88
|
import gllm_core.constants
|
|
90
89
|
import gllm_core.event
|
|
@@ -115,9 +114,7 @@ import inspect
|
|
|
115
114
|
import time
|
|
116
115
|
import jsonschema
|
|
117
116
|
import gllm_inference.lm_invoker.batch.BatchOperations
|
|
118
|
-
import gllm_inference.schema.LMEventType
|
|
119
117
|
import gllm_inference.schema.MessageContent
|
|
120
|
-
import __future__
|
|
121
118
|
import gllm_inference.schema.ActivityEvent
|
|
122
119
|
import gllm_inference.schema.CodeEvent
|
|
123
120
|
import gllm_inference.schema.CodeExecResult
|
|
@@ -139,6 +136,7 @@ import jinja2.sandbox
|
|
|
139
136
|
import gllm_inference.schema.JinjaEnvType
|
|
140
137
|
import gllm_inference.prompt_builder.format_strategy.JinjaFormatStrategy
|
|
141
138
|
import gllm_inference.prompt_builder.format_strategy.StringFormatStrategy
|
|
139
|
+
import gllm_inference.schema.HistoryFormatter
|
|
142
140
|
import transformers
|
|
143
141
|
import gllm_inference.prompt_formatter.HuggingFacePromptFormatter
|
|
144
142
|
import traceback
|
|
@@ -151,6 +149,7 @@ import pathlib
|
|
|
151
149
|
import filetype
|
|
152
150
|
import magic
|
|
153
151
|
import requests
|
|
152
|
+
import gllm_core.schema.chunk
|
|
154
153
|
import binascii
|
|
155
154
|
import fnmatch
|
|
156
155
|
import importlib
|
|
@@ -1,12 +1,12 @@
|
|
|
1
1
|
Metadata-Version: 2.2
|
|
2
2
|
Name: gllm-inference-binary
|
|
3
|
-
Version: 0.5.
|
|
3
|
+
Version: 0.5.66
|
|
4
4
|
Summary: A library containing components related to model inferences in Gen AI applications.
|
|
5
|
-
Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>,
|
|
5
|
+
Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, "Delfia N. A. Putri" <delfia.n.a.putri@gdplabs.id>
|
|
6
6
|
Requires-Python: <3.14,>=3.11
|
|
7
7
|
Description-Content-Type: text/markdown
|
|
8
8
|
Requires-Dist: poetry<3.0.0,>=2.1.3
|
|
9
|
-
Requires-Dist: gllm-core-binary<0.4.0,>=0.3.
|
|
9
|
+
Requires-Dist: gllm-core-binary<0.4.0,>=0.3.23
|
|
10
10
|
Requires-Dist: aiohttp<4.0.0,>=3.12.14
|
|
11
11
|
Requires-Dist: filetype<2.0.0,>=1.2.0
|
|
12
12
|
Requires-Dist: httpx<0.29.0,>=0.28.0
|
|
@@ -33,7 +33,7 @@ Requires-Dist: aioboto3<16.0.0,>=15.0.0; extra == "bedrock"
|
|
|
33
33
|
Provides-Extra: cohere
|
|
34
34
|
Requires-Dist: cohere<6.0.0,>=5.18.0; extra == "cohere"
|
|
35
35
|
Provides-Extra: datasaur
|
|
36
|
-
Requires-Dist: openai<
|
|
36
|
+
Requires-Dist: openai<3.0.0,>=2.7.0; extra == "datasaur"
|
|
37
37
|
Provides-Extra: google
|
|
38
38
|
Requires-Dist: google-genai<=1.36,>=1.23; extra == "google"
|
|
39
39
|
Provides-Extra: huggingface
|
|
@@ -42,7 +42,7 @@ Requires-Dist: transformers<5.0.0,>=4.52.0; extra == "huggingface"
|
|
|
42
42
|
Provides-Extra: litellm
|
|
43
43
|
Requires-Dist: litellm<2.0.0,>=1.69.2; extra == "litellm"
|
|
44
44
|
Provides-Extra: openai
|
|
45
|
-
Requires-Dist: openai<
|
|
45
|
+
Requires-Dist: openai<3.0.0,>=2.7.0; extra == "openai"
|
|
46
46
|
Provides-Extra: portkey-ai
|
|
47
47
|
Requires-Dist: portkey-ai<2.0.0,>=1.14.4; extra == "portkey-ai"
|
|
48
48
|
Provides-Extra: twelvelabs
|
|
@@ -1,11 +1,11 @@
|
|
|
1
|
-
gllm_inference.cp311-win_amd64.pyd,sha256=
|
|
2
|
-
gllm_inference.pyi,sha256=
|
|
1
|
+
gllm_inference.cp311-win_amd64.pyd,sha256=xik0kaZaQfmy3izrWe_LGbuuz8tIpFtVAtMehjg2feI,4151296
|
|
2
|
+
gllm_inference.pyi,sha256=jCxjKwYXb3t6yZrf6m7uD29V9KPXyWb8U04-lru4-JA,5173
|
|
3
3
|
gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
4
|
-
gllm_inference/constants.pyi,sha256=
|
|
4
|
+
gllm_inference/constants.pyi,sha256=8jIYOyxJYVWUYXSXF3vag9HhHwjq1iU9tzPiosRHkWk,328
|
|
5
5
|
gllm_inference/builder/__init__.pyi,sha256=-bw1uDx7CAM7pkvjvb1ZXku9zXlQ7aEAyC83KIn3bz8,506
|
|
6
6
|
gllm_inference/builder/_build_invoker.pyi,sha256=SZst1ocnRntcA7lWKp44lMJ3yRA5GP_jsUZkcsxwTjw,876
|
|
7
7
|
gllm_inference/builder/build_em_invoker.pyi,sha256=Z09hP6_ziumI2XB-7HCdyIE4AyoV1oL9DBdoU2G8izo,5879
|
|
8
|
-
gllm_inference/builder/build_lm_invoker.pyi,sha256=
|
|
8
|
+
gllm_inference/builder/build_lm_invoker.pyi,sha256=ZKNVOOQAi1DaChrPBdWGUuzPShhmkGx89oZnALhR9w0,9417
|
|
9
9
|
gllm_inference/builder/build_lm_request_processor.pyi,sha256=H7Rg88e7PTTCtuyY64r333moTmh4-ypOwgnG10gkEdY,4232
|
|
10
10
|
gllm_inference/builder/build_output_parser.pyi,sha256=sgSTrzUmSRxPzUUum0fDU7A3NXYoYhpi6bEx4Q2XMnA,965
|
|
11
11
|
gllm_inference/catalog/__init__.pyi,sha256=HWgPKWIzprpMHRKe_qN9BZSIQhVhrqiyjLjIXwvj1ho,291
|
|
@@ -15,7 +15,7 @@ gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=iViWB4SaezzjQY4UY1YxeoX
|
|
|
15
15
|
gllm_inference/em_invoker/__init__.pyi,sha256=uCWfCjh5a5DciRFcUdbHndewokM3J5hp3mbhmM5wQC8,1211
|
|
16
16
|
gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=TXC5Kgf1eZqK2FHKAyeG3LB1SEsSEStnbk9bI1mjC5k,5049
|
|
17
17
|
gllm_inference/em_invoker/bedrock_em_invoker.pyi,sha256=ZPse0qNR74qFcc-9svTN8DzVO6kCLz-KYsLDo3cEEd0,6483
|
|
18
|
-
gllm_inference/em_invoker/cohere_em_invoker.pyi,sha256=
|
|
18
|
+
gllm_inference/em_invoker/cohere_em_invoker.pyi,sha256=Z_eBDfq9XErzpO2q47rfSB8LWxZY7Wxx1TTlvVh6k28,6708
|
|
19
19
|
gllm_inference/em_invoker/em_invoker.pyi,sha256=YDYJ8TGScsz5Gg-OBnEENN1tI1RYvwoddypxUr6SAWw,5191
|
|
20
20
|
gllm_inference/em_invoker/google_em_invoker.pyi,sha256=zZYjeLp9ncwIVM4UHqDJSVOFn1eXiaz9Ba24-_fCF2c,6953
|
|
21
21
|
gllm_inference/em_invoker/jina_em_invoker.pyi,sha256=wrivFiPFRWBSw-TEAkaQKBvBRhlzN6CBaZkc0HOslRw,5754
|
|
@@ -40,33 +40,34 @@ gllm_inference/exceptions/__init__.pyi,sha256=nXOqwsuwUgsnBcJEANVuxbZ1nDfcJ6-pKU
|
|
|
40
40
|
gllm_inference/exceptions/error_parser.pyi,sha256=4aiJZhBzBOqlhdmpvaCvildGy7_XxlJzQpe3PzGt8eE,2040
|
|
41
41
|
gllm_inference/exceptions/exceptions.pyi,sha256=6y3ECgHAStqMGgQv8Dv-Ui-5PDD07mSj6qaRZeSWea4,5857
|
|
42
42
|
gllm_inference/exceptions/provider_error_map.pyi,sha256=vWa4ZIHn7qIghECGvO-dS2KzOmf3c10GRWKZ4YDPnSQ,1267
|
|
43
|
-
gllm_inference/lm_invoker/__init__.pyi,sha256=
|
|
44
|
-
gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=
|
|
45
|
-
gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=
|
|
46
|
-
gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=
|
|
47
|
-
gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=
|
|
48
|
-
gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=
|
|
49
|
-
gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=
|
|
50
|
-
gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=
|
|
51
|
-
gllm_inference/lm_invoker/lm_invoker.pyi,sha256=
|
|
52
|
-
gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=
|
|
43
|
+
gllm_inference/lm_invoker/__init__.pyi,sha256=Y0laMHRdUwGgfIXo278qe1dwj7ZMSCeO-wMToYim47c,1618
|
|
44
|
+
gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=b3VM_mcK20y4Fyo_vWDx95HA001l9h53ACo0caP1sTY,15729
|
|
45
|
+
gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=uV98H2nJsElCTsxAuInZ9KSk1jOTq6SROAGQRPR-_r0,13173
|
|
46
|
+
gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=qXmFK6zsOM3nPfueEhY5pAfG24bZytA1jqemPa63vLY,10951
|
|
47
|
+
gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=FnpayOW_Zi0pWFSawLX8XahEnknbnpsRWrkhKZe8Y3U,8035
|
|
48
|
+
gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=2ZqQl9WhehtxcnAXzUi7vFDaHV0lO-0SwHoAT7m8Lf0,20484
|
|
49
|
+
gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=ull3cX-iUT4hYMbixcxqfrNUxR8ZoR4Vt9ACVILQWSM,12126
|
|
50
|
+
gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=qG8pPTiDJZR2e7wr5Q2VyceC227tz3QybX3UPihT5ng,11400
|
|
51
|
+
gllm_inference/lm_invoker/lm_invoker.pyi,sha256=TqlnNhWnFeteC_JjnnQ1pPW66qHPePC0iA738I1LdXo,9553
|
|
52
|
+
gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=qt9DAdJM7YBB4op-6SOJB0kCouPYVxtIamGUXLGLUeA,13888
|
|
53
53
|
gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=T9sShA_9fgEuaaAuT2gJZq_EYNbEhf3IkWwMCwfszY8,4244
|
|
54
|
-
gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=
|
|
55
|
-
gllm_inference/lm_invoker/portkey_lm_invoker.pyi,sha256=
|
|
56
|
-
gllm_inference/lm_invoker/
|
|
54
|
+
gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=Wu9cvkncnfMQI1Fht5c3i3LF0Tdd6Pz0-rpCTkWRe84,24918
|
|
55
|
+
gllm_inference/lm_invoker/portkey_lm_invoker.pyi,sha256=FYOp4BaDfOtompWIRhDqzMVVSK-TiFyw7JA4TznANQE,15236
|
|
56
|
+
gllm_inference/lm_invoker/sea_lion_lm_invoker.pyi,sha256=Qqplj79JsnLGY6xjhbXYbz6Ramxur4sXfvv4svk19os,3533
|
|
57
|
+
gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=6beZsQjGUTo7TdzWBWksRzVGT58XyipErpGfiRq6NH0,13017
|
|
57
58
|
gllm_inference/lm_invoker/batch/__init__.pyi,sha256=vJOTHRJ83oq8Bq0UsMdID9_HW5JAxr06gUs4aPRZfEE,130
|
|
58
|
-
gllm_inference/lm_invoker/batch/batch_operations.pyi,sha256=
|
|
59
|
+
gllm_inference/lm_invoker/batch/batch_operations.pyi,sha256=EKwINY8DQFOXquGGrUTr_Yg24ZrSyiiAo2hX8If9vyQ,5573
|
|
59
60
|
gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
60
61
|
gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=Ol_IqeBPgrmG5iaY3NV6AD-alC2HmUvljp0vI8GApWo,1154
|
|
61
62
|
gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=rB1AWfER2BBKZ5I219211YE2EUFPF25bhzysqjdPgiY,1080
|
|
62
63
|
gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=8lmb1PRbkqBsF_l7iOffxW0K5Xxpi69GW9Z7KxyxHTE,279
|
|
63
|
-
gllm_inference/lm_invoker/schema/google.pyi,sha256=
|
|
64
|
+
gllm_inference/lm_invoker/schema/google.pyi,sha256=LQ14PJyDOe3K5TYvE-gzE1fjpZCSAy-0Sy9Lmw6fICY,827
|
|
64
65
|
gllm_inference/lm_invoker/schema/langchain.pyi,sha256=2OJOUQPlGdlUbIOTDOyiWDBOMm3MoVX-kU2nK0zQsF0,452
|
|
65
|
-
gllm_inference/lm_invoker/schema/openai.pyi,sha256=
|
|
66
|
+
gllm_inference/lm_invoker/schema/openai.pyi,sha256=yrL10gyrxvU3szXeKJC-mMkHO3nzdcaYDsBzORqK3AE,2777
|
|
66
67
|
gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=nNPb7ETC9IrJwkV5wfbGf6Co3-qdq4lhcXz0l_qYCE4,1261
|
|
67
68
|
gllm_inference/lm_invoker/schema/portkey.pyi,sha256=V2q4JIwDAR7BidqfmO01u1_1mLOMtm5OCon6sN2zNt0,662
|
|
68
69
|
gllm_inference/lm_invoker/schema/xai.pyi,sha256=jpC6ZSBDUltzm9GjD6zvSFIPwqizn_ywLnjvwSa7KuU,663
|
|
69
|
-
gllm_inference/model/__init__.pyi,sha256=
|
|
70
|
+
gllm_inference/model/__init__.pyi,sha256=J3z0Rc6PI_aC0XuA_dWB2tCBQ4pSZdS5NzQ8vPhs5c4,907
|
|
70
71
|
gllm_inference/model/em/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
71
72
|
gllm_inference/model/em/cohere_em.pyi,sha256=uF1AmDO-skQteYqzxJ3DK10SqgfdW0oW9L8Ym34eU04,505
|
|
72
73
|
gllm_inference/model/em/google_em.pyi,sha256=c53H-KNdNOK9ppPLyOSkmCA890eF5FsMd05upkPIzF0,487
|
|
@@ -78,12 +79,13 @@ gllm_inference/model/lm/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZ
|
|
|
78
79
|
gllm_inference/model/lm/anthropic_lm.pyi,sha256=ccUpxddakurLFHivl5UzJxgODLhcFgx8XC7CKa-99NE,633
|
|
79
80
|
gllm_inference/model/lm/google_lm.pyi,sha256=OLuoqT0FnJOLsNalulBMEXuCYAXoF8Y7vjfSBgjaJxA,529
|
|
80
81
|
gllm_inference/model/lm/openai_lm.pyi,sha256=yj3AJj1xDYRkNIPHX2enw46AJ9wArPZruKsxg1ME9Rg,645
|
|
82
|
+
gllm_inference/model/lm/sea_lion_lm.pyi,sha256=Gz42ym_KJA7if3sDCFmdrb1uv9b9__n2RWOfPvOxm0Y,508
|
|
81
83
|
gllm_inference/model/lm/xai_lm.pyi,sha256=O3G9Lj1Ii31CyCDrwYVkPPJN6X8V-WBF9xILUPUE-qY,525
|
|
82
84
|
gllm_inference/output_parser/__init__.pyi,sha256=dhAeRTBxc6CfS8bhnHjbtrnyqJ1iyffvUZkGp4UrJNM,132
|
|
83
85
|
gllm_inference/output_parser/json_output_parser.pyi,sha256=YtgQh8Uzy8W_Tgh8DfuR7VFFS7qvLEasiTwRfaGZZEU,2993
|
|
84
86
|
gllm_inference/output_parser/output_parser.pyi,sha256=-Xu5onKCBDqShcO-VrQh5icqAmXdihGc3rkZxL93swg,975
|
|
85
87
|
gllm_inference/prompt_builder/__init__.pyi,sha256=mPsbiafzSNHsgN-CuzjhgZpfXfi1pPC3_gdsq2p0EM4,120
|
|
86
|
-
gllm_inference/prompt_builder/prompt_builder.pyi,sha256=
|
|
88
|
+
gllm_inference/prompt_builder/prompt_builder.pyi,sha256=w99E5dsQxpJ-1FNEF_WBOt7xovmhp5WzrbTfRQ76K60,5083
|
|
87
89
|
gllm_inference/prompt_builder/format_strategy/__init__.pyi,sha256=BliNAdrN2yNoQt8muLe1IknuE78UDQys_3L_8_2Tn9E,312
|
|
88
90
|
gllm_inference/prompt_builder/format_strategy/format_strategy.pyi,sha256=Uay07qR0nMGbrOZ2twEyrIWYRKevBQL9ju_4fVfByyQ,2328
|
|
89
91
|
gllm_inference/prompt_builder/format_strategy/jinja_format_strategy.pyi,sha256=W8SdOPWPz_U355G8rQwiIQ0Ys78vomUBFEEZiz3AS5k,2270
|
|
@@ -109,18 +111,19 @@ gllm_inference/realtime_chat/output_streamer/output_streamer.pyi,sha256=5P9NQ0aJ
|
|
|
109
111
|
gllm_inference/request_processor/__init__.pyi,sha256=giEme2WFQhgyKiBZHhSet0_nKSCHwGy-_2p6NRzg0Zc,231
|
|
110
112
|
gllm_inference/request_processor/lm_request_processor.pyi,sha256=0fy1HyILCVDw6y46E-7tLnQTRYx4ppeRMe0QP6t9Jyw,5990
|
|
111
113
|
gllm_inference/request_processor/uses_lm_mixin.pyi,sha256=LYHq-zLoXEMel1LfVdYv7W3BZ8WtBLo_WWFjRf10Yto,6512
|
|
112
|
-
gllm_inference/schema/__init__.pyi,sha256=
|
|
114
|
+
gllm_inference/schema/__init__.pyi,sha256=89GWKdJ6DdPUTem42_Uc-5EsyNcfUJ-WnMVEkagpEwI,2524
|
|
113
115
|
gllm_inference/schema/activity.pyi,sha256=atrU4OwLesA9FEt1H7K3gsUWYNdOqpI5i2VdWkmo6cs,2367
|
|
114
|
-
gllm_inference/schema/attachment.pyi,sha256=
|
|
116
|
+
gllm_inference/schema/attachment.pyi,sha256=myJ_cI_h5mwUdvmMrWpSQIwj3nIxe8SD7HxO37o_3D4,4611
|
|
115
117
|
gllm_inference/schema/code_exec_result.pyi,sha256=WQ-ARoGM9r6nyRX-A0Ro1XKiqrc9R3jRYXZpu_xo5S4,573
|
|
116
118
|
gllm_inference/schema/config.pyi,sha256=NVmjQK6HipIE0dKSfx12hgIC0O-S1HEcAc-TWlXAF5A,689
|
|
117
|
-
gllm_inference/schema/enums.pyi,sha256=
|
|
118
|
-
gllm_inference/schema/events.pyi,sha256=
|
|
119
|
+
gllm_inference/schema/enums.pyi,sha256=U-lADk7xNjA5rByem0fo0xDAy9T3GsekY8B5szTlWPQ,2193
|
|
120
|
+
gllm_inference/schema/events.pyi,sha256=_CKuGNzb3j2Y1dOB2yssFkT_9FQz1AY4J_ApCwKNizU,4743
|
|
121
|
+
gllm_inference/schema/formatter.pyi,sha256=uJ1CZt_DmizfCPHD5deQwm2ftU-zcl2aj5rwJGTPe9c,1312
|
|
119
122
|
gllm_inference/schema/lm_input.pyi,sha256=HxQiZgY7zcXh_Dw8nK8LSeBTZEHMPZVwmPmnfgSsAbs,197
|
|
120
|
-
gllm_inference/schema/lm_output.pyi,sha256=
|
|
123
|
+
gllm_inference/schema/lm_output.pyi,sha256=N75CIF_2kZRdXKy2jvu9hhqzk5DrCbsHXTrhKqQ-7vo,11667
|
|
121
124
|
gllm_inference/schema/mcp.pyi,sha256=4SgQ83pEowfWm2p-w9lupV4NayqqVBOy7SuYxIFeWRs,1045
|
|
122
125
|
gllm_inference/schema/message.pyi,sha256=jJV6A0ihEcun2OhzyMtNkiHnf7d6v5R-GdpTBGfJ0AQ,2272
|
|
123
|
-
gllm_inference/schema/model_id.pyi,sha256=
|
|
126
|
+
gllm_inference/schema/model_id.pyi,sha256=JhTzjCsOzHGds_UF8PlPON3xtqmSjUEfvjpmYyf3CJ0,6018
|
|
124
127
|
gllm_inference/schema/reasoning.pyi,sha256=jbPxkDRHt0Vt-zdcc8lTT1l2hIE1Jm3HIHeNd0hfXGo,577
|
|
125
128
|
gllm_inference/schema/token_usage.pyi,sha256=WJiGQyz5qatzBK2b-sABLCyTRLCBbAvxCRcqSJOzu-8,3025
|
|
126
129
|
gllm_inference/schema/tool_call.pyi,sha256=OWT9LUqs_xfUcOkPG0aokAAqzLYYDkfnjTa0zOWvugk,403
|
|
@@ -129,9 +132,9 @@ gllm_inference/schema/type_alias.pyi,sha256=L-V0WxsFQznzAfby3DH8XMHUgjjZxQEsLw8S
|
|
|
129
132
|
gllm_inference/utils/__init__.pyi,sha256=H27RiiFjD6WQHRrYb1-sBnb2aqjVENw5_8-DdAe1k9A,396
|
|
130
133
|
gllm_inference/utils/io_utils.pyi,sha256=Eg7dvHWdXslTKdjh1j3dG50i7r35XG2zTmJ9XXvz4cI,1120
|
|
131
134
|
gllm_inference/utils/langchain.pyi,sha256=4AwFiVAO0ZpdgmqeC4Pb5NJwBt8vVr0MSUqLeCdTscc,1194
|
|
132
|
-
gllm_inference/utils/validation.pyi,sha256
|
|
135
|
+
gllm_inference/utils/validation.pyi,sha256=OWRZxeVGIuuvNU0LqLGB-9gNmypvbH-LcSJx91rnH1k,453
|
|
133
136
|
gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
|
|
134
|
-
gllm_inference_binary-0.5.
|
|
135
|
-
gllm_inference_binary-0.5.
|
|
136
|
-
gllm_inference_binary-0.5.
|
|
137
|
-
gllm_inference_binary-0.5.
|
|
137
|
+
gllm_inference_binary-0.5.66.dist-info/METADATA,sha256=JS0et7h2YvGkt2guUZO2aNaAm4rE77yALrYxMVHoKls,5953
|
|
138
|
+
gllm_inference_binary-0.5.66.dist-info/WHEEL,sha256=l2aKBREYfqJ7T2ljmr6hUiXPoNvvXF47bG4IHjuSyS4,96
|
|
139
|
+
gllm_inference_binary-0.5.66.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
|
|
140
|
+
gllm_inference_binary-0.5.66.dist-info/RECORD,,
|
|
File without changes
|
{gllm_inference_binary-0.5.51.dist-info → gllm_inference_binary-0.5.66.dist-info}/top_level.txt
RENAMED
|
File without changes
|