gllm-inference-binary 0.5.51__cp311-cp311-win_amd64.whl → 0.5.66__cp311-cp311-win_amd64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (37) hide show
  1. gllm_inference/builder/build_lm_invoker.pyi +10 -1
  2. gllm_inference/constants.pyi +0 -1
  3. gllm_inference/em_invoker/cohere_em_invoker.pyi +1 -2
  4. gllm_inference/lm_invoker/__init__.pyi +2 -1
  5. gllm_inference/lm_invoker/anthropic_lm_invoker.pyi +95 -109
  6. gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi +92 -109
  7. gllm_inference/lm_invoker/batch/batch_operations.pyi +2 -1
  8. gllm_inference/lm_invoker/bedrock_lm_invoker.pyi +51 -65
  9. gllm_inference/lm_invoker/datasaur_lm_invoker.pyi +36 -36
  10. gllm_inference/lm_invoker/google_lm_invoker.pyi +195 -110
  11. gllm_inference/lm_invoker/langchain_lm_invoker.pyi +52 -64
  12. gllm_inference/lm_invoker/litellm_lm_invoker.pyi +86 -106
  13. gllm_inference/lm_invoker/lm_invoker.pyi +20 -1
  14. gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi +87 -107
  15. gllm_inference/lm_invoker/openai_lm_invoker.pyi +235 -185
  16. gllm_inference/lm_invoker/portkey_lm_invoker.pyi +104 -68
  17. gllm_inference/lm_invoker/schema/google.pyi +12 -0
  18. gllm_inference/lm_invoker/schema/openai.pyi +22 -0
  19. gllm_inference/lm_invoker/sea_lion_lm_invoker.pyi +48 -0
  20. gllm_inference/lm_invoker/xai_lm_invoker.pyi +94 -131
  21. gllm_inference/model/__init__.pyi +2 -1
  22. gllm_inference/model/lm/sea_lion_lm.pyi +16 -0
  23. gllm_inference/prompt_builder/prompt_builder.pyi +6 -2
  24. gllm_inference/schema/__init__.pyi +4 -3
  25. gllm_inference/schema/attachment.pyi +20 -6
  26. gllm_inference/schema/enums.pyi +14 -1
  27. gllm_inference/schema/events.pyi +2 -2
  28. gllm_inference/schema/formatter.pyi +31 -0
  29. gllm_inference/schema/lm_output.pyi +245 -23
  30. gllm_inference/schema/model_id.pyi +1 -1
  31. gllm_inference/utils/validation.pyi +3 -0
  32. gllm_inference.cp311-win_amd64.pyd +0 -0
  33. gllm_inference.pyi +5 -6
  34. {gllm_inference_binary-0.5.51.dist-info → gllm_inference_binary-0.5.66.dist-info}/METADATA +5 -5
  35. {gllm_inference_binary-0.5.51.dist-info → gllm_inference_binary-0.5.66.dist-info}/RECORD +37 -34
  36. {gllm_inference_binary-0.5.51.dist-info → gllm_inference_binary-0.5.66.dist-info}/WHEEL +0 -0
  37. {gllm_inference_binary-0.5.51.dist-info → gllm_inference_binary-0.5.66.dist-info}/top_level.txt +0 -0
@@ -1,6 +1,8 @@
1
- from gllm_core.schema import Chunk as Chunk
1
+ from _typeshed import Incomplete
2
+ from gllm_core.schema import Chunk
2
3
  from gllm_inference.schema.attachment import Attachment as Attachment
3
4
  from gllm_inference.schema.code_exec_result import CodeExecResult as CodeExecResult
5
+ from gllm_inference.schema.enums import LMOutputType as LMOutputType
4
6
  from gllm_inference.schema.mcp import MCPCall as MCPCall
5
7
  from gllm_inference.schema.reasoning import Reasoning as Reasoning
6
8
  from gllm_inference.schema.token_usage import TokenUsage as TokenUsage
@@ -8,37 +10,257 @@ from gllm_inference.schema.tool_call import ToolCall as ToolCall
8
10
  from pydantic import BaseModel
9
11
  from typing import Any
10
12
 
13
+ LMOutputData = str | dict[str, Any] | BaseModel | Attachment | ToolCall | Reasoning | Chunk | CodeExecResult | MCPCall
14
+ logger: Incomplete
15
+
16
+ class LMOutputItem(BaseModel):
17
+ """Defines the output item of a language model.
18
+
19
+ Attributes:
20
+ type (str): The type of the output item.
21
+ output (LMOutputData): The output data of the output item.
22
+ """
23
+ type: str
24
+ output: LMOutputData
25
+
11
26
  class LMOutput(BaseModel):
12
27
  """Defines the output of a language model.
13
28
 
14
29
  Attributes:
15
- response (str): The text response. Defaults to an empty string.
16
- attachments (list[Attachment]): The attachments, if the language model decides to output attachments.
17
- Defaults to an empty list.
18
- tool_calls (list[ToolCall]): The tool calls, if the language model decides to invoke tools.
19
- Defaults to an empty list.
20
- structured_output (dict[str, Any] | BaseModel | None): The structured output, if a response schema is defined
21
- for the language model. Defaults to None.
30
+ outputs (list[LMOutputItem]): The outputs of the language model in sequential order. Defaults to an empty list.
22
31
  token_usage (TokenUsage | None): The token usage analytics, if requested. Defaults to None.
23
32
  duration (float | None): The duration of the invocation in seconds, if requested. Defaults to None.
24
33
  finish_details (dict[str, Any]): The details about how the generation finished, if requested.
25
34
  Defaults to an empty dictionary.
26
- reasoning (list[Reasoning]): The reasoning, if the language model is configured to output reasoning.
27
- Defaults to an empty list.
28
- citations (list[Chunk]): The citations, if the language model outputs citations. Defaults to an empty list.
29
- code_exec_results (list[CodeExecResult]): The code execution results, if the language model decides to
30
- execute code. Defaults to an empty list.
31
- mcp_calls (list[MCPCall]): The MCP calls, if the language model decides to invoke MCP tools.
32
- Defaults to an empty list.
35
+
36
+ text (str): The first text response.
37
+ structured_output (dict[str, Any] | BaseModel | None): The first structured output.
38
+
39
+ texts (list[str]): The texts from the outputs.
40
+ structured_outputs (list[dict[str, Any] | BaseModel]): The structured outputs from the outputs.
41
+ attachments (list[Attachment]): The attachments from the outputs.
42
+ tool_calls (list[ToolCall]): The tool calls from the outputs.
43
+ thinkings (list[Reasoning]): The thinkings from the outputs.
44
+ citations (list[Chunk]): The citations from the outputs.
45
+ code_exec_results (list[CodeExecResult]): The code exec results from the outputs.
46
+ mcp_calls (list[MCPCall]): The MCP calls from the outputs.
47
+
48
+ response (str): Deprecated. Replaced by `text`.
49
+ reasoning (list[Reasoning]): Deprecated. Replaced by `thinkings`.
33
50
  """
34
- response: str
35
- attachments: list[Attachment]
36
- tool_calls: list[ToolCall]
37
- structured_output: dict[str, Any] | BaseModel | None
51
+ outputs: list[LMOutputItem]
38
52
  token_usage: TokenUsage | None
39
53
  duration: float | None
40
54
  finish_details: dict[str, Any]
41
- reasoning: list[Reasoning]
42
- citations: list[Chunk]
43
- code_exec_results: list[CodeExecResult]
44
- mcp_calls: list[MCPCall]
55
+ def __init__(self, *, outputs: list[LMOutputItem] | None = None, token_usage: TokenUsage | None = None, duration: float | None = None, finish_details: dict[str, Any] | None = None, response: str = '', structured_output: dict[str, Any] | BaseModel | None = None, tool_calls: list[ToolCall] | None = None, reasoning: list[Reasoning] | None = None, attachments: list[Attachment] | None = None, citations: list[Chunk] | None = None, code_exec_results: list[CodeExecResult] | None = None, mcp_calls: list[MCPCall] | None = None) -> None:
56
+ """Initialize the LMOutput.
57
+
58
+ This constructor is created for backward compatibility with the legacy method to initialize the LMOutput.
59
+ This constructor will be removed in v0.6.
60
+
61
+ Args:
62
+ outputs (list[LMOutputItem] | None, optional): The output items. Defaults to an empty list.
63
+ token_usage (TokenUsage | None, optional): The token usage analytics. Defaults to None.
64
+ duration (float | None, optional): The duration of the invocation in seconds. Defaults to None.
65
+ finish_details (dict[str, Any] | None, optional): The details about how the generation finished.
66
+ Defaults to an empty dictionary.
67
+ response (str, optional): The first text response. Defaults to an empty string.
68
+ structured_output (dict[str, Any] | BaseModel | None, optional): The first structured output.
69
+ Defaults to None.
70
+ tool_calls (list[ToolCall] | None, optional): The tool calls. Defaults to None.
71
+ reasoning (list[Reasoning] | None, optional): The thinkings. Defaults to None.
72
+ attachments (list[Attachment] | None, optional): The attachments. Defaults to None.
73
+ citations (list[Chunk] | None, optional): The citations. Defaults to None.
74
+ code_exec_results (list[CodeExecResult] | None, optional): The code exec results. Defaults to None.
75
+ mcp_calls (list[MCPCall] | None, optional): The MCP calls. Defaults to None.
76
+ """
77
+ @property
78
+ def response(self) -> str:
79
+ """Deprecated property to get the first text response from the LMOutput.
80
+
81
+ Returns:
82
+ str: The first text response from the LMOutput.
83
+ """
84
+ @response.setter
85
+ def response(self, value: str) -> None:
86
+ """Deprecated setter to set the first text response to the LMOutput.
87
+
88
+ Args:
89
+ value (str): The first text response to set.
90
+ """
91
+ @property
92
+ def text(self) -> str:
93
+ """Get the first text from the LMOutput.
94
+
95
+ Returns:
96
+ str: The first text from the LMOutput.
97
+ """
98
+ @property
99
+ def structured_output(self) -> dict[str, Any] | BaseModel | None:
100
+ """Deprecated property to get the first structured output from the LMOutput.
101
+
102
+ Returns:
103
+ dict[str, Any] | BaseModel | None: The first structured output from the LMOutput.
104
+ """
105
+ @structured_output.setter
106
+ def structured_output(self, value: dict[str, Any] | BaseModel) -> None:
107
+ """Deprecated setter to set the first structured output to the LMOutput.
108
+
109
+ Args:
110
+ value (dict[str, Any] | BaseModel): The first structured output to set.
111
+ """
112
+ @property
113
+ def texts(self) -> list[str]:
114
+ """Get the texts from the LMOutput.
115
+
116
+ Returns:
117
+ list[str]: The texts from the LMOutput.
118
+ """
119
+ @property
120
+ def structured_outputs(self) -> list[dict[str, Any] | BaseModel]:
121
+ """Get the structured outputs from the LMOutput.
122
+
123
+ Returns:
124
+ list[dict[str, Any] | BaseModel]: The structured outputs from the LMOutput.
125
+ """
126
+ @property
127
+ def attachments(self) -> list[Attachment]:
128
+ """Get the attachments from the LMOutput.
129
+
130
+ Returns:
131
+ list[Attachment]: The attachments from the LMOutput.
132
+ """
133
+ @attachments.setter
134
+ def attachments(self, value: list[Attachment]) -> None:
135
+ """Deprecated setter to set the attachments to the LMOutput.
136
+
137
+ Args:
138
+ value (list[Attachment]): The attachments to set.
139
+ """
140
+ @property
141
+ def tool_calls(self) -> list[ToolCall]:
142
+ """Get the tool calls from the LMOutput.
143
+
144
+ Returns:
145
+ list[ToolCall]: The tool calls from the LMOutput.
146
+ """
147
+ @tool_calls.setter
148
+ def tool_calls(self, value: list[ToolCall]) -> None:
149
+ """Deprecated setter to set the tool calls to the LMOutput.
150
+
151
+ Args:
152
+ value (list[ToolCall]): The tool calls to set.
153
+ """
154
+ @property
155
+ def reasoning(self) -> list[Reasoning]:
156
+ """Deprecated property to get the thinkings from the LMOutput.
157
+
158
+ Returns:
159
+ list[Reasoning]: The thinkings from the LMOutput.
160
+ """
161
+ @reasoning.setter
162
+ def reasoning(self, value: list[Reasoning]) -> None:
163
+ """Deprecated setter to set the thinkings to the LMOutput.
164
+
165
+ Args:
166
+ value (list[Reasoning]): The thinkings to set.
167
+ """
168
+ @property
169
+ def thinkings(self) -> list[Reasoning]:
170
+ """Get the thinkings from the LMOutput.
171
+
172
+ Returns:
173
+ list[Reasoning]: The thinkings from the LMOutput.
174
+ """
175
+ @property
176
+ def citations(self) -> list[Chunk]:
177
+ """Get the citations from the LMOutput.
178
+
179
+ Returns:
180
+ list[Chunk]: The citations from the LMOutput.
181
+ """
182
+ @citations.setter
183
+ def citations(self, value: list[Chunk]) -> None:
184
+ """Deprecated setter to set the citations to the LMOutput.
185
+
186
+ Args:
187
+ value (list[Chunk]): The citations to set.
188
+ """
189
+ @property
190
+ def code_exec_results(self) -> list[CodeExecResult]:
191
+ """Get the code exec results from the LMOutput.
192
+
193
+ Returns:
194
+ list[CodeExecResult]: The code exec results from the LMOutput.
195
+ """
196
+ @code_exec_results.setter
197
+ def code_exec_results(self, value: list[CodeExecResult]) -> None:
198
+ """Deprecated setter to set the code exec results to the LMOutput.
199
+
200
+ Args:
201
+ value (list[CodeExecResult]): The code exec results to set.
202
+ """
203
+ @property
204
+ def mcp_calls(self) -> list[MCPCall]:
205
+ """Get the MCP calls from the LMOutput.
206
+
207
+ Returns:
208
+ list[MCPCall]: The MCP calls from the LMOutput.
209
+ """
210
+ @mcp_calls.setter
211
+ def mcp_calls(self, value: list[MCPCall]) -> None:
212
+ """Deprecated setter to set the MCP calls to the LMOutput.
213
+
214
+ Args:
215
+ value (list[MCPCall]): The MCP calls to set.
216
+ """
217
+ def add_text(self, text: str | list[str]) -> None:
218
+ """Add an output or a list of outputs to the LMOutput.
219
+
220
+ Args:
221
+ text (str | list[str]): The text or a list of texts to add.
222
+ """
223
+ def add_attachment(self, attachment: Attachment | list[Attachment]) -> None:
224
+ """Add an attachment or a list of attachments to the LMOutput.
225
+
226
+ Args:
227
+ attachment (Attachment | list[Attachment]): The attachment or a list of attachments to add.
228
+ """
229
+ def add_tool_call(self, tool_call: ToolCall | list[ToolCall]) -> None:
230
+ """Add a tool call or a list of tool calls to the LMOutput.
231
+
232
+ Args:
233
+ tool_call (ToolCall | list[ToolCall]): The tool call or a list of tool calls to add.
234
+ """
235
+ def add_structured(self, structured: dict[str, Any] | BaseModel | list[dict[str, Any] | BaseModel]) -> None:
236
+ """Add a structured output or a list of structured outputs to the LMOutput.
237
+
238
+ Args:
239
+ structured (dict[str, Any] | BaseModel | list[dict[str, Any] | BaseModel]): The structured output
240
+ or a list of structured outputs to add.
241
+ """
242
+ def add_thinking(self, thinking: Reasoning | list[Reasoning]) -> None:
243
+ """Add a thinking or a list of thoughts to the LMOutput.
244
+
245
+ Args:
246
+ thinking (Reasoning | list[Reasoning]): The thinking or a list of thoughts to add.
247
+ """
248
+ def add_citation(self, citation: Chunk | list[Chunk]) -> None:
249
+ """Add a citation or a list of citations to the LMOutput.
250
+
251
+ Args:
252
+ citation (Chunk | list[Chunk]): The citation or a list of citations to add.
253
+ """
254
+ def add_code_exec_result(self, code_exec_result: CodeExecResult | list[CodeExecResult]) -> None:
255
+ """Add a code exec result or a list of code exec results to the LMOutput.
256
+
257
+ Args:
258
+ code_exec_result (CodeExecResult | list[CodeExecResult]): The code exec result or a list of code exec
259
+ results to add.
260
+ """
261
+ def add_mcp_call(self, mcp_call: MCPCall | list[MCPCall]) -> None:
262
+ """Add an MCP call or a list of MCP calls to the LMOutput.
263
+
264
+ Args:
265
+ mcp_call (MCPCall | list[MCPCall]): The MCP call or a list of MCP calls to add.
266
+ """
@@ -1,6 +1,5 @@
1
1
  from _typeshed import Incomplete
2
2
  from enum import StrEnum
3
- from gllm_inference.utils import validate_string_enum as validate_string_enum
4
3
  from pydantic import BaseModel
5
4
 
6
5
  PROVIDER_SEPARATOR: str
@@ -22,6 +21,7 @@ class ModelProvider(StrEnum):
22
21
  PORTKEY = 'portkey'
23
22
  OPENAI_CHAT_COMPLETIONS = 'openai-chat-completions'
24
23
  OPENAI_COMPATIBLE = 'openai-compatible'
24
+ SEA_LION = 'sea-lion'
25
25
  TWELVELABS = 'twelvelabs'
26
26
  VOYAGE = 'voyage'
27
27
  XAI = 'xai'
@@ -1,5 +1,8 @@
1
+ from _typeshed import Incomplete
1
2
  from enum import StrEnum
2
3
 
4
+ logger: Incomplete
5
+
3
6
  def validate_string_enum(enum_type: type[StrEnum], value: str) -> None:
4
7
  """Validates that the provided value is a valid string enum value.
5
8
 
Binary file
gllm_inference.pyi CHANGED
@@ -33,6 +33,7 @@ import gllm_inference.lm_invoker.OpenAIChatCompletionsLMInvoker
33
33
  import gllm_inference.lm_invoker.OpenAICompatibleLMInvoker
34
34
  import gllm_inference.lm_invoker.OpenAILMInvoker
35
35
  import gllm_inference.lm_invoker.PortkeyLMInvoker
36
+ import gllm_inference.lm_invoker.SeaLionLMInvoker
36
37
  import gllm_inference.lm_invoker.XAILMInvoker
37
38
  import gllm_inference.prompt_builder.PromptBuilder
38
39
  import gllm_inference.output_parser.JSONOutputParser
@@ -57,7 +58,6 @@ import gllm_inference.schema.AttachmentType
57
58
  import gllm_inference.schema.EMContent
58
59
  import gllm_inference.schema.Vector
59
60
  import aioboto3
60
- import gllm_inference.utils.validate_string_enum
61
61
  import cohere
62
62
  import asyncio.CancelledError
63
63
  import gllm_inference.exceptions.convert_to_base_invoker_error
@@ -68,9 +68,7 @@ import google.genai
68
68
  import google.genai.types
69
69
  import httpx
70
70
  import gllm_inference.exceptions.ProviderInternalError
71
- import concurrent
72
- import concurrent.futures
73
- import concurrent.futures.ThreadPoolExecutor
71
+ import gllm_core.utils.concurrency
74
72
  import langchain_core
75
73
  import langchain_core.embeddings
76
74
  import gllm_inference.exceptions.InvokerRuntimeError
@@ -85,6 +83,7 @@ import voyageai
85
83
  import voyageai.client_async
86
84
  import http
87
85
  import http.HTTPStatus
86
+ import __future__
88
87
  import uuid
89
88
  import gllm_core.constants
90
89
  import gllm_core.event
@@ -115,9 +114,7 @@ import inspect
115
114
  import time
116
115
  import jsonschema
117
116
  import gllm_inference.lm_invoker.batch.BatchOperations
118
- import gllm_inference.schema.LMEventType
119
117
  import gllm_inference.schema.MessageContent
120
- import __future__
121
118
  import gllm_inference.schema.ActivityEvent
122
119
  import gllm_inference.schema.CodeEvent
123
120
  import gllm_inference.schema.CodeExecResult
@@ -139,6 +136,7 @@ import jinja2.sandbox
139
136
  import gllm_inference.schema.JinjaEnvType
140
137
  import gllm_inference.prompt_builder.format_strategy.JinjaFormatStrategy
141
138
  import gllm_inference.prompt_builder.format_strategy.StringFormatStrategy
139
+ import gllm_inference.schema.HistoryFormatter
142
140
  import transformers
143
141
  import gllm_inference.prompt_formatter.HuggingFacePromptFormatter
144
142
  import traceback
@@ -151,6 +149,7 @@ import pathlib
151
149
  import filetype
152
150
  import magic
153
151
  import requests
152
+ import gllm_core.schema.chunk
154
153
  import binascii
155
154
  import fnmatch
156
155
  import importlib
@@ -1,12 +1,12 @@
1
1
  Metadata-Version: 2.2
2
2
  Name: gllm-inference-binary
3
- Version: 0.5.51
3
+ Version: 0.5.66
4
4
  Summary: A library containing components related to model inferences in Gen AI applications.
5
- Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, Resti Febrina <resti.febrina@gdplabs.id>
5
+ Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, "Delfia N. A. Putri" <delfia.n.a.putri@gdplabs.id>
6
6
  Requires-Python: <3.14,>=3.11
7
7
  Description-Content-Type: text/markdown
8
8
  Requires-Dist: poetry<3.0.0,>=2.1.3
9
- Requires-Dist: gllm-core-binary<0.4.0,>=0.3.21
9
+ Requires-Dist: gllm-core-binary<0.4.0,>=0.3.23
10
10
  Requires-Dist: aiohttp<4.0.0,>=3.12.14
11
11
  Requires-Dist: filetype<2.0.0,>=1.2.0
12
12
  Requires-Dist: httpx<0.29.0,>=0.28.0
@@ -33,7 +33,7 @@ Requires-Dist: aioboto3<16.0.0,>=15.0.0; extra == "bedrock"
33
33
  Provides-Extra: cohere
34
34
  Requires-Dist: cohere<6.0.0,>=5.18.0; extra == "cohere"
35
35
  Provides-Extra: datasaur
36
- Requires-Dist: openai<2.0.0,>=1.98.0; extra == "datasaur"
36
+ Requires-Dist: openai<3.0.0,>=2.7.0; extra == "datasaur"
37
37
  Provides-Extra: google
38
38
  Requires-Dist: google-genai<=1.36,>=1.23; extra == "google"
39
39
  Provides-Extra: huggingface
@@ -42,7 +42,7 @@ Requires-Dist: transformers<5.0.0,>=4.52.0; extra == "huggingface"
42
42
  Provides-Extra: litellm
43
43
  Requires-Dist: litellm<2.0.0,>=1.69.2; extra == "litellm"
44
44
  Provides-Extra: openai
45
- Requires-Dist: openai<2.0.0,>=1.98.0; extra == "openai"
45
+ Requires-Dist: openai<3.0.0,>=2.7.0; extra == "openai"
46
46
  Provides-Extra: portkey-ai
47
47
  Requires-Dist: portkey-ai<2.0.0,>=1.14.4; extra == "portkey-ai"
48
48
  Provides-Extra: twelvelabs
@@ -1,11 +1,11 @@
1
- gllm_inference.cp311-win_amd64.pyd,sha256=GZ1kWbeDS9j3luJDKQ6QE6GGlQxcC6cF-YZW7wJrbpI,3872256
2
- gllm_inference.pyi,sha256=9EID8MhTbq9wGizE04ZelJSAUmiza8WdfCTqhQj3dAE,5191
1
+ gllm_inference.cp311-win_amd64.pyd,sha256=xik0kaZaQfmy3izrWe_LGbuuz8tIpFtVAtMehjg2feI,4151296
2
+ gllm_inference.pyi,sha256=jCxjKwYXb3t6yZrf6m7uD29V9KPXyWb8U04-lru4-JA,5173
3
3
  gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
4
- gllm_inference/constants.pyi,sha256=PncjVw-mkzcJ3ln1ohvVZGdJ-TD-VZy1Ygn4Va8Z7i0,350
4
+ gllm_inference/constants.pyi,sha256=8jIYOyxJYVWUYXSXF3vag9HhHwjq1iU9tzPiosRHkWk,328
5
5
  gllm_inference/builder/__init__.pyi,sha256=-bw1uDx7CAM7pkvjvb1ZXku9zXlQ7aEAyC83KIn3bz8,506
6
6
  gllm_inference/builder/_build_invoker.pyi,sha256=SZst1ocnRntcA7lWKp44lMJ3yRA5GP_jsUZkcsxwTjw,876
7
7
  gllm_inference/builder/build_em_invoker.pyi,sha256=Z09hP6_ziumI2XB-7HCdyIE4AyoV1oL9DBdoU2G8izo,5879
8
- gllm_inference/builder/build_lm_invoker.pyi,sha256=1Qgb0bD8tqobDh9kR_Kce7Vwu_Oljc2xv-HBIrC3HUY,9065
8
+ gllm_inference/builder/build_lm_invoker.pyi,sha256=ZKNVOOQAi1DaChrPBdWGUuzPShhmkGx89oZnALhR9w0,9417
9
9
  gllm_inference/builder/build_lm_request_processor.pyi,sha256=H7Rg88e7PTTCtuyY64r333moTmh4-ypOwgnG10gkEdY,4232
10
10
  gllm_inference/builder/build_output_parser.pyi,sha256=sgSTrzUmSRxPzUUum0fDU7A3NXYoYhpi6bEx4Q2XMnA,965
11
11
  gllm_inference/catalog/__init__.pyi,sha256=HWgPKWIzprpMHRKe_qN9BZSIQhVhrqiyjLjIXwvj1ho,291
@@ -15,7 +15,7 @@ gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=iViWB4SaezzjQY4UY1YxeoX
15
15
  gllm_inference/em_invoker/__init__.pyi,sha256=uCWfCjh5a5DciRFcUdbHndewokM3J5hp3mbhmM5wQC8,1211
16
16
  gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=TXC5Kgf1eZqK2FHKAyeG3LB1SEsSEStnbk9bI1mjC5k,5049
17
17
  gllm_inference/em_invoker/bedrock_em_invoker.pyi,sha256=ZPse0qNR74qFcc-9svTN8DzVO6kCLz-KYsLDo3cEEd0,6483
18
- gllm_inference/em_invoker/cohere_em_invoker.pyi,sha256=4eLqeKLoK8vJB61bGdttfWUUvNDBToBqNA6KQYBMT8s,6793
18
+ gllm_inference/em_invoker/cohere_em_invoker.pyi,sha256=Z_eBDfq9XErzpO2q47rfSB8LWxZY7Wxx1TTlvVh6k28,6708
19
19
  gllm_inference/em_invoker/em_invoker.pyi,sha256=YDYJ8TGScsz5Gg-OBnEENN1tI1RYvwoddypxUr6SAWw,5191
20
20
  gllm_inference/em_invoker/google_em_invoker.pyi,sha256=zZYjeLp9ncwIVM4UHqDJSVOFn1eXiaz9Ba24-_fCF2c,6953
21
21
  gllm_inference/em_invoker/jina_em_invoker.pyi,sha256=wrivFiPFRWBSw-TEAkaQKBvBRhlzN6CBaZkc0HOslRw,5754
@@ -40,33 +40,34 @@ gllm_inference/exceptions/__init__.pyi,sha256=nXOqwsuwUgsnBcJEANVuxbZ1nDfcJ6-pKU
40
40
  gllm_inference/exceptions/error_parser.pyi,sha256=4aiJZhBzBOqlhdmpvaCvildGy7_XxlJzQpe3PzGt8eE,2040
41
41
  gllm_inference/exceptions/exceptions.pyi,sha256=6y3ECgHAStqMGgQv8Dv-Ui-5PDD07mSj6qaRZeSWea4,5857
42
42
  gllm_inference/exceptions/provider_error_map.pyi,sha256=vWa4ZIHn7qIghECGvO-dS2KzOmf3c10GRWKZ4YDPnSQ,1267
43
- gllm_inference/lm_invoker/__init__.pyi,sha256=L2nlkj13WwWbDYEBtM0mlAj0-UbSilMjVLpCJ_0Eock,1502
44
- gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=JSgKUk9d1ZHlitv_ZjHlAk2hIW-J7u6yslVHflIeUro,16726
45
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=FYfRNPG-oD4wIfitjTHnGib1uMZL7Pid0gbrRsymAHU,14601
46
- gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=IMFD6Nr_4tkDDMy1WveOz91u_v1rjplVdtgep9UXsxw,11851
47
- gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=LR0EM4vTfufq9OWk8JVIwLyFeJFTguPNmPgJBUooSq4,8342
48
- gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=aSmEgoYj_V72Nb6erDResphw9RaHfbE5C6PhqpMfEeQ,17674
49
- gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=tJIxkFUKjLF-yz0niaDjN3L0QNCbn4sT8hmPKtERpog,12742
50
- gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=IJxRUkmgXY8oQwS7tJoskO8fiESB7M4pyvpE64pyXDo,12648
51
- gllm_inference/lm_invoker/lm_invoker.pyi,sha256=PS4cJ5VLNfHqeTgCerY-c1Xaa7ktdWAiT-f3Lnc0wvE,8665
52
- gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=uYJFgi4tJGab77232IC1gdoU9h9AqoClIUj6tM6O47s,15177
43
+ gllm_inference/lm_invoker/__init__.pyi,sha256=Y0laMHRdUwGgfIXo278qe1dwj7ZMSCeO-wMToYim47c,1618
44
+ gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=b3VM_mcK20y4Fyo_vWDx95HA001l9h53ACo0caP1sTY,15729
45
+ gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=uV98H2nJsElCTsxAuInZ9KSk1jOTq6SROAGQRPR-_r0,13173
46
+ gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=qXmFK6zsOM3nPfueEhY5pAfG24bZytA1jqemPa63vLY,10951
47
+ gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=FnpayOW_Zi0pWFSawLX8XahEnknbnpsRWrkhKZe8Y3U,8035
48
+ gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=2ZqQl9WhehtxcnAXzUi7vFDaHV0lO-0SwHoAT7m8Lf0,20484
49
+ gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=ull3cX-iUT4hYMbixcxqfrNUxR8ZoR4Vt9ACVILQWSM,12126
50
+ gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=qG8pPTiDJZR2e7wr5Q2VyceC227tz3QybX3UPihT5ng,11400
51
+ gllm_inference/lm_invoker/lm_invoker.pyi,sha256=TqlnNhWnFeteC_JjnnQ1pPW66qHPePC0iA738I1LdXo,9553
52
+ gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=qt9DAdJM7YBB4op-6SOJB0kCouPYVxtIamGUXLGLUeA,13888
53
53
  gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=T9sShA_9fgEuaaAuT2gJZq_EYNbEhf3IkWwMCwfszY8,4244
54
- gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=10iKCyleqHNbJc8M1rj3ogRcNlNxcVgyk0v6TcS6gf4,23452
55
- gllm_inference/lm_invoker/portkey_lm_invoker.pyi,sha256=BmZ5TFiQx3-6Ijf6J2ICzP6SCfnOFUVTPRLijv85oU0,13465
56
- gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=gyi12K7M9HkjNX6pU6NVv5Uq3-aHErixO-PVhHjioo8,14632
54
+ gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=Wu9cvkncnfMQI1Fht5c3i3LF0Tdd6Pz0-rpCTkWRe84,24918
55
+ gllm_inference/lm_invoker/portkey_lm_invoker.pyi,sha256=FYOp4BaDfOtompWIRhDqzMVVSK-TiFyw7JA4TznANQE,15236
56
+ gllm_inference/lm_invoker/sea_lion_lm_invoker.pyi,sha256=Qqplj79JsnLGY6xjhbXYbz6Ramxur4sXfvv4svk19os,3533
57
+ gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=6beZsQjGUTo7TdzWBWksRzVGT58XyipErpGfiRq6NH0,13017
57
58
  gllm_inference/lm_invoker/batch/__init__.pyi,sha256=vJOTHRJ83oq8Bq0UsMdID9_HW5JAxr06gUs4aPRZfEE,130
58
- gllm_inference/lm_invoker/batch/batch_operations.pyi,sha256=o2U17M41RKVFW6j_oxy-SxU1JqUtVt75pKRxrqXzorE,5499
59
+ gllm_inference/lm_invoker/batch/batch_operations.pyi,sha256=EKwINY8DQFOXquGGrUTr_Yg24ZrSyiiAo2hX8If9vyQ,5573
59
60
  gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
60
61
  gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=Ol_IqeBPgrmG5iaY3NV6AD-alC2HmUvljp0vI8GApWo,1154
61
62
  gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=rB1AWfER2BBKZ5I219211YE2EUFPF25bhzysqjdPgiY,1080
62
63
  gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=8lmb1PRbkqBsF_l7iOffxW0K5Xxpi69GW9Z7KxyxHTE,279
63
- gllm_inference/lm_invoker/schema/google.pyi,sha256=elXHrUMS46pbTsulk7hBXVVFcT022iD-_U_I590xeV8,529
64
+ gllm_inference/lm_invoker/schema/google.pyi,sha256=LQ14PJyDOe3K5TYvE-gzE1fjpZCSAy-0Sy9Lmw6fICY,827
64
65
  gllm_inference/lm_invoker/schema/langchain.pyi,sha256=2OJOUQPlGdlUbIOTDOyiWDBOMm3MoVX-kU2nK0zQsF0,452
65
- gllm_inference/lm_invoker/schema/openai.pyi,sha256=TsCr8_SM5kK2JyROeXtmH13n46TgKjLMc0agYlYUSZc,2328
66
+ gllm_inference/lm_invoker/schema/openai.pyi,sha256=yrL10gyrxvU3szXeKJC-mMkHO3nzdcaYDsBzORqK3AE,2777
66
67
  gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=nNPb7ETC9IrJwkV5wfbGf6Co3-qdq4lhcXz0l_qYCE4,1261
67
68
  gllm_inference/lm_invoker/schema/portkey.pyi,sha256=V2q4JIwDAR7BidqfmO01u1_1mLOMtm5OCon6sN2zNt0,662
68
69
  gllm_inference/lm_invoker/schema/xai.pyi,sha256=jpC6ZSBDUltzm9GjD6zvSFIPwqizn_ywLnjvwSa7KuU,663
69
- gllm_inference/model/__init__.pyi,sha256=e9Jq5V2iVPpjBh_bOEBoXdsU2LleAxKfJ0r-1rZJ5R0,822
70
+ gllm_inference/model/__init__.pyi,sha256=J3z0Rc6PI_aC0XuA_dWB2tCBQ4pSZdS5NzQ8vPhs5c4,907
70
71
  gllm_inference/model/em/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
71
72
  gllm_inference/model/em/cohere_em.pyi,sha256=uF1AmDO-skQteYqzxJ3DK10SqgfdW0oW9L8Ym34eU04,505
72
73
  gllm_inference/model/em/google_em.pyi,sha256=c53H-KNdNOK9ppPLyOSkmCA890eF5FsMd05upkPIzF0,487
@@ -78,12 +79,13 @@ gllm_inference/model/lm/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZ
78
79
  gllm_inference/model/lm/anthropic_lm.pyi,sha256=ccUpxddakurLFHivl5UzJxgODLhcFgx8XC7CKa-99NE,633
79
80
  gllm_inference/model/lm/google_lm.pyi,sha256=OLuoqT0FnJOLsNalulBMEXuCYAXoF8Y7vjfSBgjaJxA,529
80
81
  gllm_inference/model/lm/openai_lm.pyi,sha256=yj3AJj1xDYRkNIPHX2enw46AJ9wArPZruKsxg1ME9Rg,645
82
+ gllm_inference/model/lm/sea_lion_lm.pyi,sha256=Gz42ym_KJA7if3sDCFmdrb1uv9b9__n2RWOfPvOxm0Y,508
81
83
  gllm_inference/model/lm/xai_lm.pyi,sha256=O3G9Lj1Ii31CyCDrwYVkPPJN6X8V-WBF9xILUPUE-qY,525
82
84
  gllm_inference/output_parser/__init__.pyi,sha256=dhAeRTBxc6CfS8bhnHjbtrnyqJ1iyffvUZkGp4UrJNM,132
83
85
  gllm_inference/output_parser/json_output_parser.pyi,sha256=YtgQh8Uzy8W_Tgh8DfuR7VFFS7qvLEasiTwRfaGZZEU,2993
84
86
  gllm_inference/output_parser/output_parser.pyi,sha256=-Xu5onKCBDqShcO-VrQh5icqAmXdihGc3rkZxL93swg,975
85
87
  gllm_inference/prompt_builder/__init__.pyi,sha256=mPsbiafzSNHsgN-CuzjhgZpfXfi1pPC3_gdsq2p0EM4,120
86
- gllm_inference/prompt_builder/prompt_builder.pyi,sha256=67X0MXPRGb_Azifk5fM9lAsQX48rASsQk2gCPikqU5k,4626
88
+ gllm_inference/prompt_builder/prompt_builder.pyi,sha256=w99E5dsQxpJ-1FNEF_WBOt7xovmhp5WzrbTfRQ76K60,5083
87
89
  gllm_inference/prompt_builder/format_strategy/__init__.pyi,sha256=BliNAdrN2yNoQt8muLe1IknuE78UDQys_3L_8_2Tn9E,312
88
90
  gllm_inference/prompt_builder/format_strategy/format_strategy.pyi,sha256=Uay07qR0nMGbrOZ2twEyrIWYRKevBQL9ju_4fVfByyQ,2328
89
91
  gllm_inference/prompt_builder/format_strategy/jinja_format_strategy.pyi,sha256=W8SdOPWPz_U355G8rQwiIQ0Ys78vomUBFEEZiz3AS5k,2270
@@ -109,18 +111,19 @@ gllm_inference/realtime_chat/output_streamer/output_streamer.pyi,sha256=5P9NQ0aJ
109
111
  gllm_inference/request_processor/__init__.pyi,sha256=giEme2WFQhgyKiBZHhSet0_nKSCHwGy-_2p6NRzg0Zc,231
110
112
  gllm_inference/request_processor/lm_request_processor.pyi,sha256=0fy1HyILCVDw6y46E-7tLnQTRYx4ppeRMe0QP6t9Jyw,5990
111
113
  gllm_inference/request_processor/uses_lm_mixin.pyi,sha256=LYHq-zLoXEMel1LfVdYv7W3BZ8WtBLo_WWFjRf10Yto,6512
112
- gllm_inference/schema/__init__.pyi,sha256=Bpbo6a4NqSwJJnPBRKAKx2gAdGiDl1tsak-vJxfZ6UU,2284
114
+ gllm_inference/schema/__init__.pyi,sha256=89GWKdJ6DdPUTem42_Uc-5EsyNcfUJ-WnMVEkagpEwI,2524
113
115
  gllm_inference/schema/activity.pyi,sha256=atrU4OwLesA9FEt1H7K3gsUWYNdOqpI5i2VdWkmo6cs,2367
114
- gllm_inference/schema/attachment.pyi,sha256=9zgAjGXBjLfzPGaKi68FMW6b5mXdEA352nDe-ynOSvY,3385
116
+ gllm_inference/schema/attachment.pyi,sha256=myJ_cI_h5mwUdvmMrWpSQIwj3nIxe8SD7HxO37o_3D4,4611
115
117
  gllm_inference/schema/code_exec_result.pyi,sha256=WQ-ARoGM9r6nyRX-A0Ro1XKiqrc9R3jRYXZpu_xo5S4,573
116
118
  gllm_inference/schema/config.pyi,sha256=NVmjQK6HipIE0dKSfx12hgIC0O-S1HEcAc-TWlXAF5A,689
117
- gllm_inference/schema/enums.pyi,sha256=wbD5Qifv9y0c3_FR_M4WBvlDLzq0bQbWwox_Lw3f-KM,1820
118
- gllm_inference/schema/events.pyi,sha256=YStRTYGtYlM0a46AfCuBwEaijsRujTSkEusJ-M6cvSY,4810
119
+ gllm_inference/schema/enums.pyi,sha256=U-lADk7xNjA5rByem0fo0xDAy9T3GsekY8B5szTlWPQ,2193
120
+ gllm_inference/schema/events.pyi,sha256=_CKuGNzb3j2Y1dOB2yssFkT_9FQz1AY4J_ApCwKNizU,4743
121
+ gllm_inference/schema/formatter.pyi,sha256=uJ1CZt_DmizfCPHD5deQwm2ftU-zcl2aj5rwJGTPe9c,1312
119
122
  gllm_inference/schema/lm_input.pyi,sha256=HxQiZgY7zcXh_Dw8nK8LSeBTZEHMPZVwmPmnfgSsAbs,197
120
- gllm_inference/schema/lm_output.pyi,sha256=DIV8BiIOPaSnMKxzKzH_Mp7j7-MScWCvmllegJDLqFg,2479
123
+ gllm_inference/schema/lm_output.pyi,sha256=N75CIF_2kZRdXKy2jvu9hhqzk5DrCbsHXTrhKqQ-7vo,11667
121
124
  gllm_inference/schema/mcp.pyi,sha256=4SgQ83pEowfWm2p-w9lupV4NayqqVBOy7SuYxIFeWRs,1045
122
125
  gllm_inference/schema/message.pyi,sha256=jJV6A0ihEcun2OhzyMtNkiHnf7d6v5R-GdpTBGfJ0AQ,2272
123
- gllm_inference/schema/model_id.pyi,sha256=Ozu46pp1uaPPvXJ6SPKpiL1vETPJqErEjNAYfOb5b6Y,6070
126
+ gllm_inference/schema/model_id.pyi,sha256=JhTzjCsOzHGds_UF8PlPON3xtqmSjUEfvjpmYyf3CJ0,6018
124
127
  gllm_inference/schema/reasoning.pyi,sha256=jbPxkDRHt0Vt-zdcc8lTT1l2hIE1Jm3HIHeNd0hfXGo,577
125
128
  gllm_inference/schema/token_usage.pyi,sha256=WJiGQyz5qatzBK2b-sABLCyTRLCBbAvxCRcqSJOzu-8,3025
126
129
  gllm_inference/schema/tool_call.pyi,sha256=OWT9LUqs_xfUcOkPG0aokAAqzLYYDkfnjTa0zOWvugk,403
@@ -129,9 +132,9 @@ gllm_inference/schema/type_alias.pyi,sha256=L-V0WxsFQznzAfby3DH8XMHUgjjZxQEsLw8S
129
132
  gllm_inference/utils/__init__.pyi,sha256=H27RiiFjD6WQHRrYb1-sBnb2aqjVENw5_8-DdAe1k9A,396
130
133
  gllm_inference/utils/io_utils.pyi,sha256=Eg7dvHWdXslTKdjh1j3dG50i7r35XG2zTmJ9XXvz4cI,1120
131
134
  gllm_inference/utils/langchain.pyi,sha256=4AwFiVAO0ZpdgmqeC4Pb5NJwBt8vVr0MSUqLeCdTscc,1194
132
- gllm_inference/utils/validation.pyi,sha256=-RdMmb8afH7F7q4Ao7x6FbwaDfxUHn3hA3WiOgzB-3s,397
135
+ gllm_inference/utils/validation.pyi,sha256=OWRZxeVGIuuvNU0LqLGB-9gNmypvbH-LcSJx91rnH1k,453
133
136
  gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
134
- gllm_inference_binary-0.5.51.dist-info/METADATA,sha256=YYGvuy7NbswW0XC-0PcrjX1olPf4rOGAchD5hHbWgto,5945
135
- gllm_inference_binary-0.5.51.dist-info/WHEEL,sha256=l2aKBREYfqJ7T2ljmr6hUiXPoNvvXF47bG4IHjuSyS4,96
136
- gllm_inference_binary-0.5.51.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
137
- gllm_inference_binary-0.5.51.dist-info/RECORD,,
137
+ gllm_inference_binary-0.5.66.dist-info/METADATA,sha256=JS0et7h2YvGkt2guUZO2aNaAm4rE77yALrYxMVHoKls,5953
138
+ gllm_inference_binary-0.5.66.dist-info/WHEEL,sha256=l2aKBREYfqJ7T2ljmr6hUiXPoNvvXF47bG4IHjuSyS4,96
139
+ gllm_inference_binary-0.5.66.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
140
+ gllm_inference_binary-0.5.66.dist-info/RECORD,,