gllm-inference-binary 0.5.6__cp313-cp313-macosx_13_0_x86_64.whl → 0.5.8__cp313-cp313-macosx_13_0_x86_64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of gllm-inference-binary might be problematic. Click here for more details.
- gllm_inference/constants.pyi +1 -0
- gllm_inference/em_invoker/azure_openai_em_invoker.pyi +2 -1
- gllm_inference/em_invoker/google_em_invoker.pyi +1 -0
- gllm_inference/em_invoker/langchain_em_invoker.pyi +3 -1
- gllm_inference/em_invoker/openai_compatible_em_invoker.pyi +2 -0
- gllm_inference/em_invoker/openai_em_invoker.pyi +2 -0
- gllm_inference/em_invoker/schema/google.pyi +7 -0
- gllm_inference/em_invoker/schema/langchain.pyi +4 -0
- gllm_inference/em_invoker/schema/openai.pyi +7 -0
- gllm_inference/em_invoker/schema/openai_compatible.pyi +7 -0
- gllm_inference/em_invoker/schema/twelvelabs.pyi +2 -0
- gllm_inference/em_invoker/schema/voyage.pyi +4 -0
- gllm_inference/em_invoker/twelevelabs_em_invoker.pyi +1 -0
- gllm_inference/em_invoker/voyage_em_invoker.pyi +1 -0
- gllm_inference/lm_invoker/anthropic_lm_invoker.pyi +9 -6
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi +8 -6
- gllm_inference/lm_invoker/bedrock_lm_invoker.pyi +8 -6
- gllm_inference/lm_invoker/datasaur_lm_invoker.pyi +5 -4
- gllm_inference/lm_invoker/google_lm_invoker.pyi +7 -6
- gllm_inference/lm_invoker/langchain_lm_invoker.pyi +9 -6
- gllm_inference/lm_invoker/litellm_lm_invoker.pyi +6 -4
- gllm_inference/lm_invoker/lm_invoker.pyi +13 -6
- gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi +7 -5
- gllm_inference/lm_invoker/openai_lm_invoker.pyi +8 -6
- gllm_inference/lm_invoker/schema/anthropic.pyi +2 -1
- gllm_inference/lm_invoker/schema/datasaur.pyi +2 -0
- gllm_inference/lm_invoker/schema/google.pyi +3 -0
- gllm_inference/lm_invoker/schema/langchain.pyi +2 -0
- gllm_inference/lm_invoker/schema/openai.pyi +1 -1
- gllm_inference/lm_invoker/schema/openai_compatible.pyi +4 -0
- gllm_inference.cpython-313-darwin.so +0 -0
- gllm_inference.pyi +5 -4
- {gllm_inference_binary-0.5.6.dist-info → gllm_inference_binary-0.5.8.dist-info}/METADATA +1 -1
- {gllm_inference_binary-0.5.6.dist-info → gllm_inference_binary-0.5.8.dist-info}/RECORD +35 -31
- {gllm_inference_binary-0.5.6.dist-info → gllm_inference_binary-0.5.8.dist-info}/WHEEL +0 -0
gllm_inference/constants.pyi
CHANGED
|
@@ -1,7 +1,8 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
3
|
-
from gllm_inference.constants import DEFAULT_AZURE_OPENAI_API_VERSION as DEFAULT_AZURE_OPENAI_API_VERSION
|
|
3
|
+
from gllm_inference.constants import DEFAULT_AZURE_OPENAI_API_VERSION as DEFAULT_AZURE_OPENAI_API_VERSION, INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
4
4
|
from gllm_inference.em_invoker.openai_em_invoker import OpenAIEMInvoker as OpenAIEMInvoker
|
|
5
|
+
from gllm_inference.em_invoker.schema.openai import Key as Key
|
|
5
6
|
from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider
|
|
6
7
|
from typing import Any
|
|
7
8
|
|
|
@@ -2,6 +2,7 @@ from _typeshed import Incomplete
|
|
|
2
2
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
3
3
|
from gllm_inference.constants import GOOGLE_SCOPES as GOOGLE_SCOPES
|
|
4
4
|
from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
|
|
5
|
+
from gllm_inference.em_invoker.schema.google import Key as Key
|
|
5
6
|
from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider, Vector as Vector
|
|
6
7
|
from typing import Any
|
|
7
8
|
|
|
@@ -1,6 +1,8 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
|
-
from gllm_core.utils.retry import RetryConfig
|
|
2
|
+
from gllm_core.utils.retry import RetryConfig
|
|
3
|
+
from gllm_inference.constants import INVOKER_DEFAULT_TIMEOUT as INVOKER_DEFAULT_TIMEOUT, INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
3
4
|
from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
|
|
5
|
+
from gllm_inference.em_invoker.schema.langchain import Key as Key
|
|
4
6
|
from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider, Vector as Vector
|
|
5
7
|
from gllm_inference.utils import load_langchain_model as load_langchain_model, parse_model_data as parse_model_data
|
|
6
8
|
from langchain_core.embeddings import Embeddings as Embeddings
|
|
@@ -1,6 +1,8 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
3
|
+
from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
3
4
|
from gllm_inference.em_invoker.openai_em_invoker import OpenAIEMInvoker as OpenAIEMInvoker
|
|
5
|
+
from gllm_inference.em_invoker.schema.openai_compatible import Key as Key
|
|
4
6
|
from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider
|
|
5
7
|
from typing import Any
|
|
6
8
|
|
|
@@ -1,6 +1,8 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
3
|
+
from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
3
4
|
from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
|
|
5
|
+
from gllm_inference.em_invoker.schema.openai import Key as Key
|
|
4
6
|
from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider, Vector as Vector
|
|
5
7
|
from typing import Any
|
|
6
8
|
|
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
3
|
+
from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
3
4
|
from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
|
|
4
5
|
from gllm_inference.em_invoker.schema.twelvelabs import InputType as InputType, Key as Key, OutputType as OutputType
|
|
5
6
|
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EMContent as EMContent, ModelId as ModelId, ModelProvider as ModelProvider, Vector as Vector
|
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
3
|
+
from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
3
4
|
from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
|
|
4
5
|
from gllm_inference.em_invoker.schema.voyage import InputType as InputType, Key as Key
|
|
5
6
|
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EMContent as EMContent, ModelId as ModelId, ModelProvider as ModelProvider, Vector as Vector
|
|
@@ -1,10 +1,12 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.event import EventEmitter as EventEmitter
|
|
3
|
+
from gllm_core.schema.tool import Tool as Tool
|
|
3
4
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
5
|
+
from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
4
6
|
from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
|
|
5
7
|
from gllm_inference.lm_invoker.schema.anthropic import InputType as InputType, Key as Key, OutputType as OutputType
|
|
6
8
|
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EmitDataType as EmitDataType, LMOutput as LMOutput, Message as Message, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
|
|
7
|
-
from langchain_core.tools import Tool as
|
|
9
|
+
from langchain_core.tools import Tool as LangChainTool
|
|
8
10
|
from typing import Any
|
|
9
11
|
|
|
10
12
|
SUPPORTED_ATTACHMENTS: Incomplete
|
|
@@ -48,7 +50,7 @@ class AnthropicLMInvoker(BaseLMInvoker):
|
|
|
48
50
|
|
|
49
51
|
Tool calling:
|
|
50
52
|
Tool calling is a feature that allows the language model to call tools to perform tasks.
|
|
51
|
-
Tools can be passed to the via the `tools` parameter as a list of
|
|
53
|
+
Tools can be passed to the via the `tools` parameter as a list of `Tool` objects.
|
|
52
54
|
When tools are provided and the model decides to call a tool, the tool calls are stored in the
|
|
53
55
|
`tool_calls` attribute in the output.
|
|
54
56
|
|
|
@@ -221,7 +223,7 @@ class AnthropicLMInvoker(BaseLMInvoker):
|
|
|
221
223
|
client: Incomplete
|
|
222
224
|
thinking: Incomplete
|
|
223
225
|
thinking_budget: Incomplete
|
|
224
|
-
def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, thinking: bool = False, thinking_budget: int = ...) -> None:
|
|
226
|
+
def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, thinking: bool = False, thinking_budget: int = ...) -> None:
|
|
225
227
|
"""Initializes the AnthropicLmInvoker instance.
|
|
226
228
|
|
|
227
229
|
Args:
|
|
@@ -231,7 +233,8 @@ class AnthropicLMInvoker(BaseLMInvoker):
|
|
|
231
233
|
model_kwargs (dict[str, Any] | None, optional): Additional keyword arguments for the Anthropic client.
|
|
232
234
|
default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
|
|
233
235
|
Defaults to None.
|
|
234
|
-
tools (list[Tool] | None, optional): Tools provided to the model to enable tool calling.
|
|
236
|
+
tools (list[Tool | LangChainTool] | None, optional): Tools provided to the model to enable tool calling.
|
|
237
|
+
Defaults to None, in which case an empty list is used.
|
|
235
238
|
response_schema (ResponseSchema | None, optional): The schema of the response. If provided, the model will
|
|
236
239
|
output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
|
|
237
240
|
dictionary. Defaults to None.
|
|
@@ -247,13 +250,13 @@ class AnthropicLMInvoker(BaseLMInvoker):
|
|
|
247
250
|
1. `thinking` is True, but the `thinking_budget` is less than 1024.
|
|
248
251
|
3. `response_schema` is provided, but `tools` or `thinking` are also provided.
|
|
249
252
|
"""
|
|
250
|
-
def set_tools(self, tools: list[Tool]) -> None:
|
|
253
|
+
def set_tools(self, tools: list[Tool | LangChainTool]) -> None:
|
|
251
254
|
"""Sets the tools for the Anthropic language model.
|
|
252
255
|
|
|
253
256
|
This method sets the tools for the Anthropic language model. Any existing tools will be replaced.
|
|
254
257
|
|
|
255
258
|
Args:
|
|
256
|
-
tools (list[Tool]): The list of tools to be used.
|
|
259
|
+
tools (list[Tool | LangChainTool]): The list of tools to be used.
|
|
257
260
|
|
|
258
261
|
Raises:
|
|
259
262
|
ValueError: If `response_schema` exists.
|
|
@@ -1,9 +1,11 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
|
+
from gllm_core.schema.tool import Tool as Tool
|
|
2
3
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
3
|
-
from gllm_inference.constants import DEFAULT_AZURE_OPENAI_API_VERSION as DEFAULT_AZURE_OPENAI_API_VERSION
|
|
4
|
+
from gllm_inference.constants import DEFAULT_AZURE_OPENAI_API_VERSION as DEFAULT_AZURE_OPENAI_API_VERSION, INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
4
5
|
from gllm_inference.lm_invoker.openai_lm_invoker import OpenAILMInvoker as OpenAILMInvoker, ReasoningEffort as ReasoningEffort, ReasoningSummary as ReasoningSummary
|
|
6
|
+
from gllm_inference.lm_invoker.schema.openai import Key as Key
|
|
5
7
|
from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema
|
|
6
|
-
from langchain_core.tools import Tool as
|
|
8
|
+
from langchain_core.tools import Tool as LangChainTool
|
|
7
9
|
from typing import Any
|
|
8
10
|
|
|
9
11
|
class AzureOpenAILMInvoker(OpenAILMInvoker):
|
|
@@ -15,7 +17,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
|
|
|
15
17
|
model_name (str): The name of the Azure OpenAI language model deployment.
|
|
16
18
|
client (AsyncAzureOpenAI): The Azure OpenAI client instance.
|
|
17
19
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
|
|
18
|
-
tools (list[
|
|
20
|
+
tools (list[Tool]): The list of tools provided to the model to enable tool calling.
|
|
19
21
|
response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
|
|
20
22
|
structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema dictionary.
|
|
21
23
|
output_analytics (bool): Whether to output the invocation analytics.
|
|
@@ -50,7 +52,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
|
|
|
50
52
|
|
|
51
53
|
Tool calling:
|
|
52
54
|
Tool calling is a feature that allows the language model to call tools to perform tasks.
|
|
53
|
-
Tools can be passed to the via the `tools` parameter as a list of
|
|
55
|
+
Tools can be passed to the via the `tools` parameter as a list of `Tool` objects.
|
|
54
56
|
When tools are provided and the model decides to call a tool, the tool calls are stored in the
|
|
55
57
|
`tool_calls` attribute in the output.
|
|
56
58
|
|
|
@@ -218,7 +220,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
|
|
|
218
220
|
Defaults to an empty list.
|
|
219
221
|
'''
|
|
220
222
|
client: Incomplete
|
|
221
|
-
def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None) -> None:
|
|
223
|
+
def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None) -> None:
|
|
222
224
|
"""Initializes a new instance of the AzureOpenAILMInvoker class.
|
|
223
225
|
|
|
224
226
|
Args:
|
|
@@ -231,7 +233,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
|
|
|
231
233
|
model_kwargs (dict[str, Any] | None, optional): Additional model parameters. Defaults to None.
|
|
232
234
|
default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
|
|
233
235
|
Defaults to None.
|
|
234
|
-
tools (list[Tool] | None, optional): Tools provided to the
|
|
236
|
+
tools (list[Tool | LangChainTool] | None, optional): Tools provided to the model to enable tool calling.
|
|
235
237
|
Defaults to None.
|
|
236
238
|
response_schema (ResponseSchema | None, optional): The schema of the response. If provided, the model will
|
|
237
239
|
output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
|
|
@@ -1,10 +1,11 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.event import EventEmitter as EventEmitter
|
|
3
|
+
from gllm_core.schema.tool import Tool as Tool
|
|
3
4
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
4
5
|
from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
|
|
5
6
|
from gllm_inference.lm_invoker.schema.bedrock import InputType as InputType, Key as Key, OutputType as OutputType
|
|
6
7
|
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, LMOutput as LMOutput, Message as Message, ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
|
|
7
|
-
from langchain_core.tools import Tool as
|
|
8
|
+
from langchain_core.tools import Tool as LangChainTool
|
|
8
9
|
from typing import Any
|
|
9
10
|
|
|
10
11
|
SUPPORTED_ATTACHMENTS: Incomplete
|
|
@@ -49,7 +50,7 @@ class BedrockLMInvoker(BaseLMInvoker):
|
|
|
49
50
|
|
|
50
51
|
Tool calling:
|
|
51
52
|
Tool calling is a feature that allows the language model to call tools to perform tasks.
|
|
52
|
-
Tools can be passed to the via the `tools` parameter as a list of
|
|
53
|
+
Tools can be passed to the via the `tools` parameter as a list of `Tool` objects.
|
|
53
54
|
When tools are provided and the model decides to call a tool, the tool calls are stored in the
|
|
54
55
|
`tool_calls` attribute in the output.
|
|
55
56
|
|
|
@@ -179,7 +180,7 @@ class BedrockLMInvoker(BaseLMInvoker):
|
|
|
179
180
|
'''
|
|
180
181
|
session: Incomplete
|
|
181
182
|
client_kwargs: Incomplete
|
|
182
|
-
def __init__(self, model_name: str, access_key_id: str | None = None, secret_access_key: str | None = None, region_name: str = 'us-east-1', model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None) -> None:
|
|
183
|
+
def __init__(self, model_name: str, access_key_id: str | None = None, secret_access_key: str | None = None, region_name: str = 'us-east-1', model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None) -> None:
|
|
183
184
|
'''Initializes the BedrockLMInvoker instance.
|
|
184
185
|
|
|
185
186
|
Args:
|
|
@@ -192,7 +193,8 @@ class BedrockLMInvoker(BaseLMInvoker):
|
|
|
192
193
|
model_kwargs (dict[str, Any] | None, optional): Additional keyword arguments for the Bedrock client.
|
|
193
194
|
default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
|
|
194
195
|
Defaults to None.
|
|
195
|
-
tools (list[Tool] | None, optional): Tools provided to the model to enable tool calling.
|
|
196
|
+
tools (list[Tool | LangChainTool] | None, optional): Tools provided to the model to enable tool calling.
|
|
197
|
+
Defaults to None.
|
|
196
198
|
response_schema (ResponseSchema | None, optional): The schema of the response. If provided, the model will
|
|
197
199
|
output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
|
|
198
200
|
dictionary. Defaults to None.
|
|
@@ -205,13 +207,13 @@ class BedrockLMInvoker(BaseLMInvoker):
|
|
|
205
207
|
ValueError: If `access_key_id` or `secret_access_key` is neither provided nor set in the
|
|
206
208
|
`AWS_ACCESS_KEY_ID` or `AWS_SECRET_ACCESS_KEY` environment variables, respectively.
|
|
207
209
|
'''
|
|
208
|
-
def set_tools(self, tools: list[Tool]) -> None:
|
|
210
|
+
def set_tools(self, tools: list[Tool | LangChainTool]) -> None:
|
|
209
211
|
"""Sets the tools for the Bedrock language model.
|
|
210
212
|
|
|
211
213
|
This method sets the tools for the Bedrock language model. Any existing tools will be replaced.
|
|
212
214
|
|
|
213
215
|
Args:
|
|
214
|
-
tools (list[Tool]): The list of tools to be used.
|
|
216
|
+
tools (list[Tool | LangChainTool]): The list of tools to be used.
|
|
215
217
|
|
|
216
218
|
Raises:
|
|
217
219
|
ValueError: If `response_schema` exists.
|
|
@@ -1,11 +1,12 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.event import EventEmitter as EventEmitter
|
|
3
|
+
from gllm_core.schema.tool import Tool as Tool
|
|
3
4
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
4
|
-
from gllm_inference.constants import DOCUMENT_MIME_TYPES as DOCUMENT_MIME_TYPES
|
|
5
|
+
from gllm_inference.constants import DOCUMENT_MIME_TYPES as DOCUMENT_MIME_TYPES, INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
5
6
|
from gllm_inference.lm_invoker.openai_compatible_lm_invoker import OpenAICompatibleLMInvoker as OpenAICompatibleLMInvoker
|
|
6
7
|
from gllm_inference.lm_invoker.schema.datasaur import InputType as InputType, Key as Key
|
|
7
8
|
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, LMOutput as LMOutput, Message as Message, ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema, ToolCall as ToolCall, ToolResult as ToolResult
|
|
8
|
-
from langchain_core.tools import Tool as
|
|
9
|
+
from langchain_core.tools import Tool as LangChainTool
|
|
9
10
|
from typing import Any
|
|
10
11
|
|
|
11
12
|
SUPPORTED_ATTACHMENTS: Incomplete
|
|
@@ -139,14 +140,14 @@ class DatasaurLMInvoker(OpenAICompatibleLMInvoker):
|
|
|
139
140
|
Raises:
|
|
140
141
|
ValueError: If the `api_key` is not provided and the `DATASAUR_API_KEY` environment variable is not set.
|
|
141
142
|
"""
|
|
142
|
-
def set_tools(self, tools: list[Tool]) -> None:
|
|
143
|
+
def set_tools(self, tools: list[Tool | LangChainTool]) -> None:
|
|
143
144
|
"""Sets the tools for the Datasaur LLM Projects Deployment API.
|
|
144
145
|
|
|
145
146
|
This method is raises a `NotImplementedError` because the Datasaur LLM Projects Deployment API does not
|
|
146
147
|
support tools.
|
|
147
148
|
|
|
148
149
|
Args:
|
|
149
|
-
tools (list[Tool]): The list of tools to be used.
|
|
150
|
+
tools (list[Tool | LangChainTool]): The list of tools to be used.
|
|
150
151
|
|
|
151
152
|
Raises:
|
|
152
153
|
NotImplementedError: This method is not supported for the Datasaur LLM Projects Deployment API.
|
|
@@ -1,11 +1,12 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.event import EventEmitter as EventEmitter
|
|
3
|
+
from gllm_core.schema.tool import Tool
|
|
3
4
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
4
5
|
from gllm_inference.constants import GOOGLE_SCOPES as GOOGLE_SCOPES
|
|
5
6
|
from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
|
|
6
7
|
from gllm_inference.lm_invoker.schema.google import InputType as InputType, Key as Key
|
|
7
8
|
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EmitDataType as EmitDataType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
|
|
8
|
-
from langchain_core.tools import Tool
|
|
9
|
+
from langchain_core.tools import Tool as LangChainTool
|
|
9
10
|
from typing import Any
|
|
10
11
|
|
|
11
12
|
SUPPORTED_ATTACHMENTS: Incomplete
|
|
@@ -79,7 +80,7 @@ class GoogleLMInvoker(BaseLMInvoker):
|
|
|
79
80
|
|
|
80
81
|
Tool calling:
|
|
81
82
|
Tool calling is a feature that allows the language model to call tools to perform tasks.
|
|
82
|
-
Tools can be passed to the via the `tools` parameter as a list of
|
|
83
|
+
Tools can be passed to the via the `tools` parameter as a list of `Tool` objects.
|
|
83
84
|
When tools are provided and the model decides to call a tool, the tool calls are stored in the
|
|
84
85
|
`tool_calls` attribute in the output.
|
|
85
86
|
|
|
@@ -250,7 +251,7 @@ class GoogleLMInvoker(BaseLMInvoker):
|
|
|
250
251
|
client_params: Incomplete
|
|
251
252
|
thinking: Incomplete
|
|
252
253
|
thinking_budget: Incomplete
|
|
253
|
-
def __init__(self, model_name: str, api_key: str | None = None, credentials_path: str | None = None, project_id: str | None = None, location: str = 'us-central1', model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, thinking: bool | None = None, thinking_budget: int = ...) -> None:
|
|
254
|
+
def __init__(self, model_name: str, api_key: str | None = None, credentials_path: str | None = None, project_id: str | None = None, location: str = 'us-central1', model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, thinking: bool | None = None, thinking_budget: int = ...) -> None:
|
|
254
255
|
'''Initializes a new instance of the GoogleLMInvoker class.
|
|
255
256
|
|
|
256
257
|
Args:
|
|
@@ -267,7 +268,7 @@ class GoogleLMInvoker(BaseLMInvoker):
|
|
|
267
268
|
client.
|
|
268
269
|
default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
|
|
269
270
|
Defaults to None.
|
|
270
|
-
tools (list[Tool] | None, optional): Tools provided to the
|
|
271
|
+
tools (list[Tool | LangChainTool] | None, optional): Tools provided to the model to enable tool calling.
|
|
271
272
|
Defaults to None.
|
|
272
273
|
response_schema (ResponseSchema | None, optional): The schema of the response. If provided, the model will
|
|
273
274
|
output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
|
|
@@ -284,13 +285,13 @@ class GoogleLMInvoker(BaseLMInvoker):
|
|
|
284
285
|
If neither `api_key` nor `credentials_path` is provided, Google Gen AI will be used by default.
|
|
285
286
|
The `GOOGLE_API_KEY` environment variable will be used for authentication.
|
|
286
287
|
'''
|
|
287
|
-
def set_tools(self, tools: list[Tool]) -> None:
|
|
288
|
+
def set_tools(self, tools: list[Tool | LangChainTool]) -> None:
|
|
288
289
|
"""Sets the tools for the Google language model.
|
|
289
290
|
|
|
290
291
|
This method sets the tools for the Google language model. Any existing tools will be replaced.
|
|
291
292
|
|
|
292
293
|
Args:
|
|
293
|
-
tools (list[Tool]): The list of tools to be used.
|
|
294
|
+
tools (list[Tool | LangChainTool]): The list of tools to be used.
|
|
294
295
|
|
|
295
296
|
Raises:
|
|
296
297
|
ValueError: If `response_schema` exists.
|
|
@@ -1,13 +1,15 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.event import EventEmitter as EventEmitter
|
|
3
|
-
from gllm_core.
|
|
3
|
+
from gllm_core.schema.tool import Tool as Tool
|
|
4
|
+
from gllm_core.utils.retry import RetryConfig
|
|
5
|
+
from gllm_inference.constants import INVOKER_DEFAULT_TIMEOUT as INVOKER_DEFAULT_TIMEOUT, INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
4
6
|
from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
|
|
5
7
|
from gllm_inference.lm_invoker.schema.langchain import InputType as InputType, Key as Key
|
|
6
8
|
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
|
|
7
9
|
from gllm_inference.utils import load_langchain_model as load_langchain_model, parse_model_data as parse_model_data
|
|
8
10
|
from langchain_core.language_models import BaseChatModel as BaseChatModel
|
|
9
11
|
from langchain_core.messages import BaseMessage as BaseMessage
|
|
10
|
-
from langchain_core.tools import Tool as
|
|
12
|
+
from langchain_core.tools import Tool as LangChainTool
|
|
11
13
|
from typing import Any
|
|
12
14
|
|
|
13
15
|
SUPPORTED_ATTACHMENTS: Incomplete
|
|
@@ -76,7 +78,7 @@ class LangChainLMInvoker(BaseLMInvoker):
|
|
|
76
78
|
|
|
77
79
|
Tool calling:
|
|
78
80
|
Tool calling is a feature that allows the language model to call tools to perform tasks.
|
|
79
|
-
Tools can be passed to the via the `tools` parameter as a list of
|
|
81
|
+
Tools can be passed to the via the `tools` parameter as a list of `Tool` objects.
|
|
80
82
|
When tools are provided and the model decides to call a tool, the tool calls are stored in the
|
|
81
83
|
`tool_calls` attribute in the output.
|
|
82
84
|
|
|
@@ -203,7 +205,7 @@ class LangChainLMInvoker(BaseLMInvoker):
|
|
|
203
205
|
Defaults to an empty list.
|
|
204
206
|
'''
|
|
205
207
|
model: Incomplete
|
|
206
|
-
def __init__(self, model: BaseChatModel | None = None, model_class_path: str | None = None, model_name: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None) -> None:
|
|
208
|
+
def __init__(self, model: BaseChatModel | None = None, model_class_path: str | None = None, model_name: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None) -> None:
|
|
207
209
|
'''Initializes a new instance of the LangChainLMInvoker class.
|
|
208
210
|
|
|
209
211
|
Args:
|
|
@@ -218,7 +220,8 @@ class LangChainLMInvoker(BaseLMInvoker):
|
|
|
218
220
|
`model_class_path` is provided. Defaults to None.
|
|
219
221
|
default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
|
|
220
222
|
Defaults to None.
|
|
221
|
-
tools (list[Tool] | None, optional): Tools provided to the model to enable tool calling.
|
|
223
|
+
tools (list[Tool | LangChainTool] | None, optional): Tools provided to the model to enable tool calling.
|
|
224
|
+
Defaults to None.
|
|
222
225
|
response_schema (ResponseSchema | None, optional): The schema of the response. If provided, the model will
|
|
223
226
|
output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
|
|
224
227
|
dictionary. Defaults to None.
|
|
@@ -230,7 +233,7 @@ class LangChainLMInvoker(BaseLMInvoker):
|
|
|
230
233
|
ValueError: If `response_schema` is provided, but `tools` are also provided.
|
|
231
234
|
'''
|
|
232
235
|
tools: Incomplete
|
|
233
|
-
def set_tools(self, tools: list[Tool]) -> None:
|
|
236
|
+
def set_tools(self, tools: list[Tool | LangChainTool]) -> None:
|
|
234
237
|
"""Sets the tools for LangChain's BaseChatModel.
|
|
235
238
|
|
|
236
239
|
This method sets the tools for LangChain's BaseChatModel. Any existing tools will be replaced.
|
|
@@ -1,10 +1,11 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.event import EventEmitter as EventEmitter
|
|
3
|
+
from gllm_core.schema.tool import Tool as Tool
|
|
3
4
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
4
5
|
from gllm_inference.lm_invoker.openai_compatible_lm_invoker import OpenAICompatibleLMInvoker as OpenAICompatibleLMInvoker
|
|
5
6
|
from gllm_inference.lm_invoker.openai_lm_invoker import ReasoningEffort as ReasoningEffort
|
|
6
7
|
from gllm_inference.schema import AttachmentType as AttachmentType, LMOutput as LMOutput, ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema
|
|
7
|
-
from langchain_core.tools import Tool as
|
|
8
|
+
from langchain_core.tools import Tool as LangChainTool
|
|
8
9
|
from typing import Any
|
|
9
10
|
|
|
10
11
|
SUPPORTED_ATTACHMENTS: Incomplete
|
|
@@ -58,7 +59,7 @@ class LiteLLMLMInvoker(OpenAICompatibleLMInvoker):
|
|
|
58
59
|
|
|
59
60
|
Tool calling:
|
|
60
61
|
Tool calling is a feature that allows the language model to call tools to perform tasks.
|
|
61
|
-
Tools can be passed to the via the `tools` parameter as a list of
|
|
62
|
+
Tools can be passed to the via the `tools` parameter as a list of `Tool` objects.
|
|
62
63
|
When tools are provided and the model decides to call a tool, the tool calls are stored in the
|
|
63
64
|
`tool_calls` attribute in the output.
|
|
64
65
|
|
|
@@ -227,14 +228,15 @@ class LiteLLMLMInvoker(OpenAICompatibleLMInvoker):
|
|
|
227
228
|
Defaults to an empty list.
|
|
228
229
|
'''
|
|
229
230
|
completion: Incomplete
|
|
230
|
-
def __init__(self, model_id: str, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None) -> None:
|
|
231
|
+
def __init__(self, model_id: str, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None) -> None:
|
|
231
232
|
"""Initializes a new instance of the LiteLLMLMInvoker class.
|
|
232
233
|
|
|
233
234
|
Args:
|
|
234
235
|
model_id (str): The ID of the model to use. Must be in the format of `provider/model_name`.
|
|
235
236
|
default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
|
|
236
237
|
Defaults to None.
|
|
237
|
-
tools (list[Tool] | None, optional): Tools provided to the model to enable tool calling.
|
|
238
|
+
tools (list[Tool | LangChainTool] | None, optional): Tools provided to the model to enable tool calling.
|
|
239
|
+
Defaults to None.
|
|
238
240
|
response_schema (ResponseSchema | None, optional): The schema of the response. If provided, the model will
|
|
239
241
|
output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
|
|
240
242
|
dictionary. Defaults to None.
|
|
@@ -2,11 +2,12 @@ import abc
|
|
|
2
2
|
from _typeshed import Incomplete
|
|
3
3
|
from abc import ABC
|
|
4
4
|
from gllm_core.event import EventEmitter as EventEmitter
|
|
5
|
-
from gllm_core.
|
|
5
|
+
from gllm_core.schema.tool import Tool
|
|
6
|
+
from gllm_core.utils import RetryConfig
|
|
6
7
|
from gllm_inference.constants import DOCUMENT_MIME_TYPES as DOCUMENT_MIME_TYPES, INVOKER_DEFAULT_TIMEOUT as INVOKER_DEFAULT_TIMEOUT
|
|
7
8
|
from gllm_inference.exceptions import parse_error_message as parse_error_message
|
|
8
9
|
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EmitDataType as EmitDataType, LMOutput as LMOutput, Message as Message, MessageContent as MessageContent, MessageRole as MessageRole, ModelId as ModelId, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ToolCall as ToolCall, ToolResult as ToolResult
|
|
9
|
-
from langchain_core.tools import Tool as
|
|
10
|
+
from langchain_core.tools import Tool as LangChainTool
|
|
10
11
|
from typing import Any
|
|
11
12
|
|
|
12
13
|
class _Key:
|
|
@@ -16,8 +17,14 @@ class _Key:
|
|
|
16
17
|
DATA_TYPE: str
|
|
17
18
|
DATA_VALUE: str
|
|
18
19
|
DEFAULT: str
|
|
20
|
+
DESCRIPTION: str
|
|
21
|
+
FUNCTION: str
|
|
22
|
+
META: str
|
|
23
|
+
NAME: str
|
|
24
|
+
PARAMETERS: str
|
|
19
25
|
PROPERTIES: str
|
|
20
26
|
REQUIRED: str
|
|
27
|
+
TITLE: str
|
|
21
28
|
TYPE: str
|
|
22
29
|
|
|
23
30
|
class _InputType:
|
|
@@ -46,7 +53,7 @@ class BaseLMInvoker(ABC, metaclass=abc.ABCMeta):
|
|
|
46
53
|
response_schema: Incomplete
|
|
47
54
|
output_analytics: Incomplete
|
|
48
55
|
retry_config: Incomplete
|
|
49
|
-
def __init__(self, model_id: ModelId, default_hyperparameters: dict[str, Any] | None = None, supported_attachments: set[str] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None) -> None:
|
|
56
|
+
def __init__(self, model_id: ModelId, default_hyperparameters: dict[str, Any] | None = None, supported_attachments: set[str] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None) -> None:
|
|
50
57
|
"""Initializes a new instance of the BaseLMInvoker class.
|
|
51
58
|
|
|
52
59
|
Args:
|
|
@@ -55,7 +62,7 @@ class BaseLMInvoker(ABC, metaclass=abc.ABCMeta):
|
|
|
55
62
|
language model. Defaults to None, in which case an empty dictionary is used.
|
|
56
63
|
supported_attachments (set[str] | None, optional): A set of supported attachment types. Defaults to None,
|
|
57
64
|
in which case an empty set is used (indicating that no attachments are supported).
|
|
58
|
-
tools (list[Tool] | None, optional): Tools provided to the
|
|
65
|
+
tools (list[Tool | LangChainTool] | None, optional): Tools provided to the model to enable tool calling.
|
|
59
66
|
Defaults to None, in which case an empty list is used.
|
|
60
67
|
response_schema (ResponseSchema | None, optional): The schema of the response. If provided, the model will
|
|
61
68
|
output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
|
|
@@ -85,13 +92,13 @@ class BaseLMInvoker(ABC, metaclass=abc.ABCMeta):
|
|
|
85
92
|
Returns:
|
|
86
93
|
str: The name of the language model.
|
|
87
94
|
"""
|
|
88
|
-
def set_tools(self, tools: list[Tool]) -> None:
|
|
95
|
+
def set_tools(self, tools: list[Tool | LangChainTool]) -> None:
|
|
89
96
|
"""Sets the tools for the language model.
|
|
90
97
|
|
|
91
98
|
This method sets the tools for the language model. Any existing tools will be replaced.
|
|
92
99
|
|
|
93
100
|
Args:
|
|
94
|
-
tools (list[Tool]): The list of tools to be used.
|
|
101
|
+
tools (list[Tool | LangChainTool]): The list of tools to be used.
|
|
95
102
|
"""
|
|
96
103
|
def clear_tools(self) -> None:
|
|
97
104
|
"""Clears the tools for the language model.
|
|
@@ -1,11 +1,13 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.event import EventEmitter as EventEmitter
|
|
3
|
+
from gllm_core.schema.tool import Tool as Tool
|
|
3
4
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
5
|
+
from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
4
6
|
from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
|
|
5
7
|
from gllm_inference.lm_invoker.schema.openai_compatible import InputType as InputType, Key as Key, ReasoningEffort as ReasoningEffort
|
|
6
8
|
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EmitDataType as EmitDataType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
|
|
7
9
|
from gllm_inference.utils import validate_string_enum as validate_string_enum
|
|
8
|
-
from langchain_core.tools import Tool as
|
|
10
|
+
from langchain_core.tools import Tool as LangChainTool
|
|
9
11
|
from typing import Any
|
|
10
12
|
|
|
11
13
|
SUPPORTED_ATTACHMENTS: Incomplete
|
|
@@ -19,7 +21,7 @@ class OpenAICompatibleLMInvoker(BaseLMInvoker):
|
|
|
19
21
|
model_name (str): The name of the language model.
|
|
20
22
|
client (AsyncOpenAI): The OpenAI client instance.
|
|
21
23
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
|
|
22
|
-
tools (list[
|
|
24
|
+
tools (list[Tool]): The list of tools provided to the model to enable tool calling.
|
|
23
25
|
response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
|
|
24
26
|
structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema dictionary.
|
|
25
27
|
output_analytics (bool): Whether to output the invocation analytics.
|
|
@@ -62,7 +64,7 @@ class OpenAICompatibleLMInvoker(BaseLMInvoker):
|
|
|
62
64
|
|
|
63
65
|
Tool calling:
|
|
64
66
|
Tool calling is a feature that allows the language model to call tools to perform tasks.
|
|
65
|
-
Tools can be passed to the via the `tools` parameter as a list of
|
|
67
|
+
Tools can be passed to the via the `tools` parameter as a list of `Tool` objects.
|
|
66
68
|
When tools are provided and the model decides to call a tool, the tool calls are stored in the
|
|
67
69
|
`tool_calls` attribute in the output.
|
|
68
70
|
|
|
@@ -230,7 +232,7 @@ class OpenAICompatibleLMInvoker(BaseLMInvoker):
|
|
|
230
232
|
Defaults to an empty list.
|
|
231
233
|
'''
|
|
232
234
|
client: Incomplete
|
|
233
|
-
def __init__(self, model_name: str, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None) -> None:
|
|
235
|
+
def __init__(self, model_name: str, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None) -> None:
|
|
234
236
|
'''Initializes a new instance of the OpenAICompatibleLMInvoker class.
|
|
235
237
|
|
|
236
238
|
Args:
|
|
@@ -242,7 +244,7 @@ class OpenAICompatibleLMInvoker(BaseLMInvoker):
|
|
|
242
244
|
model_kwargs (dict[str, Any] | None, optional): Additional model parameters. Defaults to None.
|
|
243
245
|
default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
|
|
244
246
|
Defaults to None.
|
|
245
|
-
tools (list[Tool] | None, optional): Tools provided to the
|
|
247
|
+
tools (list[Tool | LangChainTool] | None, optional): Tools provided to the model to enable tool calling.
|
|
246
248
|
Defaults to None.
|
|
247
249
|
response_schema (ResponseSchema | None, optional): The schema of the response. If provided, the model will
|
|
248
250
|
output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
|
|
@@ -1,11 +1,13 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from gllm_core.event import EventEmitter as EventEmitter
|
|
3
|
+
from gllm_core.schema.tool import Tool as Tool
|
|
3
4
|
from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
5
|
+
from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
4
6
|
from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
|
|
5
7
|
from gllm_inference.lm_invoker.schema.openai import InputType as InputType, Key as Key, OutputType as OutputType, ReasoningEffort as ReasoningEffort, ReasoningSummary as ReasoningSummary
|
|
6
8
|
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, CodeExecResult as CodeExecResult, EmitDataType as EmitDataType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
|
|
7
9
|
from gllm_inference.utils import validate_string_enum as validate_string_enum
|
|
8
|
-
from langchain_core.tools import Tool as
|
|
10
|
+
from langchain_core.tools import Tool as LangChainTool
|
|
9
11
|
from typing import Any
|
|
10
12
|
|
|
11
13
|
SUPPORTED_ATTACHMENTS: Incomplete
|
|
@@ -19,7 +21,7 @@ class OpenAILMInvoker(BaseLMInvoker):
|
|
|
19
21
|
model_name (str): The name of the language model.
|
|
20
22
|
client (AsyncOpenAI): The OpenAI client instance.
|
|
21
23
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
|
|
22
|
-
tools (list[
|
|
24
|
+
tools (list[Tool]): The list of tools provided to the model to enable tool calling.
|
|
23
25
|
response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
|
|
24
26
|
structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema dictionary.
|
|
25
27
|
output_analytics (bool): Whether to output the invocation analytics.
|
|
@@ -51,7 +53,7 @@ class OpenAILMInvoker(BaseLMInvoker):
|
|
|
51
53
|
|
|
52
54
|
Tool calling:
|
|
53
55
|
Tool calling is a feature that allows the language model to call tools to perform tasks.
|
|
54
|
-
Tools can be passed to the via the `tools` parameter as a list of
|
|
56
|
+
Tools can be passed to the via the `tools` parameter as a list of `Tool` objects.
|
|
55
57
|
When tools are provided and the model decides to call a tool, the tool calls are stored in the
|
|
56
58
|
`tool_calls` attribute in the output.
|
|
57
59
|
|
|
@@ -318,7 +320,7 @@ class OpenAILMInvoker(BaseLMInvoker):
|
|
|
318
320
|
enabled and the language model decides to execute any codes. Defaults to an empty list.
|
|
319
321
|
'''
|
|
320
322
|
client: Incomplete
|
|
321
|
-
def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None, code_interpreter: bool = False, web_search: bool = False) -> None:
|
|
323
|
+
def __init__(self, model_name: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None, code_interpreter: bool = False, web_search: bool = False) -> None:
|
|
322
324
|
"""Initializes a new instance of the OpenAILMInvoker class.
|
|
323
325
|
|
|
324
326
|
Args:
|
|
@@ -328,8 +330,8 @@ class OpenAILMInvoker(BaseLMInvoker):
|
|
|
328
330
|
model_kwargs (dict[str, Any] | None, optional): Additional model parameters. Defaults to None.
|
|
329
331
|
default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
|
|
330
332
|
Defaults to None.
|
|
331
|
-
tools (list[Tool] | None, optional): Tools provided to the
|
|
332
|
-
Defaults to None.
|
|
333
|
+
tools (list[Tool | LangChainTool] | None, optional): Tools provided to the model to enable tool calling.
|
|
334
|
+
Defaults to None, in which case an empty list is used.
|
|
333
335
|
response_schema (ResponseSchema | None, optional): The schema of the response. If provided, the model will
|
|
334
336
|
output a structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema
|
|
335
337
|
dictionary. Defaults to None.
|
|
@@ -4,10 +4,10 @@ class Key:
|
|
|
4
4
|
CONTENT: str
|
|
5
5
|
DATA: str
|
|
6
6
|
DESCRIPTION: str
|
|
7
|
-
FUNCTION: str
|
|
8
7
|
ID: str
|
|
9
8
|
INPUT: str
|
|
10
9
|
INPUT_SCHEMA: str
|
|
10
|
+
MAX_RETRIES: str
|
|
11
11
|
MEDIA_TYPE: str
|
|
12
12
|
MAX_TOKENS: str
|
|
13
13
|
NAME: str
|
|
@@ -17,6 +17,7 @@ class Key:
|
|
|
17
17
|
SOURCE: str
|
|
18
18
|
STOP_REASON: str
|
|
19
19
|
SYSTEM: str
|
|
20
|
+
TIMEOUT: str
|
|
20
21
|
THINKING: str
|
|
21
22
|
TOOLS: str
|
|
22
23
|
TOOL_CHOICE: str
|
|
@@ -6,9 +6,12 @@ class Key:
|
|
|
6
6
|
FINISH_REASON: str
|
|
7
7
|
FUNCTION: str
|
|
8
8
|
FUNCTION_CALL: str
|
|
9
|
+
HTTP_OPTIONS: str
|
|
9
10
|
NAME: str
|
|
11
|
+
RETRY_OPTIONS: str
|
|
10
12
|
SYSTEM_INSTRUCTION: str
|
|
11
13
|
THINKING_CONFIG: str
|
|
14
|
+
TIMEOUT: str
|
|
12
15
|
TOOLS: str
|
|
13
16
|
RESPONSE_SCHEMA: str
|
|
14
17
|
RESPONSE_MIME_TYPE: str
|
|
@@ -7,6 +7,7 @@ class Key:
|
|
|
7
7
|
CHOICES: str
|
|
8
8
|
DATA: str
|
|
9
9
|
DEFS: str
|
|
10
|
+
DESCRIPTION: str
|
|
10
11
|
EFFORT: str
|
|
11
12
|
FINISH_REASON: str
|
|
12
13
|
FORMAT: str
|
|
@@ -15,13 +16,16 @@ class Key:
|
|
|
15
16
|
IMAGE_URL: str
|
|
16
17
|
INPUT_AUDIO: str
|
|
17
18
|
JSON_SCHEMA: str
|
|
19
|
+
MAX_RETRIES: str
|
|
18
20
|
MESSAGE: str
|
|
19
21
|
NAME: str
|
|
22
|
+
PARAMETERS: str
|
|
20
23
|
RESPONSE_FORMAT: str
|
|
21
24
|
ROLE: str
|
|
22
25
|
SCHEMA: str
|
|
23
26
|
STRICT: str
|
|
24
27
|
TEXT: str
|
|
28
|
+
TIMEOUT: str
|
|
25
29
|
TITLE: str
|
|
26
30
|
TOOLS: str
|
|
27
31
|
TOOL_CALLS: str
|
|
Binary file
|
gllm_inference.pyi
CHANGED
|
@@ -59,6 +59,7 @@ import langchain_core.embeddings
|
|
|
59
59
|
import gllm_inference.utils.load_langchain_model
|
|
60
60
|
import gllm_inference.utils.parse_model_data
|
|
61
61
|
import io
|
|
62
|
+
import httpx
|
|
62
63
|
import twelvelabs
|
|
63
64
|
import base64
|
|
64
65
|
import sys
|
|
@@ -70,14 +71,13 @@ import enum
|
|
|
70
71
|
import http
|
|
71
72
|
import http.HTTPStatus
|
|
72
73
|
import aiohttp
|
|
73
|
-
import httpx
|
|
74
74
|
import requests
|
|
75
75
|
import gllm_inference.schema.ErrorResponse
|
|
76
76
|
import gllm_core.constants
|
|
77
77
|
import gllm_core.event
|
|
78
|
+
import gllm_core.schema
|
|
79
|
+
import gllm_core.schema.tool
|
|
78
80
|
import langchain_core.tools
|
|
79
|
-
import langchain_core.utils
|
|
80
|
-
import langchain_core.utils.function_calling
|
|
81
81
|
import gllm_inference.schema.EmitDataType
|
|
82
82
|
import gllm_inference.schema.LMOutput
|
|
83
83
|
import gllm_inference.schema.Message
|
|
@@ -88,13 +88,14 @@ import gllm_inference.schema.ToolCall
|
|
|
88
88
|
import gllm_inference.schema.ToolResult
|
|
89
89
|
import anthropic
|
|
90
90
|
import aioboto3
|
|
91
|
-
import gllm_core.schema
|
|
92
91
|
import gllm_inference.schema.MessageRole
|
|
93
92
|
import langchain_core.language_models
|
|
94
93
|
import langchain_core.messages
|
|
95
94
|
import litellm
|
|
96
95
|
import time
|
|
97
96
|
import jsonschema
|
|
97
|
+
import langchain_core.utils
|
|
98
|
+
import langchain_core.utils.function_calling
|
|
98
99
|
import gllm_inference.schema.MessageContent
|
|
99
100
|
import gllm_inference.utils.validate_string_enum
|
|
100
101
|
import gllm_inference.schema.CodeExecResult
|
|
@@ -8,43 +8,47 @@ gllm_inference/catalog/__init__.pyi,sha256=JBkPGTyiiZ30GECzJBW-mW8LekWyY2qyzal3e
|
|
|
8
8
|
gllm_inference/catalog/catalog.pyi,sha256=a4RNG1lKv51GxQpOqh47tz-PAROMPaeP2o5XNLBSZaU,4790
|
|
9
9
|
gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=ranHMbG9--DZj9FJRhIUa6U8e-L-Tm-_hSBpzJ6DDs4,5428
|
|
10
10
|
gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=OU8k_4HbqjZEzHZlzSM3uzGQZJmM2uGD76Csqom0CEQ,3197
|
|
11
|
-
gllm_inference/constants.pyi,sha256=
|
|
11
|
+
gllm_inference/constants.pyi,sha256=gPlwRHKIjUdyQoLdoog8ca76YmjrQL0SCDTEn8UEWSY,256
|
|
12
12
|
gllm_inference/em_invoker/__init__.pyi,sha256=XESsrYo1PZeeHe7AMRyuzKoV7XDD5oN89ZTH01zRf4k,873
|
|
13
|
-
gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=
|
|
13
|
+
gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=OEkVu5nv92ITqdhDtgDg4MiLSDRWDmLSnAhYtXpCn6E,4602
|
|
14
14
|
gllm_inference/em_invoker/em_invoker.pyi,sha256=hiH8FB5R-KxhI8Ds2htF3cjRcIcH92yHPcOdpgc4FDo,4341
|
|
15
|
-
gllm_inference/em_invoker/google_em_invoker.pyi,sha256=
|
|
15
|
+
gllm_inference/em_invoker/google_em_invoker.pyi,sha256=pn05VdOZXZOffeg89wlAQOBAyHyt6dLaS0Sd6LoH05M,6130
|
|
16
16
|
gllm_inference/em_invoker/langchain/__init__.pyi,sha256=VYGKE5OgU0my1RlhgzkU_A7-GLGnUDDnNFuctuRwILE,148
|
|
17
17
|
gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi,sha256=6nASLqi0FXCpqyYPl7kM3g7hAW-xS5ZwsS3GFudns98,2347
|
|
18
|
-
gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=
|
|
19
|
-
gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=
|
|
20
|
-
gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=
|
|
18
|
+
gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=lXiTTGcNOIwurZx3_6vWLE1DQyioK8Z1fOcPuUATnxA,2782
|
|
19
|
+
gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=Qz2Qx1KRKhzXr8IseDWcF_6yC-SNtfsXvQuGuKnqVe8,4978
|
|
20
|
+
gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=SFuS2DsvMHcibxFnpQOOchlZUyNRRlI2uMhVEUfifas,4235
|
|
21
21
|
gllm_inference/em_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
22
|
-
gllm_inference/em_invoker/schema/
|
|
23
|
-
gllm_inference/em_invoker/schema/
|
|
24
|
-
gllm_inference/em_invoker/
|
|
25
|
-
gllm_inference/em_invoker/
|
|
22
|
+
gllm_inference/em_invoker/schema/google.pyi,sha256=MUmgtjMmjSpzmzaAOx6JGZbcdRxgMUhOpvcVQIo-oGs,146
|
|
23
|
+
gllm_inference/em_invoker/schema/langchain.pyi,sha256=onpZutqa2xw2g8rdJTdycy3ub58lkPBVB3KvVVPpyds,92
|
|
24
|
+
gllm_inference/em_invoker/schema/openai.pyi,sha256=Q_dsEcodkOXYXPdrkOkW0LnuLhfeq8tEbtZAGMz2ajA,139
|
|
25
|
+
gllm_inference/em_invoker/schema/openai_compatible.pyi,sha256=gmvGtsWoOMBelke_tZjC6dKimFBW9f4Vrgv0Ig0OM9Q,150
|
|
26
|
+
gllm_inference/em_invoker/schema/twelvelabs.pyi,sha256=F6wKHgG01bYskJpKoheBSpRpHUfFpteKn9sj9n5YfcU,372
|
|
27
|
+
gllm_inference/em_invoker/schema/voyage.pyi,sha256=HVpor0fqNy-IwapCICfsgFmqf1FJXCOMIxS2vOXhHd8,289
|
|
28
|
+
gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=6b8oJ8aLNBn45o85dka-Xbq9nWZqtcMgFls39ekEZ6o,5066
|
|
29
|
+
gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=7akaf8GxOA8Trokad0xmlYKr49rY3Egm3_4gpW-vly8,5134
|
|
26
30
|
gllm_inference/exceptions/__init__.pyi,sha256=v9uxjW5DssIn7n_bKqT7L83CeqFET2Z45GFOvi78UuE,977
|
|
27
31
|
gllm_inference/exceptions/error_parser.pyi,sha256=4RkVfS2Fl9kjz_h2bK9eoAeI-Y-VkHcUqXWj68BsYig,2393
|
|
28
32
|
gllm_inference/exceptions/exceptions.pyi,sha256=5YRackwVNvyOJjOtiVszqu8q87s8ioXTa-XwaYmeiC4,4643
|
|
29
33
|
gllm_inference/lm_invoker/__init__.pyi,sha256=8oUFgavpItvjRUqsLqLSHqQVIPyTKCgNQ_euf58H2zY,1104
|
|
30
|
-
gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=
|
|
31
|
-
gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=
|
|
32
|
-
gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=
|
|
33
|
-
gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=
|
|
34
|
-
gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=
|
|
35
|
-
gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=
|
|
36
|
-
gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=
|
|
37
|
-
gllm_inference/lm_invoker/lm_invoker.pyi,sha256=
|
|
38
|
-
gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=
|
|
39
|
-
gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=
|
|
34
|
+
gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=MsF3OmDo0L9aEHuTJYTgsoDILi2B_IgKtPpDcDMduWc,14925
|
|
35
|
+
gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=EV_yrj6mnV_rCDEqSZaIikfw76_rXXyDlC-w_y-m7K0,14603
|
|
36
|
+
gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=IuLxgCThOSBHx1AXqnhL6yVu5_JV6hAeGBWWm5P1JCo,12423
|
|
37
|
+
gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=QS84w3WpD3Oyl5HdxrucsadCmsHE8gn6Ewl3l01DCgI,9203
|
|
38
|
+
gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=Hy5DYaIaKPqMjSjDyfHy0HofHhogAn3McN7SeDs67g4,16491
|
|
39
|
+
gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=skcbX34ZosGMFli3SLmGITJTzMj9HxkQBWGrd-zYCbU,13264
|
|
40
|
+
gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=_c56ewpEQ-Ywj5ofFzRYBvQgefR7Q_WkcQt97lnIFgg,13128
|
|
41
|
+
gllm_inference/lm_invoker/lm_invoker.pyi,sha256=NLLzJY5YaWckR0DHgNNxZhg-Lif41ZX-DZZcICq7NIU,7809
|
|
42
|
+
gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=HWuoklm6DMvyiFnuq9iBbkR2YU4d5xXMW7-haU9Okt4,14928
|
|
43
|
+
gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=y9IiSvffDVA7cQh24nFZa2qbl-YOBT8A0rBWUcH8I6c,19531
|
|
40
44
|
gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
41
|
-
gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=
|
|
45
|
+
gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=zNbm4RV454dBEEUUN-Vyl_4cO628wUMPZyrO27O_DfM,991
|
|
42
46
|
gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=FOUMZkBi6KRa__mYoy3FNJ5sP0EC6rgLuhoijnwelIg,927
|
|
43
|
-
gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=
|
|
44
|
-
gllm_inference/lm_invoker/schema/google.pyi,sha256=
|
|
45
|
-
gllm_inference/lm_invoker/schema/langchain.pyi,sha256
|
|
46
|
-
gllm_inference/lm_invoker/schema/openai.pyi,sha256=
|
|
47
|
-
gllm_inference/lm_invoker/schema/openai_compatible.pyi,sha256=
|
|
47
|
+
gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=aA4DhTXIezwLvFzphR24a5ueVln2FCBIloP9Hbt3iz4,230
|
|
48
|
+
gllm_inference/lm_invoker/schema/google.pyi,sha256=AIsNgq0ZZuicHmx4bL7z6q-946T05nWts3HUeA8hhHQ,505
|
|
49
|
+
gllm_inference/lm_invoker/schema/langchain.pyi,sha256=l2kHU7S3vmG3-NCt8B26krp_i4Br3waES_CekkgrKSA,409
|
|
50
|
+
gllm_inference/lm_invoker/schema/openai.pyi,sha256=w_XSzOohX_7vnR81hVXdgDbBqk737LmMAN_AvODGk2I,1902
|
|
51
|
+
gllm_inference/lm_invoker/schema/openai_compatible.pyi,sha256=U2Q6F5Jw_-czGdEF0clWcjY3ciNsCL_a46XGLM7OSpk,1101
|
|
48
52
|
gllm_inference/model/__init__.pyi,sha256=qClHIgljqhPPCKlGTKmHsWgYb4_hADybxtC2q1U8a5Q,593
|
|
49
53
|
gllm_inference/model/em/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
50
54
|
gllm_inference/model/em/google_em.pyi,sha256=ZPN5LmReO0bcTfnZixFooUTzgD-daNFPzfxzZ-5WzQQ,471
|
|
@@ -86,8 +90,8 @@ gllm_inference/utils/__init__.pyi,sha256=npmBmmlBv7cPHMg1hdL3S2_RelD6vk_LhCsGELh
|
|
|
86
90
|
gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
|
|
87
91
|
gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
|
|
88
92
|
gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
|
|
89
|
-
gllm_inference.cpython-313-darwin.so,sha256=
|
|
90
|
-
gllm_inference.pyi,sha256=
|
|
91
|
-
gllm_inference_binary-0.5.
|
|
92
|
-
gllm_inference_binary-0.5.
|
|
93
|
-
gllm_inference_binary-0.5.
|
|
93
|
+
gllm_inference.cpython-313-darwin.so,sha256=fQ0LauFXcB_VhD5HnxcdszxW3OdxXJUCIpepROsmfAc,3768944
|
|
94
|
+
gllm_inference.pyi,sha256=2rrMuhWOyLvODZQ7_NT7rt409eCv7iKzZZR3OWJ3KX4,3344
|
|
95
|
+
gllm_inference_binary-0.5.8.dist-info/METADATA,sha256=czkvBDQEqltXhBF08Zq4Ik6KjTwuRDDvbafy9pXvs0Y,4531
|
|
96
|
+
gllm_inference_binary-0.5.8.dist-info/WHEEL,sha256=PCOZcL_jcbAVhuFR5ylE4Mr-7HPGHAcfJk9OpuMh7RQ,107
|
|
97
|
+
gllm_inference_binary-0.5.8.dist-info/RECORD,,
|
|
File without changes
|