gllm-inference-binary 0.5.36__cp312-cp312-macosx_13_0_x86_64.whl → 0.5.38__cp312-cp312-macosx_13_0_x86_64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of gllm-inference-binary might be problematic. Click here for more details.
- gllm_inference/em_invoker/azure_openai_em_invoker.pyi +2 -2
- gllm_inference/em_invoker/openai_compatible_em_invoker.pyi +2 -2
- gllm_inference/em_invoker/openai_em_invoker.pyi +2 -2
- gllm_inference/lm_invoker/anthropic_lm_invoker.pyi +1 -1
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi +2 -2
- gllm_inference/lm_invoker/datasaur_lm_invoker.pyi +2 -2
- gllm_inference/lm_invoker/google_lm_invoker.pyi +1 -1
- gllm_inference/lm_invoker/lm_invoker.pyi +3 -2
- gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi +3 -3
- gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi +1 -1
- gllm_inference/lm_invoker/openai_lm_invoker.pyi +3 -3
- gllm_inference/lm_invoker/schema/datasaur.pyi +2 -0
- gllm_inference/lm_invoker/schema/openai.pyi +2 -0
- gllm_inference/lm_invoker/schema/openai_chat_completions.pyi +2 -0
- gllm_inference/lm_invoker/xai_lm_invoker.pyi +1 -1
- gllm_inference/schema/__init__.pyi +3 -3
- gllm_inference/schema/activity.pyi +55 -2
- gllm_inference/schema/enums.pyi +17 -0
- gllm_inference/schema/events.pyi +89 -20
- gllm_inference.cpython-312-darwin.so +0 -0
- gllm_inference.pyi +8 -2
- {gllm_inference_binary-0.5.36.dist-info → gllm_inference_binary-0.5.38.dist-info}/METADATA +1 -1
- {gllm_inference_binary-0.5.36.dist-info → gllm_inference_binary-0.5.38.dist-info}/RECORD +25 -25
- {gllm_inference_binary-0.5.36.dist-info → gllm_inference_binary-0.5.38.dist-info}/WHEEL +0 -0
- {gllm_inference_binary-0.5.36.dist-info → gllm_inference_binary-0.5.38.dist-info}/top_level.txt +0 -0
|
@@ -13,7 +13,7 @@ class AzureOpenAIEMInvoker(OpenAIEMInvoker):
|
|
|
13
13
|
model_id (str): The model ID of the embedding model.
|
|
14
14
|
model_provider (str): The provider of the embedding model.
|
|
15
15
|
model_name (str): The name of the Azure OpenAI embedding model deployment.
|
|
16
|
-
|
|
16
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the Azure OpenAI client.
|
|
17
17
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the embedding model.
|
|
18
18
|
retry_config (RetryConfig): The retry configuration for the embedding model.
|
|
19
19
|
truncation_config (TruncationConfig | None): The truncation configuration for the embedding model.
|
|
@@ -68,7 +68,7 @@ class AzureOpenAIEMInvoker(OpenAIEMInvoker):
|
|
|
68
68
|
em_invoker = AzureOpenAIEMInvoker(..., retry_config=retry_config)
|
|
69
69
|
```
|
|
70
70
|
'''
|
|
71
|
-
|
|
71
|
+
client_kwargs: Incomplete
|
|
72
72
|
def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
|
|
73
73
|
"""Initializes a new instance of the AzureOpenAIEMInvoker class.
|
|
74
74
|
|
|
@@ -15,14 +15,14 @@ class OpenAICompatibleEMInvoker(OpenAIEMInvoker):
|
|
|
15
15
|
model_id (str): The model ID of the embedding model.
|
|
16
16
|
model_provider (str): The provider of the embedding model.
|
|
17
17
|
model_name (str): The name of the embedding model.
|
|
18
|
-
|
|
18
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
|
|
19
19
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the embedding model.
|
|
20
20
|
retry_config (RetryConfig): The retry configuration for the embedding model.
|
|
21
21
|
truncation_config (TruncationConfig | None): The truncation configuration for the embedding model.
|
|
22
22
|
|
|
23
23
|
This class is deprecated and will be removed in v0.6. Please use the `OpenAIEMInvoker` class instead.
|
|
24
24
|
"""
|
|
25
|
-
|
|
25
|
+
client_kwargs: Incomplete
|
|
26
26
|
def __init__(self, model_name: str, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
|
|
27
27
|
"""Initializes a new instance of the OpenAICompatibleEMInvoker class.
|
|
28
28
|
|
|
@@ -16,7 +16,7 @@ class OpenAIEMInvoker(BaseEMInvoker):
|
|
|
16
16
|
model_id (str): The model ID of the embedding model.
|
|
17
17
|
model_provider (str): The provider of the embedding model.
|
|
18
18
|
model_name (str): The name of the embedding model.
|
|
19
|
-
|
|
19
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
|
|
20
20
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the embedding model.
|
|
21
21
|
retry_config (RetryConfig): The retry configuration for the embedding model.
|
|
22
22
|
truncation_config (TruncationConfig | None): The truncation configuration for the embedding model.
|
|
@@ -96,7 +96,7 @@ class OpenAIEMInvoker(BaseEMInvoker):
|
|
|
96
96
|
em_invoker = OpenAIEMInvoker(..., retry_config=retry_config)
|
|
97
97
|
```
|
|
98
98
|
'''
|
|
99
|
-
|
|
99
|
+
client_kwargs: Incomplete
|
|
100
100
|
def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
|
|
101
101
|
'''Initializes a new instance of the OpenAIEMInvoker class.
|
|
102
102
|
|
|
@@ -5,7 +5,7 @@ from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
|
5
5
|
from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
|
|
6
6
|
from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
|
|
7
7
|
from gllm_inference.lm_invoker.schema.anthropic import InputType as InputType, Key as Key, OutputType as OutputType
|
|
8
|
-
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, BatchStatus as BatchStatus,
|
|
8
|
+
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, BatchStatus as BatchStatus, LMInput as LMInput, LMOutput as LMOutput, Message as Message, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ThinkingEvent as ThinkingEvent, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
|
|
9
9
|
from langchain_core.tools import Tool as LangChainTool
|
|
10
10
|
from typing import Any
|
|
11
11
|
|
|
@@ -15,7 +15,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
|
|
|
15
15
|
model_id (str): The model ID of the language model.
|
|
16
16
|
model_provider (str): The provider of the language model.
|
|
17
17
|
model_name (str): The name of the Azure OpenAI language model deployment.
|
|
18
|
-
|
|
18
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the Azure OpenAI client.
|
|
19
19
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
|
|
20
20
|
tools (list[Tool]): The list of tools provided to the model to enable tool calling.
|
|
21
21
|
response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
|
|
@@ -226,7 +226,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
|
|
|
226
226
|
Defaults to an empty list.
|
|
227
227
|
2.10. mcp_calls (list[MCPCall]): The MCP calls. Currently not supported. Defaults to an empty list.
|
|
228
228
|
'''
|
|
229
|
-
|
|
229
|
+
client_kwargs: Incomplete
|
|
230
230
|
def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None) -> None:
|
|
231
231
|
"""Initializes a new instance of the AzureOpenAILMInvoker class.
|
|
232
232
|
|
|
@@ -18,7 +18,7 @@ class DatasaurLMInvoker(OpenAICompatibleLMInvoker):
|
|
|
18
18
|
model_id (str): The model ID of the language model.
|
|
19
19
|
model_provider (str): The provider of the language model.
|
|
20
20
|
model_name (str): The name of the language model.
|
|
21
|
-
|
|
21
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
|
|
22
22
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
|
|
23
23
|
tools (list[Any]): The list of tools provided to the model to enable tool calling. Currently not supported.
|
|
24
24
|
response_schema (ResponseSchema | None): The schema of the response. Currently not supported.
|
|
@@ -121,7 +121,7 @@ class DatasaurLMInvoker(OpenAICompatibleLMInvoker):
|
|
|
121
121
|
Defaults to an empty list.
|
|
122
122
|
2.10. mcp_calls (list[MCPCall]): The MCP calls. Currently not supported. Defaults to an empty list.
|
|
123
123
|
'''
|
|
124
|
-
|
|
124
|
+
client_kwargs: Incomplete
|
|
125
125
|
citations: Incomplete
|
|
126
126
|
def __init__(self, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, citations: bool = False) -> None:
|
|
127
127
|
"""Initializes a new instance of the DatasaurLMInvoker class.
|
|
@@ -7,7 +7,7 @@ from gllm_inference.exceptions import BaseInvokerError as BaseInvokerError, conv
|
|
|
7
7
|
from gllm_inference.exceptions.provider_error_map import GOOGLE_ERROR_MAPPING as GOOGLE_ERROR_MAPPING
|
|
8
8
|
from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
|
|
9
9
|
from gllm_inference.lm_invoker.schema.google import InputType as InputType, Key as Key
|
|
10
|
-
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType,
|
|
10
|
+
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ThinkingEvent as ThinkingEvent, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
|
|
11
11
|
from langchain_core.tools import Tool as LangChainTool
|
|
12
12
|
from typing import Any
|
|
13
13
|
|
|
@@ -2,12 +2,12 @@ import abc
|
|
|
2
2
|
from _typeshed import Incomplete
|
|
3
3
|
from abc import ABC
|
|
4
4
|
from gllm_core.event import EventEmitter as EventEmitter
|
|
5
|
-
from gllm_core.schema
|
|
5
|
+
from gllm_core.schema import Event as Event, Tool
|
|
6
6
|
from gllm_core.utils import RetryConfig
|
|
7
7
|
from gllm_inference.constants import DOCUMENT_MIME_TYPES as DOCUMENT_MIME_TYPES, INVOKER_DEFAULT_TIMEOUT as INVOKER_DEFAULT_TIMEOUT
|
|
8
8
|
from gllm_inference.exceptions import BaseInvokerError as BaseInvokerError, convert_to_base_invoker_error as convert_to_base_invoker_error
|
|
9
9
|
from gllm_inference.lm_invoker.batch import BatchOperations as BatchOperations
|
|
10
|
-
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, BatchStatus as BatchStatus,
|
|
10
|
+
from gllm_inference.schema import Activity as Activity, Attachment as Attachment, AttachmentType as AttachmentType, BatchStatus as BatchStatus, LMInput as LMInput, LMOutput as LMOutput, Message as Message, MessageContent as MessageContent, MessageRole as MessageRole, ModelId as ModelId, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ToolCall as ToolCall, ToolResult as ToolResult
|
|
11
11
|
from langchain_core.tools import Tool as LangChainTool
|
|
12
12
|
from typing import Any
|
|
13
13
|
|
|
@@ -23,6 +23,7 @@ class Key:
|
|
|
23
23
|
DEFAULT: str
|
|
24
24
|
DESCRIPTION: str
|
|
25
25
|
FUNC: str
|
|
26
|
+
ID: str
|
|
26
27
|
NAME: str
|
|
27
28
|
PROPERTIES: str
|
|
28
29
|
REQUIRED: str
|
|
@@ -5,7 +5,7 @@ from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
|
5
5
|
from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES, OPENAI_DEFAULT_URL as OPENAI_DEFAULT_URL
|
|
6
6
|
from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
|
|
7
7
|
from gllm_inference.lm_invoker.schema.openai_chat_completions import InputType as InputType, Key as Key, ReasoningEffort as ReasoningEffort
|
|
8
|
-
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType,
|
|
8
|
+
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ThinkingEvent as ThinkingEvent, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
|
|
9
9
|
from gllm_inference.utils import validate_string_enum as validate_string_enum
|
|
10
10
|
from langchain_core.tools import Tool as LangChainTool
|
|
11
11
|
from typing import Any
|
|
@@ -23,7 +23,7 @@ class OpenAIChatCompletionsLMInvoker(BaseLMInvoker):
|
|
|
23
23
|
model_id (str): The model ID of the language model.
|
|
24
24
|
model_provider (str): The provider of the language model.
|
|
25
25
|
model_name (str): The name of the language model.
|
|
26
|
-
|
|
26
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
|
|
27
27
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
|
|
28
28
|
tools (list[Tool]): The list of tools provided to the model to enable tool calling.
|
|
29
29
|
response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
|
|
@@ -243,7 +243,7 @@ class OpenAIChatCompletionsLMInvoker(BaseLMInvoker):
|
|
|
243
243
|
Defaults to an empty list.
|
|
244
244
|
2.10. mcp_calls (list[MCPCall]): The MCP calls. Currently not supported. Defaults to an empty list.
|
|
245
245
|
'''
|
|
246
|
-
|
|
246
|
+
client_kwargs: Incomplete
|
|
247
247
|
def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None) -> None:
|
|
248
248
|
'''Initializes a new instance of the OpenAIChatCompletionsLMInvoker class.
|
|
249
249
|
|
|
@@ -15,7 +15,7 @@ class OpenAICompatibleLMInvoker(OpenAIChatCompletionsLMInvoker):
|
|
|
15
15
|
model_id (str): The model ID of the language model.
|
|
16
16
|
model_provider (str): The provider of the language model.
|
|
17
17
|
model_name (str): The name of the language model.
|
|
18
|
-
|
|
18
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
|
|
19
19
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
|
|
20
20
|
tools (list[Tool]): The list of tools provided to the model to enable tool calling.
|
|
21
21
|
response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
|
|
@@ -5,7 +5,7 @@ from gllm_core.utils.retry import RetryConfig as RetryConfig
|
|
|
5
5
|
from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES, OPENAI_DEFAULT_URL as OPENAI_DEFAULT_URL
|
|
6
6
|
from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
|
|
7
7
|
from gllm_inference.lm_invoker.schema.openai import InputType as InputType, Key as Key, OutputType as OutputType, ReasoningEffort as ReasoningEffort, ReasoningSummary as ReasoningSummary
|
|
8
|
-
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType,
|
|
8
|
+
from gllm_inference.schema import ActivityEvent as ActivityEvent, Attachment as Attachment, AttachmentType as AttachmentType, CodeEvent as CodeEvent, CodeExecResult as CodeExecResult, LMOutput as LMOutput, MCPCall as MCPCall, MCPCallActivity as MCPCallActivity, MCPListToolsActivity as MCPListToolsActivity, MCPServer as MCPServer, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ThinkingEvent as ThinkingEvent, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult, WebSearchActivity as WebSearchActivity
|
|
9
9
|
from gllm_inference.utils import validate_string_enum as validate_string_enum
|
|
10
10
|
from langchain_core.tools import Tool as LangChainTool
|
|
11
11
|
from typing import Any
|
|
@@ -25,7 +25,7 @@ class OpenAILMInvoker(BaseLMInvoker):
|
|
|
25
25
|
model_id (str): The model ID of the language model.
|
|
26
26
|
model_provider (str): The provider of the language model.
|
|
27
27
|
model_name (str): The name of the language model.
|
|
28
|
-
|
|
28
|
+
client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
|
|
29
29
|
default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
|
|
30
30
|
tools (list[Tool]): The list of tools provided to the model to enable tool calling.
|
|
31
31
|
response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
|
|
@@ -391,7 +391,7 @@ class OpenAILMInvoker(BaseLMInvoker):
|
|
|
391
391
|
2.10. mcp_calls (list[MCPCall]): The MCP calls, if the MCP servers are provided and the language model
|
|
392
392
|
decides to invoke MCP tools. Defaults to an empty list.
|
|
393
393
|
'''
|
|
394
|
-
|
|
394
|
+
client_kwargs: Incomplete
|
|
395
395
|
def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None, mcp_servers: list[MCPServer] | None = None, code_interpreter: bool = False, web_search: bool = False) -> None:
|
|
396
396
|
'''Initializes a new instance of the OpenAILMInvoker class.
|
|
397
397
|
|
|
@@ -7,7 +7,7 @@ from gllm_inference.exceptions import BaseInvokerError as BaseInvokerError, Invo
|
|
|
7
7
|
from gllm_inference.exceptions.provider_error_map import GRPC_STATUS_CODE_MAPPING as GRPC_STATUS_CODE_MAPPING
|
|
8
8
|
from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
|
|
9
9
|
from gllm_inference.lm_invoker.schema.xai import Key as Key, ReasoningEffort as ReasoningEffort
|
|
10
|
-
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType,
|
|
10
|
+
from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ThinkingEvent as ThinkingEvent, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
|
|
11
11
|
from gllm_inference.utils.validation import validate_string_enum as validate_string_enum
|
|
12
12
|
from langchain_core.tools import Tool as LangChainTool
|
|
13
13
|
from typing import Any
|
|
@@ -1,9 +1,9 @@
|
|
|
1
|
-
from gllm_inference.schema.activity import Activity as Activity
|
|
1
|
+
from gllm_inference.schema.activity import Activity as Activity, MCPCallActivity as MCPCallActivity, MCPListToolsActivity as MCPListToolsActivity, WebSearchActivity as WebSearchActivity
|
|
2
2
|
from gllm_inference.schema.attachment import Attachment as Attachment
|
|
3
3
|
from gllm_inference.schema.code_exec_result import CodeExecResult as CodeExecResult
|
|
4
4
|
from gllm_inference.schema.config import TruncationConfig as TruncationConfig
|
|
5
5
|
from gllm_inference.schema.enums import AttachmentType as AttachmentType, BatchStatus as BatchStatus, EmitDataType as EmitDataType, MessageRole as MessageRole, TruncateSide as TruncateSide
|
|
6
|
-
from gllm_inference.schema.events import ActivityEvent as ActivityEvent, CodeEvent as CodeEvent,
|
|
6
|
+
from gllm_inference.schema.events import ActivityEvent as ActivityEvent, CodeEvent as CodeEvent, ThinkingEvent as ThinkingEvent
|
|
7
7
|
from gllm_inference.schema.lm_input import LMInput as LMInput
|
|
8
8
|
from gllm_inference.schema.lm_output import LMOutput as LMOutput
|
|
9
9
|
from gllm_inference.schema.mcp import MCPCall as MCPCall, MCPServer as MCPServer
|
|
@@ -15,4 +15,4 @@ from gllm_inference.schema.tool_call import ToolCall as ToolCall
|
|
|
15
15
|
from gllm_inference.schema.tool_result import ToolResult as ToolResult
|
|
16
16
|
from gllm_inference.schema.type_alias import EMContent as EMContent, MessageContent as MessageContent, ResponseSchema as ResponseSchema, Vector as Vector
|
|
17
17
|
|
|
18
|
-
__all__ = ['Activity', 'ActivityEvent', 'Attachment', 'AttachmentType', 'BatchStatus', 'CodeEvent', 'CodeExecResult', 'EMContent', 'EmitDataType', '
|
|
18
|
+
__all__ = ['Activity', 'ActivityEvent', 'Attachment', 'AttachmentType', 'BatchStatus', 'CodeEvent', 'CodeExecResult', 'EMContent', 'EmitDataType', 'InputTokenDetails', 'LMInput', 'LMOutput', 'MCPCall', 'MCPCallActivity', 'MCPListToolsActivity', 'MCPServer', 'Message', 'MessageContent', 'MessageRole', 'ModelId', 'ModelProvider', 'OutputTokenDetails', 'Reasoning', 'ThinkingEvent', 'ResponseSchema', 'TokenUsage', 'ToolCall', 'ToolResult', 'TruncateSide', 'TruncationConfig', 'Vector', 'WebSearchActivity']
|
|
@@ -1,9 +1,62 @@
|
|
|
1
|
+
from _typeshed import Incomplete
|
|
2
|
+
from gllm_inference.schema.enums import ActivityType as ActivityType, WebSearchKey as WebSearchKey
|
|
1
3
|
from pydantic import BaseModel
|
|
4
|
+
from typing import Literal
|
|
5
|
+
|
|
6
|
+
WEB_SEARCH_VISIBLE_FIELDS: Incomplete
|
|
2
7
|
|
|
3
8
|
class Activity(BaseModel):
|
|
4
9
|
"""Base schema for any activity.
|
|
5
10
|
|
|
6
11
|
Attributes:
|
|
7
|
-
|
|
12
|
+
type (str): The type of activity being performed.
|
|
13
|
+
"""
|
|
14
|
+
type: str
|
|
15
|
+
|
|
16
|
+
class MCPListToolsActivity(Activity):
|
|
17
|
+
"""Schema for listing tools in MCP.
|
|
18
|
+
|
|
19
|
+
Attributes:
|
|
20
|
+
server_name (str): The name of the MCP server.
|
|
21
|
+
tools (list[dict[str, str]] | None): The tools in the MCP server.
|
|
22
|
+
type (str): The type of activity being performed.
|
|
23
|
+
"""
|
|
24
|
+
type: Literal[ActivityType.MCP_LIST_TOOLS]
|
|
25
|
+
server_name: str
|
|
26
|
+
tools: list[dict[str, str]] | None
|
|
27
|
+
|
|
28
|
+
class MCPCallActivity(Activity):
|
|
29
|
+
"""Schema for MCP tool call.
|
|
30
|
+
|
|
31
|
+
Attributes:
|
|
32
|
+
server_name (str): The name of the MCP server.
|
|
33
|
+
tool_name (str): The name of the tool.
|
|
34
|
+
args (dict[str, str]): The arguments of the tool.
|
|
35
|
+
type (str): The type of activity being performed.
|
|
8
36
|
"""
|
|
9
|
-
|
|
37
|
+
type: Literal[ActivityType.MCP_CALL]
|
|
38
|
+
server_name: str
|
|
39
|
+
tool_name: str
|
|
40
|
+
args: dict[str, str]
|
|
41
|
+
|
|
42
|
+
class WebSearchActivity(Activity):
|
|
43
|
+
"""Schema for web search tool call.
|
|
44
|
+
|
|
45
|
+
Attributes:
|
|
46
|
+
type (str): The type of activity being performed.
|
|
47
|
+
pattern (str): The pattern of the web search.
|
|
48
|
+
url (str): The URL of the page.
|
|
49
|
+
query (str): The query of the web search.
|
|
50
|
+
sources (list[dict[str, str]] | None): The sources of the web search.
|
|
51
|
+
"""
|
|
52
|
+
type: Literal[ActivityType.FIND_IN_PAGE, ActivityType.OPEN_PAGE, ActivityType.SEARCH]
|
|
53
|
+
url: str | None
|
|
54
|
+
pattern: str | None
|
|
55
|
+
query: str | None
|
|
56
|
+
sources: list[dict[str, str]] | None
|
|
57
|
+
def model_dump(self, *args, **kwargs) -> dict[str, str]:
|
|
58
|
+
"""Serialize the activity for display.
|
|
59
|
+
|
|
60
|
+
Returns:
|
|
61
|
+
dict[str, str]: The serialized activity.
|
|
62
|
+
"""
|
gllm_inference/schema/enums.pyi
CHANGED
|
@@ -24,6 +24,15 @@ class EmitDataType(StrEnum):
|
|
|
24
24
|
THINKING_START = 'thinking_start'
|
|
25
25
|
THINKING_END = 'thinking_end'
|
|
26
26
|
|
|
27
|
+
class ActivityType(StrEnum):
|
|
28
|
+
"""Defines valid activity types."""
|
|
29
|
+
FIND_IN_PAGE = 'find_in_page'
|
|
30
|
+
MCP_CALL = 'mcp_call'
|
|
31
|
+
MCP_LIST_TOOLS = 'mcp_list_tools'
|
|
32
|
+
OPEN_PAGE = 'open_page'
|
|
33
|
+
SEARCH = 'search'
|
|
34
|
+
WEB_SEARCH = 'web_search'
|
|
35
|
+
|
|
27
36
|
class MessageRole(StrEnum):
|
|
28
37
|
"""Defines valid message roles."""
|
|
29
38
|
SYSTEM = 'system'
|
|
@@ -34,3 +43,11 @@ class TruncateSide(StrEnum):
|
|
|
34
43
|
"""Enumeration for truncation sides."""
|
|
35
44
|
RIGHT = 'RIGHT'
|
|
36
45
|
LEFT = 'LEFT'
|
|
46
|
+
|
|
47
|
+
class WebSearchKey(StrEnum):
|
|
48
|
+
"""Defines valid web search keys."""
|
|
49
|
+
PATTERN = 'pattern'
|
|
50
|
+
QUERY = 'query'
|
|
51
|
+
SOURCES = 'sources'
|
|
52
|
+
TYPE = 'type'
|
|
53
|
+
URL = 'url'
|
gllm_inference/schema/events.pyi
CHANGED
|
@@ -1,40 +1,109 @@
|
|
|
1
|
+
from gllm_core.constants import EventLevel
|
|
1
2
|
from gllm_core.schema import Event
|
|
2
3
|
from gllm_inference.schema.activity import Activity as Activity
|
|
3
4
|
from gllm_inference.schema.enums import EmitDataType as EmitDataType
|
|
4
5
|
from typing import Literal
|
|
5
6
|
|
|
6
|
-
class ReasoningEvent(Event):
|
|
7
|
-
"""Event schema for model reasoning.
|
|
8
|
-
|
|
9
|
-
Attributes:
|
|
10
|
-
id (str): The unique identifier for the thinking event. Defaults to an UUID string.
|
|
11
|
-
data_type (Literal): The type of thinking event (thinking, thinking_start, or thinking_end).
|
|
12
|
-
data_value (str): The thinking content or message.
|
|
13
|
-
"""
|
|
14
|
-
id: str
|
|
15
|
-
data_type: Literal[EmitDataType.THINKING, EmitDataType.THINKING_START, EmitDataType.THINKING_END]
|
|
16
|
-
data_value: str
|
|
17
|
-
|
|
18
7
|
class ActivityEvent(Event):
|
|
19
8
|
"""Event schema for model-triggered activities (e.g. web search, MCP).
|
|
20
9
|
|
|
21
10
|
Attributes:
|
|
22
11
|
id (str): The unique identifier for the activity event. Defaults to an UUID string.
|
|
23
|
-
|
|
24
|
-
|
|
12
|
+
type (Literal): The type of event, always 'activity'.
|
|
13
|
+
value (Activity): The activity data containing message and type.
|
|
14
|
+
level (EventLevel): The severity level of the event. Defined through the EventLevel constants.
|
|
25
15
|
"""
|
|
26
16
|
id: str
|
|
27
|
-
|
|
28
|
-
|
|
17
|
+
type: Literal[EmitDataType.ACTIVITY]
|
|
18
|
+
value: Activity
|
|
19
|
+
level: EventLevel
|
|
29
20
|
|
|
30
21
|
class CodeEvent(Event):
|
|
31
22
|
"""Event schema for model-triggered code execution.
|
|
32
23
|
|
|
33
24
|
Attributes:
|
|
34
25
|
id (str): The unique identifier for the code event. Defaults to an UUID string.
|
|
35
|
-
|
|
36
|
-
|
|
26
|
+
type (Literal): The type of event (code, code_start, or code_end).
|
|
27
|
+
value (str): The code content.
|
|
28
|
+
level (EventLevel): The severity level of the event. Defined through the EventLevel constants.
|
|
29
|
+
"""
|
|
30
|
+
id: str
|
|
31
|
+
type: Literal[EmitDataType.CODE, EmitDataType.CODE_START, EmitDataType.CODE_END]
|
|
32
|
+
value: str
|
|
33
|
+
level: EventLevel
|
|
34
|
+
@classmethod
|
|
35
|
+
def start(cls, id: str | None = None) -> CodeEvent:
|
|
36
|
+
"""Create a code start event.
|
|
37
|
+
|
|
38
|
+
Args:
|
|
39
|
+
id (str | None): The unique identifier for the code event. Defaults to an UUID string.
|
|
40
|
+
|
|
41
|
+
Returns:
|
|
42
|
+
CodeEvent: The code start event.
|
|
43
|
+
"""
|
|
44
|
+
@classmethod
|
|
45
|
+
def content(cls, id: str | None = None, value: str = '') -> CodeEvent:
|
|
46
|
+
"""Create a code content event.
|
|
47
|
+
|
|
48
|
+
Args:
|
|
49
|
+
id (str | None): The unique identifier for the code event. Defaults to an UUID string.
|
|
50
|
+
value (str): The code content.
|
|
51
|
+
|
|
52
|
+
Returns:
|
|
53
|
+
CodeEvent: The code value event.
|
|
54
|
+
"""
|
|
55
|
+
@classmethod
|
|
56
|
+
def end(cls, id: str | None = None) -> CodeEvent:
|
|
57
|
+
"""Create a code end event.
|
|
58
|
+
|
|
59
|
+
Args:
|
|
60
|
+
id (str | None): The unique identifier for the code event. Defaults to an UUID string.
|
|
61
|
+
|
|
62
|
+
Returns:
|
|
63
|
+
CodeEvent: The code end event.
|
|
64
|
+
"""
|
|
65
|
+
|
|
66
|
+
class ThinkingEvent(Event):
|
|
67
|
+
"""Event schema for model thinking.
|
|
68
|
+
|
|
69
|
+
Attributes:
|
|
70
|
+
id (str): The unique identifier for the thinking event. Defaults to an UUID string.
|
|
71
|
+
type (Literal): The type of thinking event (thinking, thinking_start, or thinking_end).
|
|
72
|
+
value (str): The thinking content or message.
|
|
73
|
+
level (EventLevel): The severity level of the event. Defined through the EventLevel constants.
|
|
37
74
|
"""
|
|
38
75
|
id: str
|
|
39
|
-
|
|
40
|
-
|
|
76
|
+
type: Literal[EmitDataType.THINKING, EmitDataType.THINKING_START, EmitDataType.THINKING_END]
|
|
77
|
+
value: str
|
|
78
|
+
level: EventLevel
|
|
79
|
+
@classmethod
|
|
80
|
+
def start(cls, id: str | None = None) -> ThinkingEvent:
|
|
81
|
+
"""Create a thinking start event.
|
|
82
|
+
|
|
83
|
+
Args:
|
|
84
|
+
id (str | None): The unique identifier for the thinking event. Defaults to an UUID string.
|
|
85
|
+
|
|
86
|
+
Returns:
|
|
87
|
+
ThinkingEvent: The thinking start event.
|
|
88
|
+
"""
|
|
89
|
+
@classmethod
|
|
90
|
+
def content(cls, id: str | None = None, value: str = '') -> ThinkingEvent:
|
|
91
|
+
"""Create a thinking value event.
|
|
92
|
+
|
|
93
|
+
Args:
|
|
94
|
+
id (str | None): The unique identifier for the thinking event. Defaults to an UUID string.
|
|
95
|
+
value (str): The thinking content or message.
|
|
96
|
+
|
|
97
|
+
Returns:
|
|
98
|
+
ThinkingEvent: The thinking value event.
|
|
99
|
+
"""
|
|
100
|
+
@classmethod
|
|
101
|
+
def end(cls, id: str | None = None) -> ThinkingEvent:
|
|
102
|
+
"""Create a thinking end event.
|
|
103
|
+
|
|
104
|
+
Args:
|
|
105
|
+
id (str | None): The unique identifier for the thinking event. Defaults to an UUID string.
|
|
106
|
+
|
|
107
|
+
Returns:
|
|
108
|
+
ThinkingEvent: The thinking end event.
|
|
109
|
+
"""
|
|
Binary file
|
gllm_inference.pyi
CHANGED
|
@@ -44,7 +44,6 @@ import gllm_core.utils.imports
|
|
|
44
44
|
import gllm_inference.schema.ModelId
|
|
45
45
|
import gllm_inference.schema.ModelProvider
|
|
46
46
|
import gllm_inference.schema.TruncationConfig
|
|
47
|
-
import openai
|
|
48
47
|
import asyncio
|
|
49
48
|
import enum
|
|
50
49
|
import gllm_inference.exceptions.BaseInvokerError
|
|
@@ -70,6 +69,7 @@ import gllm_inference.exceptions.InvokerRuntimeError
|
|
|
70
69
|
import gllm_inference.exceptions.build_debug_info
|
|
71
70
|
import gllm_inference.utils.load_langchain_model
|
|
72
71
|
import gllm_inference.utils.parse_model_data
|
|
72
|
+
import openai
|
|
73
73
|
import io
|
|
74
74
|
import httpx
|
|
75
75
|
import twelvelabs
|
|
@@ -85,12 +85,12 @@ import gllm_core.schema
|
|
|
85
85
|
import gllm_core.schema.tool
|
|
86
86
|
import langchain_core.tools
|
|
87
87
|
import gllm_inference.schema.BatchStatus
|
|
88
|
-
import gllm_inference.schema.EmitDataType
|
|
89
88
|
import gllm_inference.schema.LMInput
|
|
90
89
|
import gllm_inference.schema.LMOutput
|
|
91
90
|
import gllm_inference.schema.Message
|
|
92
91
|
import gllm_inference.schema.Reasoning
|
|
93
92
|
import gllm_inference.schema.ResponseSchema
|
|
93
|
+
import gllm_inference.schema.ThinkingEvent
|
|
94
94
|
import gllm_inference.schema.TokenUsage
|
|
95
95
|
import gllm_inference.schema.ToolCall
|
|
96
96
|
import gllm_inference.schema.ToolResult
|
|
@@ -108,12 +108,18 @@ import inspect
|
|
|
108
108
|
import time
|
|
109
109
|
import jsonschema
|
|
110
110
|
import gllm_inference.lm_invoker.batch.BatchOperations
|
|
111
|
+
import gllm_inference.schema.Activity
|
|
111
112
|
import gllm_inference.schema.MessageContent
|
|
112
113
|
import gllm_inference.utils.validate_string_enum
|
|
113
114
|
import __future__
|
|
115
|
+
import gllm_inference.schema.ActivityEvent
|
|
116
|
+
import gllm_inference.schema.CodeEvent
|
|
114
117
|
import gllm_inference.schema.CodeExecResult
|
|
115
118
|
import gllm_inference.schema.MCPCall
|
|
119
|
+
import gllm_inference.schema.MCPCallActivity
|
|
120
|
+
import gllm_inference.schema.MCPListToolsActivity
|
|
116
121
|
import gllm_inference.schema.MCPServer
|
|
122
|
+
import gllm_inference.schema.WebSearchActivity
|
|
117
123
|
import xai_sdk
|
|
118
124
|
import xai_sdk.chat
|
|
119
125
|
import xai_sdk.search
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.2
|
|
2
2
|
Name: gllm-inference-binary
|
|
3
|
-
Version: 0.5.
|
|
3
|
+
Version: 0.5.38
|
|
4
4
|
Summary: A library containing components related to model inferences in Gen AI applications.
|
|
5
5
|
Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, Resti Febrina <resti.febrina@gdplabs.id>
|
|
6
6
|
Requires-Python: <3.14,>=3.11
|
|
@@ -1,5 +1,5 @@
|
|
|
1
|
-
gllm_inference.cpython-312-darwin.so,sha256=
|
|
2
|
-
gllm_inference.pyi,sha256=
|
|
1
|
+
gllm_inference.cpython-312-darwin.so,sha256=FjvDDRLkbPYpfjheV0kRZOVhVd5fL8Gh742VIkLppP0,5018056
|
|
2
|
+
gllm_inference.pyi,sha256=esCvmd7d4XUM_kXqLmlXmHAm_wrcKJuNu30Yx1UCPQo,4730
|
|
3
3
|
gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
4
4
|
gllm_inference/constants.pyi,sha256=uCf0Rc8GkKwUKeMC5wx5WG_x0iV_piKhjZQl1gNffSQ,291
|
|
5
5
|
gllm_inference/builder/__init__.pyi,sha256=usz2lvfwO4Yk-ZGKXbCWG1cEr3nlQXxMNDNC-2yc1NM,500
|
|
@@ -12,13 +12,13 @@ gllm_inference/catalog/catalog.pyi,sha256=a4RNG1lKv51GxQpOqh47tz-PAROMPaeP2o5XNL
|
|
|
12
12
|
gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=WW1j8jWujnatF0c9rCk94CyobtFe3gOky2vjoGCV6nw,5424
|
|
13
13
|
gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=OU8k_4HbqjZEzHZlzSM3uzGQZJmM2uGD76Csqom0CEQ,3197
|
|
14
14
|
gllm_inference/em_invoker/__init__.pyi,sha256=83QVCkMjS2-jMKdAvmZska4LuJ-un755lAxjuVSLZ9o,987
|
|
15
|
-
gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=
|
|
15
|
+
gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=gICS8iadpItNaND3wzDIuQyYyHxz71QKXOlNboN8UEw,4959
|
|
16
16
|
gllm_inference/em_invoker/bedrock_em_invoker.pyi,sha256=77omAUXnGU_62KPC5sKOnWTwdqsoEXX38IM1JScd6K0,5723
|
|
17
17
|
gllm_inference/em_invoker/em_invoker.pyi,sha256=KGjLiAWGIA3ziV50zMwSzx6lTVUbYspQCl4LFxqbDlY,5101
|
|
18
18
|
gllm_inference/em_invoker/google_em_invoker.pyi,sha256=VVpizDo47kznqRk07t6e-Lp_K6Ojfn5KDkIKoqn2boE,6822
|
|
19
19
|
gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=BhZjkYZoxQXPJjP0PgA8W0xrtwiqDkW5E6NpGit0h1E,3498
|
|
20
|
-
gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=
|
|
21
|
-
gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=
|
|
20
|
+
gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=z4_jEuaprna6AJF2wXH-JgmynpVFXeb2HQZr-0ATmMw,2882
|
|
21
|
+
gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=FxQNV1xGEnHEWUak95OWVAwFPx65z0fphLLQ7R1qJLA,6193
|
|
22
22
|
gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=9F37VuLYTH5xsPSS_uzhrsPJug0QifX_qh2GwG0jSTU,5345
|
|
23
23
|
gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=zJZqMvvFKu3sHdrNM773UYjfHVlnwE2w2BmvdFcHzV0,5515
|
|
24
24
|
gllm_inference/em_invoker/langchain/__init__.pyi,sha256=VYGKE5OgU0my1RlhgzkU_A7-GLGnUDDnNFuctuRwILE,148
|
|
@@ -36,28 +36,28 @@ gllm_inference/exceptions/error_parser.pyi,sha256=IOfa--NpLUW5E9Qq0mwWi6ZpTAbUyy
|
|
|
36
36
|
gllm_inference/exceptions/exceptions.pyi,sha256=Bv996qLa_vju0Qjf4GewMxdkq8CV9LRZb0S6289DldA,5725
|
|
37
37
|
gllm_inference/exceptions/provider_error_map.pyi,sha256=P1WnhWkM103FW6hqMfNZBOmYSWOmsJtll3VQV8DGb8E,1210
|
|
38
38
|
gllm_inference/lm_invoker/__init__.pyi,sha256=IGF3h8Z7Yr0bLrkDMRTDVPNBU6Y3liJabW3acjstJDY,1374
|
|
39
|
-
gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=
|
|
40
|
-
gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=
|
|
39
|
+
gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=inf_fuiy89t9CPiwCs_gsF2k7G7qZ9hMKEKYsGXLtNQ,17237
|
|
40
|
+
gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=Ex91dB8-2lEo2ZKHUJghNGehAfGstQqcGzIKQ75jSUk,15056
|
|
41
41
|
gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=uZ9wpzOKSOvgu1ICMLqEXcrOE3RIbUmqHmgtuwBekPg,12802
|
|
42
|
-
gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=
|
|
43
|
-
gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=
|
|
42
|
+
gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=T4p2abiP4Vv7psroJ7_rtcSqgzBkFp_B0HGFeIZW3H8,9349
|
|
43
|
+
gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=u4egvGHRG30IotgAN6r_jrjDWFujDxjTdyoIlNLXjYA,17166
|
|
44
44
|
gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=Bxi9hT4-sIZTz-VhAlxd5JyjAjQZQbUq2xtGhUL6oIs,13705
|
|
45
45
|
gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=5dEgqLvMB04W30PO2MEUJ4h7rEMmM7ETdQ7LOlLpeXg,13238
|
|
46
|
-
gllm_inference/lm_invoker/lm_invoker.pyi,sha256=
|
|
47
|
-
gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=
|
|
48
|
-
gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=
|
|
49
|
-
gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=
|
|
50
|
-
gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=
|
|
46
|
+
gllm_inference/lm_invoker/lm_invoker.pyi,sha256=IDqmtIe9T47mrqJzb0e7dynH-9gbwqhmcb7EeVzPFos,8167
|
|
47
|
+
gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=Z3eAshWJzD8k1Phyo78GUWV8LkdOYOvZqo5ZFWYdJqE,15772
|
|
48
|
+
gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=by0LR_YqP6-uL3VMRq94t_ykO03wGz6Vv71UpWavQtk,3864
|
|
49
|
+
gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=vBo4XEZCmZBiScA6-n5aALJzIU90vsIPRah2p9neEzY,23841
|
|
50
|
+
gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=KtCGKADVbrLfvkFOPHoeLNjlA2TKRf8F_kQHekSvQQI,15689
|
|
51
51
|
gllm_inference/lm_invoker/batch/__init__.pyi,sha256=W4W-_yfk7lL20alREJai6GnwuQvdlKRfwQCX4mQK4XI,127
|
|
52
52
|
gllm_inference/lm_invoker/batch/batch_operations.pyi,sha256=Oo7hoyPSfPZdy1mXvSdvtRndvq-XTIbPIjEoGvJj5C0,5372
|
|
53
53
|
gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
54
54
|
gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=6lreMyHKRfZzX5NBYKnQf1Z6RzXBjTvqZj2VbMeaTLQ,1098
|
|
55
55
|
gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=FJLY-ZkkLUYDV48pfsLatnot4ev_xxz9xAayLK28CpU,1027
|
|
56
|
-
gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=
|
|
56
|
+
gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=WSuwOqL1j2ZioCZFC-gbB7vTRIZHQ3sU40c3ool5L6c,265
|
|
57
57
|
gllm_inference/lm_invoker/schema/google.pyi,sha256=AIsNgq0ZZuicHmx4bL7z6q-946T05nWts3HUeA8hhHQ,505
|
|
58
58
|
gllm_inference/lm_invoker/schema/langchain.pyi,sha256=rZcIxuvABI4pKfyVvkRBRqfJJogZ67EFPydpubHt49c,429
|
|
59
|
-
gllm_inference/lm_invoker/schema/openai.pyi,sha256=
|
|
60
|
-
gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=
|
|
59
|
+
gllm_inference/lm_invoker/schema/openai.pyi,sha256=J_rT5Z3rx0hLIae-me1ENeemOESpavcRmYI5pgpkhhk,2222
|
|
60
|
+
gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=8byBRZ4xyTidIQJsZqiSjp5t1X875Obe-aEbT0yYfuA,1199
|
|
61
61
|
gllm_inference/lm_invoker/schema/xai.pyi,sha256=cWnbJmDtllqRH3NXpQbiXgkNBcUXr8ksDSDywcgJebE,632
|
|
62
62
|
gllm_inference/model/__init__.pyi,sha256=qClHIgljqhPPCKlGTKmHsWgYb4_hADybxtC2q1U8a5Q,593
|
|
63
63
|
gllm_inference/model/em/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
@@ -95,13 +95,13 @@ gllm_inference/realtime_chat/output_streamer/output_streamer.pyi,sha256=GPAw1wPS
|
|
|
95
95
|
gllm_inference/request_processor/__init__.pyi,sha256=hVnfdNZnkTBJHnmLtN3Na4ANP0yK6AstWdIizVr2Apo,227
|
|
96
96
|
gllm_inference/request_processor/lm_request_processor.pyi,sha256=VnYc8E3Iayyhw-rPnGPfTKuO3ohgFsS8HPrZJeyES5I,5889
|
|
97
97
|
gllm_inference/request_processor/uses_lm_mixin.pyi,sha256=Yu0XPNuHxq1tWBviHTPw1oThojneFwGHepvGjBXxKQA,6382
|
|
98
|
-
gllm_inference/schema/__init__.pyi,sha256=
|
|
99
|
-
gllm_inference/schema/activity.pyi,sha256=
|
|
98
|
+
gllm_inference/schema/__init__.pyi,sha256=bV9_1E-3ce6Dm6Q7xJ6TynPmxQAE1-ujfp9fkRbjI0E,2116
|
|
99
|
+
gllm_inference/schema/activity.pyi,sha256=xGZbfceV8iTJcLtRJrwHfK-dYdhQRmhNwWm7KuckLh4,1974
|
|
100
100
|
gllm_inference/schema/attachment.pyi,sha256=jApuzjOHJDCz4lr4MlHzBgIndh559nbWu2Xp1fk3hso,3297
|
|
101
101
|
gllm_inference/schema/code_exec_result.pyi,sha256=ZTHh6JtRrPIdQ059P1UAiD2L-tAO1_S5YcMsAXfJ5A0,559
|
|
102
102
|
gllm_inference/schema/config.pyi,sha256=rAL_UeXyQeXVk1P2kqd8vFWOMwmKenfpQLtvMP74t9s,674
|
|
103
|
-
gllm_inference/schema/enums.pyi,sha256=
|
|
104
|
-
gllm_inference/schema/events.pyi,sha256=
|
|
103
|
+
gllm_inference/schema/enums.pyi,sha256=5shfuJs76BqiAIZpA1wW1gXe5_KuFikQvuW3ALbQAmM,1322
|
|
104
|
+
gllm_inference/schema/events.pyi,sha256=Ly7B0N0_B3UztJNpT2oqOZ7YArEQp2VZpZ-VFeGc3ok,3911
|
|
105
105
|
gllm_inference/schema/lm_input.pyi,sha256=A5pjz1id6tP9XRNhzQrbmzd66C_q3gzo0UP8rCemz6Q,193
|
|
106
106
|
gllm_inference/schema/lm_output.pyi,sha256=15y-M0lpqM_fSlErPKiN1Pj-ikl5NtFBcWLMYsRidt8,2182
|
|
107
107
|
gllm_inference/schema/mcp.pyi,sha256=Vwu8E2BDl6FvvnI42gIyY3Oki1BdwRE3Uh3aV0rmhQU,1014
|
|
@@ -117,7 +117,7 @@ gllm_inference/utils/io_utils.pyi,sha256=7kUTacHAVRYoemFUOjCH7-Qmw-YsQGd6rGYxjf_
|
|
|
117
117
|
gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
|
|
118
118
|
gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
|
|
119
119
|
gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
|
|
120
|
-
gllm_inference_binary-0.5.
|
|
121
|
-
gllm_inference_binary-0.5.
|
|
122
|
-
gllm_inference_binary-0.5.
|
|
123
|
-
gllm_inference_binary-0.5.
|
|
120
|
+
gllm_inference_binary-0.5.38.dist-info/METADATA,sha256=FU3f7oxGCu59ZA1CbA4zZi2in2Bt21J5hJtEJ8zbtUI,5636
|
|
121
|
+
gllm_inference_binary-0.5.38.dist-info/WHEEL,sha256=ar3KUKk5QtasLek_3_fVX4zgPzX-lxtBGErPgJ515rA,105
|
|
122
|
+
gllm_inference_binary-0.5.38.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
|
|
123
|
+
gllm_inference_binary-0.5.38.dist-info/RECORD,,
|
|
File without changes
|
{gllm_inference_binary-0.5.36.dist-info → gllm_inference_binary-0.5.38.dist-info}/top_level.txt
RENAMED
|
File without changes
|