gllm-inference-binary 0.5.36__cp312-cp312-macosx_13_0_x86_64.whl → 0.5.38__cp312-cp312-macosx_13_0_x86_64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of gllm-inference-binary might be problematic. Click here for more details.

Files changed (25) hide show
  1. gllm_inference/em_invoker/azure_openai_em_invoker.pyi +2 -2
  2. gllm_inference/em_invoker/openai_compatible_em_invoker.pyi +2 -2
  3. gllm_inference/em_invoker/openai_em_invoker.pyi +2 -2
  4. gllm_inference/lm_invoker/anthropic_lm_invoker.pyi +1 -1
  5. gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi +2 -2
  6. gllm_inference/lm_invoker/datasaur_lm_invoker.pyi +2 -2
  7. gllm_inference/lm_invoker/google_lm_invoker.pyi +1 -1
  8. gllm_inference/lm_invoker/lm_invoker.pyi +3 -2
  9. gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi +3 -3
  10. gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi +1 -1
  11. gllm_inference/lm_invoker/openai_lm_invoker.pyi +3 -3
  12. gllm_inference/lm_invoker/schema/datasaur.pyi +2 -0
  13. gllm_inference/lm_invoker/schema/openai.pyi +2 -0
  14. gllm_inference/lm_invoker/schema/openai_chat_completions.pyi +2 -0
  15. gllm_inference/lm_invoker/xai_lm_invoker.pyi +1 -1
  16. gllm_inference/schema/__init__.pyi +3 -3
  17. gllm_inference/schema/activity.pyi +55 -2
  18. gllm_inference/schema/enums.pyi +17 -0
  19. gllm_inference/schema/events.pyi +89 -20
  20. gllm_inference.cpython-312-darwin.so +0 -0
  21. gllm_inference.pyi +8 -2
  22. {gllm_inference_binary-0.5.36.dist-info → gllm_inference_binary-0.5.38.dist-info}/METADATA +1 -1
  23. {gllm_inference_binary-0.5.36.dist-info → gllm_inference_binary-0.5.38.dist-info}/RECORD +25 -25
  24. {gllm_inference_binary-0.5.36.dist-info → gllm_inference_binary-0.5.38.dist-info}/WHEEL +0 -0
  25. {gllm_inference_binary-0.5.36.dist-info → gllm_inference_binary-0.5.38.dist-info}/top_level.txt +0 -0
@@ -13,7 +13,7 @@ class AzureOpenAIEMInvoker(OpenAIEMInvoker):
13
13
  model_id (str): The model ID of the embedding model.
14
14
  model_provider (str): The provider of the embedding model.
15
15
  model_name (str): The name of the Azure OpenAI embedding model deployment.
16
- client (AsyncAzureOpenAI): The client for the Azure OpenAI API.
16
+ client_kwargs (dict[str, Any]): The keyword arguments for the Azure OpenAI client.
17
17
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the embedding model.
18
18
  retry_config (RetryConfig): The retry configuration for the embedding model.
19
19
  truncation_config (TruncationConfig | None): The truncation configuration for the embedding model.
@@ -68,7 +68,7 @@ class AzureOpenAIEMInvoker(OpenAIEMInvoker):
68
68
  em_invoker = AzureOpenAIEMInvoker(..., retry_config=retry_config)
69
69
  ```
70
70
  '''
71
- client: Incomplete
71
+ client_kwargs: Incomplete
72
72
  def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
73
73
  """Initializes a new instance of the AzureOpenAIEMInvoker class.
74
74
 
@@ -15,14 +15,14 @@ class OpenAICompatibleEMInvoker(OpenAIEMInvoker):
15
15
  model_id (str): The model ID of the embedding model.
16
16
  model_provider (str): The provider of the embedding model.
17
17
  model_name (str): The name of the embedding model.
18
- client (AsyncOpenAI): The OpenAI client instance.
18
+ client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
19
19
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the embedding model.
20
20
  retry_config (RetryConfig): The retry configuration for the embedding model.
21
21
  truncation_config (TruncationConfig | None): The truncation configuration for the embedding model.
22
22
 
23
23
  This class is deprecated and will be removed in v0.6. Please use the `OpenAIEMInvoker` class instead.
24
24
  """
25
- client: Incomplete
25
+ client_kwargs: Incomplete
26
26
  def __init__(self, model_name: str, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
27
27
  """Initializes a new instance of the OpenAICompatibleEMInvoker class.
28
28
 
@@ -16,7 +16,7 @@ class OpenAIEMInvoker(BaseEMInvoker):
16
16
  model_id (str): The model ID of the embedding model.
17
17
  model_provider (str): The provider of the embedding model.
18
18
  model_name (str): The name of the embedding model.
19
- client (AsyncOpenAI): The client for the OpenAI API.
19
+ client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
20
20
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the embedding model.
21
21
  retry_config (RetryConfig): The retry configuration for the embedding model.
22
22
  truncation_config (TruncationConfig | None): The truncation configuration for the embedding model.
@@ -96,7 +96,7 @@ class OpenAIEMInvoker(BaseEMInvoker):
96
96
  em_invoker = OpenAIEMInvoker(..., retry_config=retry_config)
97
97
  ```
98
98
  '''
99
- client: Incomplete
99
+ client_kwargs: Incomplete
100
100
  def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
101
101
  '''Initializes a new instance of the OpenAIEMInvoker class.
102
102
 
@@ -5,7 +5,7 @@ from gllm_core.utils.retry import RetryConfig as RetryConfig
5
5
  from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
6
6
  from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
7
7
  from gllm_inference.lm_invoker.schema.anthropic import InputType as InputType, Key as Key, OutputType as OutputType
8
- from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, BatchStatus as BatchStatus, EmitDataType as EmitDataType, LMInput as LMInput, LMOutput as LMOutput, Message as Message, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
8
+ from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, BatchStatus as BatchStatus, LMInput as LMInput, LMOutput as LMOutput, Message as Message, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ThinkingEvent as ThinkingEvent, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
9
9
  from langchain_core.tools import Tool as LangChainTool
10
10
  from typing import Any
11
11
 
@@ -15,7 +15,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
15
15
  model_id (str): The model ID of the language model.
16
16
  model_provider (str): The provider of the language model.
17
17
  model_name (str): The name of the Azure OpenAI language model deployment.
18
- client (AsyncAzureOpenAI): The Azure OpenAI client instance.
18
+ client_kwargs (dict[str, Any]): The keyword arguments for the Azure OpenAI client.
19
19
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
20
20
  tools (list[Tool]): The list of tools provided to the model to enable tool calling.
21
21
  response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
@@ -226,7 +226,7 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
226
226
  Defaults to an empty list.
227
227
  2.10. mcp_calls (list[MCPCall]): The MCP calls. Currently not supported. Defaults to an empty list.
228
228
  '''
229
- client: Incomplete
229
+ client_kwargs: Incomplete
230
230
  def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None) -> None:
231
231
  """Initializes a new instance of the AzureOpenAILMInvoker class.
232
232
 
@@ -18,7 +18,7 @@ class DatasaurLMInvoker(OpenAICompatibleLMInvoker):
18
18
  model_id (str): The model ID of the language model.
19
19
  model_provider (str): The provider of the language model.
20
20
  model_name (str): The name of the language model.
21
- client (AsyncOpenAI): The OpenAI client instance.
21
+ client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
22
22
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
23
23
  tools (list[Any]): The list of tools provided to the model to enable tool calling. Currently not supported.
24
24
  response_schema (ResponseSchema | None): The schema of the response. Currently not supported.
@@ -121,7 +121,7 @@ class DatasaurLMInvoker(OpenAICompatibleLMInvoker):
121
121
  Defaults to an empty list.
122
122
  2.10. mcp_calls (list[MCPCall]): The MCP calls. Currently not supported. Defaults to an empty list.
123
123
  '''
124
- client: Incomplete
124
+ client_kwargs: Incomplete
125
125
  citations: Incomplete
126
126
  def __init__(self, base_url: str, api_key: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, citations: bool = False) -> None:
127
127
  """Initializes a new instance of the DatasaurLMInvoker class.
@@ -7,7 +7,7 @@ from gllm_inference.exceptions import BaseInvokerError as BaseInvokerError, conv
7
7
  from gllm_inference.exceptions.provider_error_map import GOOGLE_ERROR_MAPPING as GOOGLE_ERROR_MAPPING
8
8
  from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
9
9
  from gllm_inference.lm_invoker.schema.google import InputType as InputType, Key as Key
10
- from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EmitDataType as EmitDataType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
10
+ from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ThinkingEvent as ThinkingEvent, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
11
11
  from langchain_core.tools import Tool as LangChainTool
12
12
  from typing import Any
13
13
 
@@ -2,12 +2,12 @@ import abc
2
2
  from _typeshed import Incomplete
3
3
  from abc import ABC
4
4
  from gllm_core.event import EventEmitter as EventEmitter
5
- from gllm_core.schema.tool import Tool
5
+ from gllm_core.schema import Event as Event, Tool
6
6
  from gllm_core.utils import RetryConfig
7
7
  from gllm_inference.constants import DOCUMENT_MIME_TYPES as DOCUMENT_MIME_TYPES, INVOKER_DEFAULT_TIMEOUT as INVOKER_DEFAULT_TIMEOUT
8
8
  from gllm_inference.exceptions import BaseInvokerError as BaseInvokerError, convert_to_base_invoker_error as convert_to_base_invoker_error
9
9
  from gllm_inference.lm_invoker.batch import BatchOperations as BatchOperations
10
- from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, BatchStatus as BatchStatus, EmitDataType as EmitDataType, LMInput as LMInput, LMOutput as LMOutput, Message as Message, MessageContent as MessageContent, MessageRole as MessageRole, ModelId as ModelId, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ToolCall as ToolCall, ToolResult as ToolResult
10
+ from gllm_inference.schema import Activity as Activity, Attachment as Attachment, AttachmentType as AttachmentType, BatchStatus as BatchStatus, LMInput as LMInput, LMOutput as LMOutput, Message as Message, MessageContent as MessageContent, MessageRole as MessageRole, ModelId as ModelId, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ToolCall as ToolCall, ToolResult as ToolResult
11
11
  from langchain_core.tools import Tool as LangChainTool
12
12
  from typing import Any
13
13
 
@@ -23,6 +23,7 @@ class Key:
23
23
  DEFAULT: str
24
24
  DESCRIPTION: str
25
25
  FUNC: str
26
+ ID: str
26
27
  NAME: str
27
28
  PROPERTIES: str
28
29
  REQUIRED: str
@@ -5,7 +5,7 @@ from gllm_core.utils.retry import RetryConfig as RetryConfig
5
5
  from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES, OPENAI_DEFAULT_URL as OPENAI_DEFAULT_URL
6
6
  from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
7
7
  from gllm_inference.lm_invoker.schema.openai_chat_completions import InputType as InputType, Key as Key, ReasoningEffort as ReasoningEffort
8
- from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EmitDataType as EmitDataType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
8
+ from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ThinkingEvent as ThinkingEvent, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
9
9
  from gllm_inference.utils import validate_string_enum as validate_string_enum
10
10
  from langchain_core.tools import Tool as LangChainTool
11
11
  from typing import Any
@@ -23,7 +23,7 @@ class OpenAIChatCompletionsLMInvoker(BaseLMInvoker):
23
23
  model_id (str): The model ID of the language model.
24
24
  model_provider (str): The provider of the language model.
25
25
  model_name (str): The name of the language model.
26
- client (AsyncOpenAI): The OpenAI client instance.
26
+ client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
27
27
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
28
28
  tools (list[Tool]): The list of tools provided to the model to enable tool calling.
29
29
  response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
@@ -243,7 +243,7 @@ class OpenAIChatCompletionsLMInvoker(BaseLMInvoker):
243
243
  Defaults to an empty list.
244
244
  2.10. mcp_calls (list[MCPCall]): The MCP calls. Currently not supported. Defaults to an empty list.
245
245
  '''
246
- client: Incomplete
246
+ client_kwargs: Incomplete
247
247
  def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None) -> None:
248
248
  '''Initializes a new instance of the OpenAIChatCompletionsLMInvoker class.
249
249
 
@@ -15,7 +15,7 @@ class OpenAICompatibleLMInvoker(OpenAIChatCompletionsLMInvoker):
15
15
  model_id (str): The model ID of the language model.
16
16
  model_provider (str): The provider of the language model.
17
17
  model_name (str): The name of the language model.
18
- client (AsyncOpenAI): The OpenAI client instance.
18
+ client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
19
19
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
20
20
  tools (list[Tool]): The list of tools provided to the model to enable tool calling.
21
21
  response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
@@ -5,7 +5,7 @@ from gllm_core.utils.retry import RetryConfig as RetryConfig
5
5
  from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES, OPENAI_DEFAULT_URL as OPENAI_DEFAULT_URL
6
6
  from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
7
7
  from gllm_inference.lm_invoker.schema.openai import InputType as InputType, Key as Key, OutputType as OutputType, ReasoningEffort as ReasoningEffort, ReasoningSummary as ReasoningSummary
8
- from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, CodeExecResult as CodeExecResult, EmitDataType as EmitDataType, LMOutput as LMOutput, MCPCall as MCPCall, MCPServer as MCPServer, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
8
+ from gllm_inference.schema import ActivityEvent as ActivityEvent, Attachment as Attachment, AttachmentType as AttachmentType, CodeEvent as CodeEvent, CodeExecResult as CodeExecResult, LMOutput as LMOutput, MCPCall as MCPCall, MCPCallActivity as MCPCallActivity, MCPListToolsActivity as MCPListToolsActivity, MCPServer as MCPServer, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ThinkingEvent as ThinkingEvent, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult, WebSearchActivity as WebSearchActivity
9
9
  from gllm_inference.utils import validate_string_enum as validate_string_enum
10
10
  from langchain_core.tools import Tool as LangChainTool
11
11
  from typing import Any
@@ -25,7 +25,7 @@ class OpenAILMInvoker(BaseLMInvoker):
25
25
  model_id (str): The model ID of the language model.
26
26
  model_provider (str): The provider of the language model.
27
27
  model_name (str): The name of the language model.
28
- client (AsyncOpenAI): The OpenAI client instance.
28
+ client_kwargs (dict[str, Any]): The keyword arguments for the OpenAI client.
29
29
  default_hyperparameters (dict[str, Any]): Default hyperparameters for invoking the model.
30
30
  tools (list[Tool]): The list of tools provided to the model to enable tool calling.
31
31
  response_schema (ResponseSchema | None): The schema of the response. If provided, the model will output a
@@ -391,7 +391,7 @@ class OpenAILMInvoker(BaseLMInvoker):
391
391
  2.10. mcp_calls (list[MCPCall]): The MCP calls, if the MCP servers are provided and the language model
392
392
  decides to invoke MCP tools. Defaults to an empty list.
393
393
  '''
394
- client: Incomplete
394
+ client_kwargs: Incomplete
395
395
  def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None, mcp_servers: list[MCPServer] | None = None, code_interpreter: bool = False, web_search: bool = False) -> None:
396
396
  '''Initializes a new instance of the OpenAILMInvoker class.
397
397
 
@@ -1,5 +1,7 @@
1
1
  class Key:
2
2
  """Defines valid keys in Datasaur."""
3
+ API_KEY: str
4
+ BASE_URL: str
3
5
  CONTEXTS: str
4
6
  MAX_RETRIES: str
5
7
  NAME: str
@@ -3,8 +3,10 @@ from enum import StrEnum
3
3
  class Key:
4
4
  """Defines valid keys in OpenAI."""
5
5
  ALLOWED_TOOLS: str
6
+ API_KEY: str
6
7
  ARGS: str
7
8
  ARGUMENTS: str
9
+ BASE_URL: str
8
10
  CALL_ID: str
9
11
  CONTAINER: str
10
12
  CONTENT: str
@@ -2,7 +2,9 @@ from enum import StrEnum
2
2
 
3
3
  class Key:
4
4
  """Defines valid keys in OpenAI Chat Completions."""
5
+ API_KEY: str
5
6
  ARGUMENTS: str
7
+ BASE_URL: str
6
8
  CONTENT: str
7
9
  CHOICES: str
8
10
  DATA: str
@@ -7,7 +7,7 @@ from gllm_inference.exceptions import BaseInvokerError as BaseInvokerError, Invo
7
7
  from gllm_inference.exceptions.provider_error_map import GRPC_STATUS_CODE_MAPPING as GRPC_STATUS_CODE_MAPPING
8
8
  from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
9
9
  from gllm_inference.lm_invoker.schema.xai import Key as Key, ReasoningEffort as ReasoningEffort
10
- from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EmitDataType as EmitDataType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
10
+ from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, LMOutput as LMOutput, Message as Message, MessageRole as MessageRole, ModelId as ModelId, ModelProvider as ModelProvider, Reasoning as Reasoning, ResponseSchema as ResponseSchema, ThinkingEvent as ThinkingEvent, TokenUsage as TokenUsage, ToolCall as ToolCall, ToolResult as ToolResult
11
11
  from gllm_inference.utils.validation import validate_string_enum as validate_string_enum
12
12
  from langchain_core.tools import Tool as LangChainTool
13
13
  from typing import Any
@@ -1,9 +1,9 @@
1
- from gllm_inference.schema.activity import Activity as Activity
1
+ from gllm_inference.schema.activity import Activity as Activity, MCPCallActivity as MCPCallActivity, MCPListToolsActivity as MCPListToolsActivity, WebSearchActivity as WebSearchActivity
2
2
  from gllm_inference.schema.attachment import Attachment as Attachment
3
3
  from gllm_inference.schema.code_exec_result import CodeExecResult as CodeExecResult
4
4
  from gllm_inference.schema.config import TruncationConfig as TruncationConfig
5
5
  from gllm_inference.schema.enums import AttachmentType as AttachmentType, BatchStatus as BatchStatus, EmitDataType as EmitDataType, MessageRole as MessageRole, TruncateSide as TruncateSide
6
- from gllm_inference.schema.events import ActivityEvent as ActivityEvent, CodeEvent as CodeEvent, ReasoningEvent as ReasoningEvent
6
+ from gllm_inference.schema.events import ActivityEvent as ActivityEvent, CodeEvent as CodeEvent, ThinkingEvent as ThinkingEvent
7
7
  from gllm_inference.schema.lm_input import LMInput as LMInput
8
8
  from gllm_inference.schema.lm_output import LMOutput as LMOutput
9
9
  from gllm_inference.schema.mcp import MCPCall as MCPCall, MCPServer as MCPServer
@@ -15,4 +15,4 @@ from gllm_inference.schema.tool_call import ToolCall as ToolCall
15
15
  from gllm_inference.schema.tool_result import ToolResult as ToolResult
16
16
  from gllm_inference.schema.type_alias import EMContent as EMContent, MessageContent as MessageContent, ResponseSchema as ResponseSchema, Vector as Vector
17
17
 
18
- __all__ = ['Activity', 'ActivityEvent', 'Attachment', 'AttachmentType', 'BatchStatus', 'CodeEvent', 'CodeExecResult', 'EMContent', 'EmitDataType', 'MCPCall', 'MCPServer', 'InputTokenDetails', 'MessageContent', 'LMInput', 'LMOutput', 'ModelId', 'ModelProvider', 'Message', 'MessageRole', 'OutputTokenDetails', 'Reasoning', 'ReasoningEvent', 'ResponseSchema', 'TokenUsage', 'ToolCall', 'ToolResult', 'TruncateSide', 'TruncationConfig', 'Vector']
18
+ __all__ = ['Activity', 'ActivityEvent', 'Attachment', 'AttachmentType', 'BatchStatus', 'CodeEvent', 'CodeExecResult', 'EMContent', 'EmitDataType', 'InputTokenDetails', 'LMInput', 'LMOutput', 'MCPCall', 'MCPCallActivity', 'MCPListToolsActivity', 'MCPServer', 'Message', 'MessageContent', 'MessageRole', 'ModelId', 'ModelProvider', 'OutputTokenDetails', 'Reasoning', 'ThinkingEvent', 'ResponseSchema', 'TokenUsage', 'ToolCall', 'ToolResult', 'TruncateSide', 'TruncationConfig', 'Vector', 'WebSearchActivity']
@@ -1,9 +1,62 @@
1
+ from _typeshed import Incomplete
2
+ from gllm_inference.schema.enums import ActivityType as ActivityType, WebSearchKey as WebSearchKey
1
3
  from pydantic import BaseModel
4
+ from typing import Literal
5
+
6
+ WEB_SEARCH_VISIBLE_FIELDS: Incomplete
2
7
 
3
8
  class Activity(BaseModel):
4
9
  """Base schema for any activity.
5
10
 
6
11
  Attributes:
7
- activity_type (str): The type of activity being performed.
12
+ type (str): The type of activity being performed.
13
+ """
14
+ type: str
15
+
16
+ class MCPListToolsActivity(Activity):
17
+ """Schema for listing tools in MCP.
18
+
19
+ Attributes:
20
+ server_name (str): The name of the MCP server.
21
+ tools (list[dict[str, str]] | None): The tools in the MCP server.
22
+ type (str): The type of activity being performed.
23
+ """
24
+ type: Literal[ActivityType.MCP_LIST_TOOLS]
25
+ server_name: str
26
+ tools: list[dict[str, str]] | None
27
+
28
+ class MCPCallActivity(Activity):
29
+ """Schema for MCP tool call.
30
+
31
+ Attributes:
32
+ server_name (str): The name of the MCP server.
33
+ tool_name (str): The name of the tool.
34
+ args (dict[str, str]): The arguments of the tool.
35
+ type (str): The type of activity being performed.
8
36
  """
9
- activity_type: str
37
+ type: Literal[ActivityType.MCP_CALL]
38
+ server_name: str
39
+ tool_name: str
40
+ args: dict[str, str]
41
+
42
+ class WebSearchActivity(Activity):
43
+ """Schema for web search tool call.
44
+
45
+ Attributes:
46
+ type (str): The type of activity being performed.
47
+ pattern (str): The pattern of the web search.
48
+ url (str): The URL of the page.
49
+ query (str): The query of the web search.
50
+ sources (list[dict[str, str]] | None): The sources of the web search.
51
+ """
52
+ type: Literal[ActivityType.FIND_IN_PAGE, ActivityType.OPEN_PAGE, ActivityType.SEARCH]
53
+ url: str | None
54
+ pattern: str | None
55
+ query: str | None
56
+ sources: list[dict[str, str]] | None
57
+ def model_dump(self, *args, **kwargs) -> dict[str, str]:
58
+ """Serialize the activity for display.
59
+
60
+ Returns:
61
+ dict[str, str]: The serialized activity.
62
+ """
@@ -24,6 +24,15 @@ class EmitDataType(StrEnum):
24
24
  THINKING_START = 'thinking_start'
25
25
  THINKING_END = 'thinking_end'
26
26
 
27
+ class ActivityType(StrEnum):
28
+ """Defines valid activity types."""
29
+ FIND_IN_PAGE = 'find_in_page'
30
+ MCP_CALL = 'mcp_call'
31
+ MCP_LIST_TOOLS = 'mcp_list_tools'
32
+ OPEN_PAGE = 'open_page'
33
+ SEARCH = 'search'
34
+ WEB_SEARCH = 'web_search'
35
+
27
36
  class MessageRole(StrEnum):
28
37
  """Defines valid message roles."""
29
38
  SYSTEM = 'system'
@@ -34,3 +43,11 @@ class TruncateSide(StrEnum):
34
43
  """Enumeration for truncation sides."""
35
44
  RIGHT = 'RIGHT'
36
45
  LEFT = 'LEFT'
46
+
47
+ class WebSearchKey(StrEnum):
48
+ """Defines valid web search keys."""
49
+ PATTERN = 'pattern'
50
+ QUERY = 'query'
51
+ SOURCES = 'sources'
52
+ TYPE = 'type'
53
+ URL = 'url'
@@ -1,40 +1,109 @@
1
+ from gllm_core.constants import EventLevel
1
2
  from gllm_core.schema import Event
2
3
  from gllm_inference.schema.activity import Activity as Activity
3
4
  from gllm_inference.schema.enums import EmitDataType as EmitDataType
4
5
  from typing import Literal
5
6
 
6
- class ReasoningEvent(Event):
7
- """Event schema for model reasoning.
8
-
9
- Attributes:
10
- id (str): The unique identifier for the thinking event. Defaults to an UUID string.
11
- data_type (Literal): The type of thinking event (thinking, thinking_start, or thinking_end).
12
- data_value (str): The thinking content or message.
13
- """
14
- id: str
15
- data_type: Literal[EmitDataType.THINKING, EmitDataType.THINKING_START, EmitDataType.THINKING_END]
16
- data_value: str
17
-
18
7
  class ActivityEvent(Event):
19
8
  """Event schema for model-triggered activities (e.g. web search, MCP).
20
9
 
21
10
  Attributes:
22
11
  id (str): The unique identifier for the activity event. Defaults to an UUID string.
23
- data_type (Literal): The type of event, always 'activity'.
24
- data_value (Activity): The activity data containing message and type.
12
+ type (Literal): The type of event, always 'activity'.
13
+ value (Activity): The activity data containing message and type.
14
+ level (EventLevel): The severity level of the event. Defined through the EventLevel constants.
25
15
  """
26
16
  id: str
27
- data_type: Literal[EmitDataType.ACTIVITY]
28
- data_value: Activity
17
+ type: Literal[EmitDataType.ACTIVITY]
18
+ value: Activity
19
+ level: EventLevel
29
20
 
30
21
  class CodeEvent(Event):
31
22
  """Event schema for model-triggered code execution.
32
23
 
33
24
  Attributes:
34
25
  id (str): The unique identifier for the code event. Defaults to an UUID string.
35
- data_type (Literal): The type of event (code, code_start, or code_end).
36
- data_value (str): The code content.
26
+ type (Literal): The type of event (code, code_start, or code_end).
27
+ value (str): The code content.
28
+ level (EventLevel): The severity level of the event. Defined through the EventLevel constants.
29
+ """
30
+ id: str
31
+ type: Literal[EmitDataType.CODE, EmitDataType.CODE_START, EmitDataType.CODE_END]
32
+ value: str
33
+ level: EventLevel
34
+ @classmethod
35
+ def start(cls, id: str | None = None) -> CodeEvent:
36
+ """Create a code start event.
37
+
38
+ Args:
39
+ id (str | None): The unique identifier for the code event. Defaults to an UUID string.
40
+
41
+ Returns:
42
+ CodeEvent: The code start event.
43
+ """
44
+ @classmethod
45
+ def content(cls, id: str | None = None, value: str = '') -> CodeEvent:
46
+ """Create a code content event.
47
+
48
+ Args:
49
+ id (str | None): The unique identifier for the code event. Defaults to an UUID string.
50
+ value (str): The code content.
51
+
52
+ Returns:
53
+ CodeEvent: The code value event.
54
+ """
55
+ @classmethod
56
+ def end(cls, id: str | None = None) -> CodeEvent:
57
+ """Create a code end event.
58
+
59
+ Args:
60
+ id (str | None): The unique identifier for the code event. Defaults to an UUID string.
61
+
62
+ Returns:
63
+ CodeEvent: The code end event.
64
+ """
65
+
66
+ class ThinkingEvent(Event):
67
+ """Event schema for model thinking.
68
+
69
+ Attributes:
70
+ id (str): The unique identifier for the thinking event. Defaults to an UUID string.
71
+ type (Literal): The type of thinking event (thinking, thinking_start, or thinking_end).
72
+ value (str): The thinking content or message.
73
+ level (EventLevel): The severity level of the event. Defined through the EventLevel constants.
37
74
  """
38
75
  id: str
39
- data_type: Literal[EmitDataType.CODE, EmitDataType.CODE_START, EmitDataType.CODE_END]
40
- data_value: str
76
+ type: Literal[EmitDataType.THINKING, EmitDataType.THINKING_START, EmitDataType.THINKING_END]
77
+ value: str
78
+ level: EventLevel
79
+ @classmethod
80
+ def start(cls, id: str | None = None) -> ThinkingEvent:
81
+ """Create a thinking start event.
82
+
83
+ Args:
84
+ id (str | None): The unique identifier for the thinking event. Defaults to an UUID string.
85
+
86
+ Returns:
87
+ ThinkingEvent: The thinking start event.
88
+ """
89
+ @classmethod
90
+ def content(cls, id: str | None = None, value: str = '') -> ThinkingEvent:
91
+ """Create a thinking value event.
92
+
93
+ Args:
94
+ id (str | None): The unique identifier for the thinking event. Defaults to an UUID string.
95
+ value (str): The thinking content or message.
96
+
97
+ Returns:
98
+ ThinkingEvent: The thinking value event.
99
+ """
100
+ @classmethod
101
+ def end(cls, id: str | None = None) -> ThinkingEvent:
102
+ """Create a thinking end event.
103
+
104
+ Args:
105
+ id (str | None): The unique identifier for the thinking event. Defaults to an UUID string.
106
+
107
+ Returns:
108
+ ThinkingEvent: The thinking end event.
109
+ """
Binary file
gllm_inference.pyi CHANGED
@@ -44,7 +44,6 @@ import gllm_core.utils.imports
44
44
  import gllm_inference.schema.ModelId
45
45
  import gllm_inference.schema.ModelProvider
46
46
  import gllm_inference.schema.TruncationConfig
47
- import openai
48
47
  import asyncio
49
48
  import enum
50
49
  import gllm_inference.exceptions.BaseInvokerError
@@ -70,6 +69,7 @@ import gllm_inference.exceptions.InvokerRuntimeError
70
69
  import gllm_inference.exceptions.build_debug_info
71
70
  import gllm_inference.utils.load_langchain_model
72
71
  import gllm_inference.utils.parse_model_data
72
+ import openai
73
73
  import io
74
74
  import httpx
75
75
  import twelvelabs
@@ -85,12 +85,12 @@ import gllm_core.schema
85
85
  import gllm_core.schema.tool
86
86
  import langchain_core.tools
87
87
  import gllm_inference.schema.BatchStatus
88
- import gllm_inference.schema.EmitDataType
89
88
  import gllm_inference.schema.LMInput
90
89
  import gllm_inference.schema.LMOutput
91
90
  import gllm_inference.schema.Message
92
91
  import gllm_inference.schema.Reasoning
93
92
  import gllm_inference.schema.ResponseSchema
93
+ import gllm_inference.schema.ThinkingEvent
94
94
  import gllm_inference.schema.TokenUsage
95
95
  import gllm_inference.schema.ToolCall
96
96
  import gllm_inference.schema.ToolResult
@@ -108,12 +108,18 @@ import inspect
108
108
  import time
109
109
  import jsonschema
110
110
  import gllm_inference.lm_invoker.batch.BatchOperations
111
+ import gllm_inference.schema.Activity
111
112
  import gllm_inference.schema.MessageContent
112
113
  import gllm_inference.utils.validate_string_enum
113
114
  import __future__
115
+ import gllm_inference.schema.ActivityEvent
116
+ import gllm_inference.schema.CodeEvent
114
117
  import gllm_inference.schema.CodeExecResult
115
118
  import gllm_inference.schema.MCPCall
119
+ import gllm_inference.schema.MCPCallActivity
120
+ import gllm_inference.schema.MCPListToolsActivity
116
121
  import gllm_inference.schema.MCPServer
122
+ import gllm_inference.schema.WebSearchActivity
117
123
  import xai_sdk
118
124
  import xai_sdk.chat
119
125
  import xai_sdk.search
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.2
2
2
  Name: gllm-inference-binary
3
- Version: 0.5.36
3
+ Version: 0.5.38
4
4
  Summary: A library containing components related to model inferences in Gen AI applications.
5
5
  Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, Resti Febrina <resti.febrina@gdplabs.id>
6
6
  Requires-Python: <3.14,>=3.11
@@ -1,5 +1,5 @@
1
- gllm_inference.cpython-312-darwin.so,sha256=BsIPawxmSj4KcBpNwu39bem2oEt7YcXexLZpLSz1oPk,4913496
2
- gllm_inference.pyi,sha256=iM7lOtfdpE_YvCEfI5mcK57dnBO8zNOZlzaPborXUCI,4467
1
+ gllm_inference.cpython-312-darwin.so,sha256=FjvDDRLkbPYpfjheV0kRZOVhVd5fL8Gh742VIkLppP0,5018056
2
+ gllm_inference.pyi,sha256=esCvmd7d4XUM_kXqLmlXmHAm_wrcKJuNu30Yx1UCPQo,4730
3
3
  gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
4
4
  gllm_inference/constants.pyi,sha256=uCf0Rc8GkKwUKeMC5wx5WG_x0iV_piKhjZQl1gNffSQ,291
5
5
  gllm_inference/builder/__init__.pyi,sha256=usz2lvfwO4Yk-ZGKXbCWG1cEr3nlQXxMNDNC-2yc1NM,500
@@ -12,13 +12,13 @@ gllm_inference/catalog/catalog.pyi,sha256=a4RNG1lKv51GxQpOqh47tz-PAROMPaeP2o5XNL
12
12
  gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=WW1j8jWujnatF0c9rCk94CyobtFe3gOky2vjoGCV6nw,5424
13
13
  gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=OU8k_4HbqjZEzHZlzSM3uzGQZJmM2uGD76Csqom0CEQ,3197
14
14
  gllm_inference/em_invoker/__init__.pyi,sha256=83QVCkMjS2-jMKdAvmZska4LuJ-un755lAxjuVSLZ9o,987
15
- gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=Tp92kRqUmB2FpqdnWdJXDGZ_ibOzgFeZAEey4coaD5E,4933
15
+ gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=gICS8iadpItNaND3wzDIuQyYyHxz71QKXOlNboN8UEw,4959
16
16
  gllm_inference/em_invoker/bedrock_em_invoker.pyi,sha256=77omAUXnGU_62KPC5sKOnWTwdqsoEXX38IM1JScd6K0,5723
17
17
  gllm_inference/em_invoker/em_invoker.pyi,sha256=KGjLiAWGIA3ziV50zMwSzx6lTVUbYspQCl4LFxqbDlY,5101
18
18
  gllm_inference/em_invoker/google_em_invoker.pyi,sha256=VVpizDo47kznqRk07t6e-Lp_K6Ojfn5KDkIKoqn2boE,6822
19
19
  gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=BhZjkYZoxQXPJjP0PgA8W0xrtwiqDkW5E6NpGit0h1E,3498
20
- gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=WMJTEYdkRh5Mro8hcnZ1TIrL6rXk7bDYmsKzjGU8wwA,2848
21
- gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=m7zhI75nXXa9OwZSSnt59Aoy-V2KuApUI2EESm1Puqk,6162
20
+ gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=z4_jEuaprna6AJF2wXH-JgmynpVFXeb2HQZr-0ATmMw,2882
21
+ gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=FxQNV1xGEnHEWUak95OWVAwFPx65z0fphLLQ7R1qJLA,6193
22
22
  gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=9F37VuLYTH5xsPSS_uzhrsPJug0QifX_qh2GwG0jSTU,5345
23
23
  gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=zJZqMvvFKu3sHdrNM773UYjfHVlnwE2w2BmvdFcHzV0,5515
24
24
  gllm_inference/em_invoker/langchain/__init__.pyi,sha256=VYGKE5OgU0my1RlhgzkU_A7-GLGnUDDnNFuctuRwILE,148
@@ -36,28 +36,28 @@ gllm_inference/exceptions/error_parser.pyi,sha256=IOfa--NpLUW5E9Qq0mwWi6ZpTAbUyy
36
36
  gllm_inference/exceptions/exceptions.pyi,sha256=Bv996qLa_vju0Qjf4GewMxdkq8CV9LRZb0S6289DldA,5725
37
37
  gllm_inference/exceptions/provider_error_map.pyi,sha256=P1WnhWkM103FW6hqMfNZBOmYSWOmsJtll3VQV8DGb8E,1210
38
38
  gllm_inference/lm_invoker/__init__.pyi,sha256=IGF3h8Z7Yr0bLrkDMRTDVPNBU6Y3liJabW3acjstJDY,1374
39
- gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=rJeQ9jpUIvcf5z1BB9Lksqf37ZgUzcnFqDMstOl3-kk,17235
40
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=7zF8pj3LbuOwu1QArPX0ra6IrqUq5AkkGbC_wXiGhlA,15027
39
+ gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=inf_fuiy89t9CPiwCs_gsF2k7G7qZ9hMKEKYsGXLtNQ,17237
40
+ gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=Ex91dB8-2lEo2ZKHUJghNGehAfGstQqcGzIKQ75jSUk,15056
41
41
  gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=uZ9wpzOKSOvgu1ICMLqEXcrOE3RIbUmqHmgtuwBekPg,12802
42
- gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=J_tfnIgVDr-zQ7YE5_TKMyZyA336ly04g1l-ZKnr1As,9315
43
- gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=4-3CwfBcDh6thxkidRcYbGVp9bCDkQTemat6VBHsUC8,17164
42
+ gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=T4p2abiP4Vv7psroJ7_rtcSqgzBkFp_B0HGFeIZW3H8,9349
43
+ gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=u4egvGHRG30IotgAN6r_jrjDWFujDxjTdyoIlNLXjYA,17166
44
44
  gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=Bxi9hT4-sIZTz-VhAlxd5JyjAjQZQbUq2xtGhUL6oIs,13705
45
45
  gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=5dEgqLvMB04W30PO2MEUJ4h7rEMmM7ETdQ7LOlLpeXg,13238
46
- gllm_inference/lm_invoker/lm_invoker.pyi,sha256=hjolpN8BzUrhgy8MSpnYxhrlWPJO1LXeCFGlBhQ-eBw,8152
47
- gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=IPx3HXbvEti0LIXBEq1ZPUk6FChVtzBpmISslC9F_Qo,15736
48
- gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=PFOqWhLGHFcoZROW-yreldVaZjHKTzPhuZY2ELM6_SY,3837
49
- gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=oqd50KXEScC3HnS14UhcmPDBgFqL7fw6p1OqGloyxEs,23627
50
- gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=rV8D3E730OUmwK7jELKSziMUl7MnpbfxMAvMuq8-Aew,15687
46
+ gllm_inference/lm_invoker/lm_invoker.pyi,sha256=IDqmtIe9T47mrqJzb0e7dynH-9gbwqhmcb7EeVzPFos,8167
47
+ gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=Z3eAshWJzD8k1Phyo78GUWV8LkdOYOvZqo5ZFWYdJqE,15772
48
+ gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=by0LR_YqP6-uL3VMRq94t_ykO03wGz6Vv71UpWavQtk,3864
49
+ gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=vBo4XEZCmZBiScA6-n5aALJzIU90vsIPRah2p9neEzY,23841
50
+ gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=KtCGKADVbrLfvkFOPHoeLNjlA2TKRf8F_kQHekSvQQI,15689
51
51
  gllm_inference/lm_invoker/batch/__init__.pyi,sha256=W4W-_yfk7lL20alREJai6GnwuQvdlKRfwQCX4mQK4XI,127
52
52
  gllm_inference/lm_invoker/batch/batch_operations.pyi,sha256=Oo7hoyPSfPZdy1mXvSdvtRndvq-XTIbPIjEoGvJj5C0,5372
53
53
  gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
54
54
  gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=6lreMyHKRfZzX5NBYKnQf1Z6RzXBjTvqZj2VbMeaTLQ,1098
55
55
  gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=FJLY-ZkkLUYDV48pfsLatnot4ev_xxz9xAayLK28CpU,1027
56
- gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=aA4DhTXIezwLvFzphR24a5ueVln2FCBIloP9Hbt3iz4,230
56
+ gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=WSuwOqL1j2ZioCZFC-gbB7vTRIZHQ3sU40c3ool5L6c,265
57
57
  gllm_inference/lm_invoker/schema/google.pyi,sha256=AIsNgq0ZZuicHmx4bL7z6q-946T05nWts3HUeA8hhHQ,505
58
58
  gllm_inference/lm_invoker/schema/langchain.pyi,sha256=rZcIxuvABI4pKfyVvkRBRqfJJogZ67EFPydpubHt49c,429
59
- gllm_inference/lm_invoker/schema/openai.pyi,sha256=oju4itbH6mm-yMCqX3m-448XJra4cg6oHHq7abYGM_g,2187
60
- gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=RyULzoGmIHuPwIwlSCg6dFdAYrHhZ0OttNxP2TCMWY8,1164
59
+ gllm_inference/lm_invoker/schema/openai.pyi,sha256=J_rT5Z3rx0hLIae-me1ENeemOESpavcRmYI5pgpkhhk,2222
60
+ gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=8byBRZ4xyTidIQJsZqiSjp5t1X875Obe-aEbT0yYfuA,1199
61
61
  gllm_inference/lm_invoker/schema/xai.pyi,sha256=cWnbJmDtllqRH3NXpQbiXgkNBcUXr8ksDSDywcgJebE,632
62
62
  gllm_inference/model/__init__.pyi,sha256=qClHIgljqhPPCKlGTKmHsWgYb4_hADybxtC2q1U8a5Q,593
63
63
  gllm_inference/model/em/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
@@ -95,13 +95,13 @@ gllm_inference/realtime_chat/output_streamer/output_streamer.pyi,sha256=GPAw1wPS
95
95
  gllm_inference/request_processor/__init__.pyi,sha256=hVnfdNZnkTBJHnmLtN3Na4ANP0yK6AstWdIizVr2Apo,227
96
96
  gllm_inference/request_processor/lm_request_processor.pyi,sha256=VnYc8E3Iayyhw-rPnGPfTKuO3ohgFsS8HPrZJeyES5I,5889
97
97
  gllm_inference/request_processor/uses_lm_mixin.pyi,sha256=Yu0XPNuHxq1tWBviHTPw1oThojneFwGHepvGjBXxKQA,6382
98
- gllm_inference/schema/__init__.pyi,sha256=BJUDYiIo_jPjR5GjWzpSaZ2yOahUFry3FPhgrv-NjK8,1933
99
- gllm_inference/schema/activity.pyi,sha256=kcqSAWAdtyAIyKy2hhLtDgDzZzsNOaUAmHpMAVlGI-s,211
98
+ gllm_inference/schema/__init__.pyi,sha256=bV9_1E-3ce6Dm6Q7xJ6TynPmxQAE1-ujfp9fkRbjI0E,2116
99
+ gllm_inference/schema/activity.pyi,sha256=xGZbfceV8iTJcLtRJrwHfK-dYdhQRmhNwWm7KuckLh4,1974
100
100
  gllm_inference/schema/attachment.pyi,sha256=jApuzjOHJDCz4lr4MlHzBgIndh559nbWu2Xp1fk3hso,3297
101
101
  gllm_inference/schema/code_exec_result.pyi,sha256=ZTHh6JtRrPIdQ059P1UAiD2L-tAO1_S5YcMsAXfJ5A0,559
102
102
  gllm_inference/schema/config.pyi,sha256=rAL_UeXyQeXVk1P2kqd8vFWOMwmKenfpQLtvMP74t9s,674
103
- gllm_inference/schema/enums.pyi,sha256=XQpohUC7_9nFdEmSZHj_4YmOAwM_C5jvTWw_RN-JiFk,901
104
- gllm_inference/schema/events.pyi,sha256=wW14Gb4dS6_YcIus8sBrWvkt_iyw7Xc8E3DrnsQNa08,1566
103
+ gllm_inference/schema/enums.pyi,sha256=5shfuJs76BqiAIZpA1wW1gXe5_KuFikQvuW3ALbQAmM,1322
104
+ gllm_inference/schema/events.pyi,sha256=Ly7B0N0_B3UztJNpT2oqOZ7YArEQp2VZpZ-VFeGc3ok,3911
105
105
  gllm_inference/schema/lm_input.pyi,sha256=A5pjz1id6tP9XRNhzQrbmzd66C_q3gzo0UP8rCemz6Q,193
106
106
  gllm_inference/schema/lm_output.pyi,sha256=15y-M0lpqM_fSlErPKiN1Pj-ikl5NtFBcWLMYsRidt8,2182
107
107
  gllm_inference/schema/mcp.pyi,sha256=Vwu8E2BDl6FvvnI42gIyY3Oki1BdwRE3Uh3aV0rmhQU,1014
@@ -117,7 +117,7 @@ gllm_inference/utils/io_utils.pyi,sha256=7kUTacHAVRYoemFUOjCH7-Qmw-YsQGd6rGYxjf_
117
117
  gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
118
118
  gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
119
119
  gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
120
- gllm_inference_binary-0.5.36.dist-info/METADATA,sha256=FVOLq7mG-Lu_C3as9f33x5eZgNBjLzk5CqsQjyEpGq8,5636
121
- gllm_inference_binary-0.5.36.dist-info/WHEEL,sha256=ar3KUKk5QtasLek_3_fVX4zgPzX-lxtBGErPgJ515rA,105
122
- gllm_inference_binary-0.5.36.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
123
- gllm_inference_binary-0.5.36.dist-info/RECORD,,
120
+ gllm_inference_binary-0.5.38.dist-info/METADATA,sha256=FU3f7oxGCu59ZA1CbA4zZi2in2Bt21J5hJtEJ8zbtUI,5636
121
+ gllm_inference_binary-0.5.38.dist-info/WHEEL,sha256=ar3KUKk5QtasLek_3_fVX4zgPzX-lxtBGErPgJ515rA,105
122
+ gllm_inference_binary-0.5.38.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
123
+ gllm_inference_binary-0.5.38.dist-info/RECORD,,