gllm-inference-binary 0.5.55__cp313-cp313-macosx_13_0_arm64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of gllm-inference-binary might be problematic. Click here for more details.

Files changed (137) hide show
  1. gllm_inference/__init__.pyi +0 -0
  2. gllm_inference/builder/__init__.pyi +6 -0
  3. gllm_inference/builder/_build_invoker.pyi +28 -0
  4. gllm_inference/builder/build_em_invoker.pyi +130 -0
  5. gllm_inference/builder/build_lm_invoker.pyi +213 -0
  6. gllm_inference/builder/build_lm_request_processor.pyi +88 -0
  7. gllm_inference/builder/build_output_parser.pyi +29 -0
  8. gllm_inference/catalog/__init__.pyi +4 -0
  9. gllm_inference/catalog/catalog.pyi +121 -0
  10. gllm_inference/catalog/lm_request_processor_catalog.pyi +112 -0
  11. gllm_inference/catalog/prompt_builder_catalog.pyi +82 -0
  12. gllm_inference/constants.pyi +12 -0
  13. gllm_inference/em_invoker/__init__.pyi +12 -0
  14. gllm_inference/em_invoker/azure_openai_em_invoker.pyi +88 -0
  15. gllm_inference/em_invoker/bedrock_em_invoker.pyi +118 -0
  16. gllm_inference/em_invoker/cohere_em_invoker.pyi +128 -0
  17. gllm_inference/em_invoker/em_invoker.pyi +90 -0
  18. gllm_inference/em_invoker/google_em_invoker.pyi +129 -0
  19. gllm_inference/em_invoker/jina_em_invoker.pyi +103 -0
  20. gllm_inference/em_invoker/langchain/__init__.pyi +3 -0
  21. gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi +84 -0
  22. gllm_inference/em_invoker/langchain_em_invoker.pyi +46 -0
  23. gllm_inference/em_invoker/openai_compatible_em_invoker.pyi +41 -0
  24. gllm_inference/em_invoker/openai_em_invoker.pyi +118 -0
  25. gllm_inference/em_invoker/schema/__init__.pyi +0 -0
  26. gllm_inference/em_invoker/schema/bedrock.pyi +29 -0
  27. gllm_inference/em_invoker/schema/cohere.pyi +20 -0
  28. gllm_inference/em_invoker/schema/google.pyi +9 -0
  29. gllm_inference/em_invoker/schema/jina.pyi +29 -0
  30. gllm_inference/em_invoker/schema/langchain.pyi +5 -0
  31. gllm_inference/em_invoker/schema/openai.pyi +7 -0
  32. gllm_inference/em_invoker/schema/openai_compatible.pyi +7 -0
  33. gllm_inference/em_invoker/schema/twelvelabs.pyi +17 -0
  34. gllm_inference/em_invoker/schema/voyage.pyi +15 -0
  35. gllm_inference/em_invoker/twelevelabs_em_invoker.pyi +101 -0
  36. gllm_inference/em_invoker/voyage_em_invoker.pyi +104 -0
  37. gllm_inference/exceptions/__init__.pyi +4 -0
  38. gllm_inference/exceptions/error_parser.pyi +41 -0
  39. gllm_inference/exceptions/exceptions.pyi +132 -0
  40. gllm_inference/exceptions/provider_error_map.pyi +24 -0
  41. gllm_inference/lm_invoker/__init__.pyi +14 -0
  42. gllm_inference/lm_invoker/anthropic_lm_invoker.pyi +318 -0
  43. gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi +237 -0
  44. gllm_inference/lm_invoker/batch/__init__.pyi +3 -0
  45. gllm_inference/lm_invoker/batch/batch_operations.pyi +127 -0
  46. gllm_inference/lm_invoker/bedrock_lm_invoker.pyi +212 -0
  47. gllm_inference/lm_invoker/datasaur_lm_invoker.pyi +157 -0
  48. gllm_inference/lm_invoker/google_lm_invoker.pyi +327 -0
  49. gllm_inference/lm_invoker/langchain_lm_invoker.pyi +239 -0
  50. gllm_inference/lm_invoker/litellm_lm_invoker.pyi +224 -0
  51. gllm_inference/lm_invoker/lm_invoker.pyi +165 -0
  52. gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi +253 -0
  53. gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi +52 -0
  54. gllm_inference/lm_invoker/openai_lm_invoker.pyi +404 -0
  55. gllm_inference/lm_invoker/portkey_lm_invoker.pyi +296 -0
  56. gllm_inference/lm_invoker/schema/__init__.pyi +0 -0
  57. gllm_inference/lm_invoker/schema/anthropic.pyi +56 -0
  58. gllm_inference/lm_invoker/schema/bedrock.pyi +53 -0
  59. gllm_inference/lm_invoker/schema/datasaur.pyi +14 -0
  60. gllm_inference/lm_invoker/schema/google.pyi +24 -0
  61. gllm_inference/lm_invoker/schema/langchain.pyi +23 -0
  62. gllm_inference/lm_invoker/schema/openai.pyi +106 -0
  63. gllm_inference/lm_invoker/schema/openai_chat_completions.pyi +62 -0
  64. gllm_inference/lm_invoker/schema/portkey.pyi +31 -0
  65. gllm_inference/lm_invoker/schema/xai.pyi +31 -0
  66. gllm_inference/lm_invoker/xai_lm_invoker.pyi +253 -0
  67. gllm_inference/model/__init__.pyi +12 -0
  68. gllm_inference/model/em/__init__.pyi +0 -0
  69. gllm_inference/model/em/cohere_em.pyi +17 -0
  70. gllm_inference/model/em/google_em.pyi +16 -0
  71. gllm_inference/model/em/jina_em.pyi +22 -0
  72. gllm_inference/model/em/openai_em.pyi +15 -0
  73. gllm_inference/model/em/twelvelabs_em.pyi +13 -0
  74. gllm_inference/model/em/voyage_em.pyi +20 -0
  75. gllm_inference/model/lm/__init__.pyi +0 -0
  76. gllm_inference/model/lm/anthropic_lm.pyi +22 -0
  77. gllm_inference/model/lm/google_lm.pyi +18 -0
  78. gllm_inference/model/lm/openai_lm.pyi +27 -0
  79. gllm_inference/model/lm/xai_lm.pyi +19 -0
  80. gllm_inference/output_parser/__init__.pyi +3 -0
  81. gllm_inference/output_parser/json_output_parser.pyi +60 -0
  82. gllm_inference/output_parser/output_parser.pyi +27 -0
  83. gllm_inference/prompt_builder/__init__.pyi +3 -0
  84. gllm_inference/prompt_builder/format_strategy/__init__.pyi +4 -0
  85. gllm_inference/prompt_builder/format_strategy/format_strategy.pyi +55 -0
  86. gllm_inference/prompt_builder/format_strategy/jinja_format_strategy.pyi +45 -0
  87. gllm_inference/prompt_builder/format_strategy/string_format_strategy.pyi +20 -0
  88. gllm_inference/prompt_builder/prompt_builder.pyi +69 -0
  89. gllm_inference/prompt_formatter/__init__.pyi +7 -0
  90. gllm_inference/prompt_formatter/agnostic_prompt_formatter.pyi +49 -0
  91. gllm_inference/prompt_formatter/huggingface_prompt_formatter.pyi +55 -0
  92. gllm_inference/prompt_formatter/llama_prompt_formatter.pyi +59 -0
  93. gllm_inference/prompt_formatter/mistral_prompt_formatter.pyi +53 -0
  94. gllm_inference/prompt_formatter/openai_prompt_formatter.pyi +35 -0
  95. gllm_inference/prompt_formatter/prompt_formatter.pyi +30 -0
  96. gllm_inference/realtime_chat/__init__.pyi +3 -0
  97. gllm_inference/realtime_chat/google_realtime_chat.pyi +205 -0
  98. gllm_inference/realtime_chat/input_streamer/__init__.pyi +4 -0
  99. gllm_inference/realtime_chat/input_streamer/input_streamer.pyi +36 -0
  100. gllm_inference/realtime_chat/input_streamer/keyboard_input_streamer.pyi +27 -0
  101. gllm_inference/realtime_chat/input_streamer/linux_mic_input_streamer.pyi +36 -0
  102. gllm_inference/realtime_chat/output_streamer/__init__.pyi +4 -0
  103. gllm_inference/realtime_chat/output_streamer/console_output_streamer.pyi +21 -0
  104. gllm_inference/realtime_chat/output_streamer/linux_speaker_output_streamer.pyi +42 -0
  105. gllm_inference/realtime_chat/output_streamer/output_streamer.pyi +33 -0
  106. gllm_inference/realtime_chat/realtime_chat.pyi +28 -0
  107. gllm_inference/request_processor/__init__.pyi +4 -0
  108. gllm_inference/request_processor/lm_request_processor.pyi +101 -0
  109. gllm_inference/request_processor/uses_lm_mixin.pyi +130 -0
  110. gllm_inference/schema/__init__.pyi +18 -0
  111. gllm_inference/schema/activity.pyi +64 -0
  112. gllm_inference/schema/attachment.pyi +88 -0
  113. gllm_inference/schema/code_exec_result.pyi +14 -0
  114. gllm_inference/schema/config.pyi +15 -0
  115. gllm_inference/schema/enums.pyi +80 -0
  116. gllm_inference/schema/events.pyi +105 -0
  117. gllm_inference/schema/lm_input.pyi +4 -0
  118. gllm_inference/schema/lm_output.pyi +188 -0
  119. gllm_inference/schema/mcp.pyi +31 -0
  120. gllm_inference/schema/message.pyi +52 -0
  121. gllm_inference/schema/model_id.pyi +176 -0
  122. gllm_inference/schema/reasoning.pyi +15 -0
  123. gllm_inference/schema/token_usage.pyi +75 -0
  124. gllm_inference/schema/tool_call.pyi +14 -0
  125. gllm_inference/schema/tool_result.pyi +11 -0
  126. gllm_inference/schema/type_alias.pyi +11 -0
  127. gllm_inference/utils/__init__.pyi +5 -0
  128. gllm_inference/utils/io_utils.pyi +26 -0
  129. gllm_inference/utils/langchain.pyi +30 -0
  130. gllm_inference/utils/validation.pyi +12 -0
  131. gllm_inference.build/.gitignore +1 -0
  132. gllm_inference.cpython-313-darwin.so +0 -0
  133. gllm_inference.pyi +154 -0
  134. gllm_inference_binary-0.5.55.dist-info/METADATA +138 -0
  135. gllm_inference_binary-0.5.55.dist-info/RECORD +137 -0
  136. gllm_inference_binary-0.5.55.dist-info/WHEEL +5 -0
  137. gllm_inference_binary-0.5.55.dist-info/top_level.txt +1 -0
@@ -0,0 +1,75 @@
1
+ from pydantic import BaseModel
2
+
3
+ class InputTokenDetails(BaseModel):
4
+ """Defines the input token details schema.
5
+
6
+ Attributes:
7
+ cached_tokens (int): The number of cached tokens. Defaults to 0.
8
+ uncached_tokens (int): The number of uncached tokens. Defaults to 0.
9
+ """
10
+ cached_tokens: int
11
+ uncached_tokens: int
12
+ def __add__(self, other: InputTokenDetails) -> InputTokenDetails:
13
+ """Add two InputTokenDetails objects together.
14
+
15
+ Args:
16
+ other (InputTokenDetails): The other InputTokenDetails object to add.
17
+
18
+ Returns:
19
+ InputTokenDetails: A new InputTokenDetails object with summed values.
20
+ """
21
+
22
+ class OutputTokenDetails(BaseModel):
23
+ """Defines the output token details schema.
24
+
25
+ Attributes:
26
+ reasoning_tokens (int): The number of reasoning tokens. Defaults to 0.
27
+ response_tokens (int): The number of response tokens. Defaults to 0.
28
+ """
29
+ reasoning_tokens: int
30
+ response_tokens: int
31
+ def __add__(self, other: OutputTokenDetails) -> OutputTokenDetails:
32
+ """Add two OutputTokenDetails objects together.
33
+
34
+ Args:
35
+ other (OutputTokenDetails): The other OutputTokenDetails object to add.
36
+
37
+ Returns:
38
+ OutputTokenDetails: A new OutputTokenDetails object with summed values.
39
+ """
40
+
41
+ class TokenUsage(BaseModel):
42
+ """Defines the token usage data structure of a language model.
43
+
44
+ Attributes:
45
+ input_tokens (int): The number of input tokens. Defaults to 0.
46
+ output_tokens (int): The number of output tokens. Defaults to 0.
47
+ input_token_details (InputTokenDetails | None): The details of the input tokens. Defaults to None.
48
+ output_token_details (OutputTokenDetails | None): The details of the output tokens. Defaults to None.
49
+ """
50
+ input_tokens: int
51
+ output_tokens: int
52
+ input_token_details: InputTokenDetails | None
53
+ output_token_details: OutputTokenDetails | None
54
+ @classmethod
55
+ def from_token_details(cls, input_tokens: int | None = None, output_tokens: int | None = None, cached_tokens: int | None = None, reasoning_tokens: int | None = None) -> TokenUsage:
56
+ """Creates a TokenUsage from token details.
57
+
58
+ Args:
59
+ input_tokens (int | None): The number of input tokens. Defaults to None.
60
+ output_tokens (int | None): The number of output tokens. Defaults to None.
61
+ cached_tokens (int | None): The number of cached tokens. Defaults to None.
62
+ reasoning_tokens (int | None): The number of reasoning tokens. Defaults to None.
63
+
64
+ Returns:
65
+ TokenUsage: The instantiated TokenUsage.
66
+ """
67
+ def __add__(self, other: TokenUsage) -> TokenUsage:
68
+ """Add two TokenUsage objects together.
69
+
70
+ Args:
71
+ other (TokenUsage): The other TokenUsage object to add.
72
+
73
+ Returns:
74
+ TokenUsage: A new TokenUsage object with summed values.
75
+ """
@@ -0,0 +1,14 @@
1
+ from pydantic import BaseModel
2
+ from typing import Any
3
+
4
+ class ToolCall(BaseModel):
5
+ """Defines a tool call request when a language model decides to invoke a tool.
6
+
7
+ Attributes:
8
+ id (str): The ID of the tool call.
9
+ name (str): The name of the tool.
10
+ args (dict[str, Any]): The arguments of the tool call.
11
+ """
12
+ id: str
13
+ name: str
14
+ args: dict[str, Any]
@@ -0,0 +1,11 @@
1
+ from pydantic import BaseModel
2
+
3
+ class ToolResult(BaseModel):
4
+ """Defines a tool result to be sent back to the language model.
5
+
6
+ Attributes:
7
+ id (str): The ID of the tool call.
8
+ output (str): The output of the tool call.
9
+ """
10
+ id: str
11
+ output: str
@@ -0,0 +1,11 @@
1
+ from gllm_inference.schema.attachment import Attachment as Attachment
2
+ from gllm_inference.schema.reasoning import Reasoning as Reasoning
3
+ from gllm_inference.schema.tool_call import ToolCall as ToolCall
4
+ from gllm_inference.schema.tool_result import ToolResult as ToolResult
5
+ from pydantic import BaseModel
6
+ from typing import Any
7
+
8
+ ResponseSchema = dict[str, Any] | type[BaseModel]
9
+ MessageContent = str | Attachment | ToolCall | ToolResult | Reasoning
10
+ EMContent = str | Attachment | tuple[str | Attachment, ...]
11
+ Vector = list[float]
@@ -0,0 +1,5 @@
1
+ from gllm_inference.utils.io_utils import base64_to_bytes as base64_to_bytes
2
+ from gllm_inference.utils.langchain import load_langchain_model as load_langchain_model, parse_model_data as parse_model_data
3
+ from gllm_inference.utils.validation import validate_string_enum as validate_string_enum
4
+
5
+ __all__ = ['base64_to_bytes', 'load_langchain_model', 'parse_model_data', 'validate_string_enum']
@@ -0,0 +1,26 @@
1
+ from _typeshed import Incomplete
2
+
3
+ logger: Incomplete
4
+ DEFAULT_BASE64_ALLOWED_MIMETYPES: Incomplete
5
+
6
+ def base64_to_bytes(value: str, *, allowed_mimetypes: tuple[str, ...] | None = ...) -> str | bytes:
7
+ '''Decode a base64 string to bytes based on allowed MIME type.
8
+
9
+ The conversion steps are as follows:
10
+ 1. The function first attempts to decode the given string from base64.
11
+ 2. If decoding succeeds, it checks the MIME type of the decoded content.
12
+ 3. When the MIME type matches one of the allowed patterns (e.g., ``"image/*"``),
13
+ the raw bytes are returned. Otherwise, the original string is returned unchanged.
14
+
15
+ Args:
16
+ value (str): Input data to decode.
17
+ allowed_mimetypes (tuple[str, ...], optional): MIME type prefixes that are allowed
18
+ to be decoded into bytes. Defaults to ("image/*", "audio/*", "video/*").
19
+
20
+ Returns:
21
+ str | bytes: Base64-encoded string or raw bytes if MIME type is allowed;
22
+ otherwise returns original string.
23
+
24
+ Raises:
25
+ ValueError: If the input is not a string.
26
+ '''
@@ -0,0 +1,30 @@
1
+ from _typeshed import Incomplete
2
+ from langchain_core.embeddings import Embeddings as Embeddings
3
+ from langchain_core.language_models import BaseChatModel as BaseChatModel
4
+ from typing import Any
5
+
6
+ MODEL_NAME_KEYS: Incomplete
7
+
8
+ def load_langchain_model(model_class_path: str, model_name: str, model_kwargs: dict[str, Any]) -> BaseChatModel | Embeddings:
9
+ '''Loads the LangChain\'s model instance.
10
+
11
+ Args:
12
+ model_class_path (str): The path to the LangChain\'s class, e.g. "langchain_openai.ChatOpenAI".
13
+ model_name (str): The model name.
14
+ model_kwargs (dict[str, Any]): The additional keyword arguments.
15
+
16
+ Returns:
17
+ BaseChatModel | Embeddings: The LangChain\'s model instance.
18
+ '''
19
+ def parse_model_data(model: BaseChatModel | Embeddings) -> dict[str, str]:
20
+ """Parses the model data from LangChain's BaseChatModel or Embeddings instance.
21
+
22
+ Args:
23
+ model (BaseChatModel | Embeddings): The LangChain's BaseChatModel or Embeddings instance.
24
+
25
+ Returns:
26
+ dict[str, str]: The dictionary containing the model name and path.
27
+
28
+ Raises:
29
+ ValueError: If the model name is not found in the model data.
30
+ """
@@ -0,0 +1,12 @@
1
+ from enum import StrEnum
2
+
3
+ def validate_string_enum(enum_type: type[StrEnum], value: str) -> None:
4
+ """Validates that the provided value is a valid string enum value.
5
+
6
+ Args:
7
+ enum_type (type[StrEnum]): The type of the string enum.
8
+ value (str): The value to validate.
9
+
10
+ Raises:
11
+ ValueError: If the provided value is not a valid string enum value.
12
+ """
@@ -0,0 +1 @@
1
+ *
Binary file
gllm_inference.pyi ADDED
@@ -0,0 +1,154 @@
1
+ # This file was generated by Nuitka
2
+
3
+ # Stubs included by default
4
+
5
+
6
+ __name__ = ...
7
+
8
+
9
+
10
+ # Modules used internally, to allow implicit dependencies to be seen:
11
+ import os
12
+ import typing
13
+ import gllm_core
14
+ import gllm_core.utils
15
+ import gllm_inference.em_invoker.AzureOpenAIEMInvoker
16
+ import gllm_inference.em_invoker.BedrockEMInvoker
17
+ import gllm_inference.em_invoker.CohereEMInvoker
18
+ import gllm_inference.em_invoker.GoogleEMInvoker
19
+ import gllm_inference.em_invoker.JinaEMInvoker
20
+ import gllm_inference.em_invoker.LangChainEMInvoker
21
+ import gllm_inference.em_invoker.OpenAICompatibleEMInvoker
22
+ import gllm_inference.em_invoker.OpenAIEMInvoker
23
+ import gllm_inference.em_invoker.TwelveLabsEMInvoker
24
+ import gllm_inference.em_invoker.VoyageEMInvoker
25
+ import gllm_inference.lm_invoker.AnthropicLMInvoker
26
+ import gllm_inference.lm_invoker.AzureOpenAILMInvoker
27
+ import gllm_inference.lm_invoker.BedrockLMInvoker
28
+ import gllm_inference.lm_invoker.DatasaurLMInvoker
29
+ import gllm_inference.lm_invoker.GoogleLMInvoker
30
+ import gllm_inference.lm_invoker.LangChainLMInvoker
31
+ import gllm_inference.lm_invoker.LiteLLMLMInvoker
32
+ import gllm_inference.lm_invoker.OpenAIChatCompletionsLMInvoker
33
+ import gllm_inference.lm_invoker.OpenAICompatibleLMInvoker
34
+ import gllm_inference.lm_invoker.OpenAILMInvoker
35
+ import gllm_inference.lm_invoker.PortkeyLMInvoker
36
+ import gllm_inference.lm_invoker.XAILMInvoker
37
+ import gllm_inference.prompt_builder.PromptBuilder
38
+ import gllm_inference.output_parser.JSONOutputParser
39
+ import json
40
+ import abc
41
+ import pandas
42
+ import pydantic
43
+ import re
44
+ import gllm_core.utils.retry
45
+ import gllm_inference.request_processor.LMRequestProcessor
46
+ import gllm_core.utils.imports
47
+ import gllm_inference.schema.ModelId
48
+ import gllm_inference.schema.ModelProvider
49
+ import gllm_inference.schema.TruncationConfig
50
+ import asyncio
51
+ import base64
52
+ import enum
53
+ import gllm_inference.exceptions.BaseInvokerError
54
+ import gllm_inference.exceptions.convert_http_status_to_base_invoker_error
55
+ import gllm_inference.schema.Attachment
56
+ import gllm_inference.schema.AttachmentType
57
+ import gllm_inference.schema.EMContent
58
+ import gllm_inference.schema.Vector
59
+ import aioboto3
60
+ import gllm_inference.utils.validate_string_enum
61
+ import cohere
62
+ import asyncio.CancelledError
63
+ import gllm_inference.exceptions.convert_to_base_invoker_error
64
+ import gllm_inference.schema.TruncateSide
65
+ import google
66
+ import google.auth
67
+ import google.genai
68
+ import google.genai.types
69
+ import httpx
70
+ import gllm_inference.exceptions.ProviderInternalError
71
+ import gllm_core.utils.concurrency
72
+ import langchain_core
73
+ import langchain_core.embeddings
74
+ import gllm_inference.exceptions.InvokerRuntimeError
75
+ import gllm_inference.exceptions.build_debug_info
76
+ import gllm_inference.utils.load_langchain_model
77
+ import gllm_inference.utils.parse_model_data
78
+ import openai
79
+ import io
80
+ import twelvelabs
81
+ import sys
82
+ import voyageai
83
+ import voyageai.client_async
84
+ import http
85
+ import http.HTTPStatus
86
+ import uuid
87
+ import gllm_core.constants
88
+ import gllm_core.event
89
+ import gllm_core.schema
90
+ import gllm_core.schema.tool
91
+ import langchain_core.tools
92
+ import gllm_inference.schema.BatchStatus
93
+ import gllm_inference.schema.LMInput
94
+ import gllm_inference.schema.LMOutput
95
+ import gllm_inference.schema.Message
96
+ import gllm_inference.schema.Reasoning
97
+ import gllm_inference.schema.ResponseSchema
98
+ import gllm_inference.schema.ThinkingEvent
99
+ import gllm_inference.schema.TokenUsage
100
+ import gllm_inference.schema.ToolCall
101
+ import gllm_inference.schema.ToolResult
102
+ import anthropic
103
+ import anthropic.types
104
+ import anthropic.types.message_create_params
105
+ import anthropic.types.messages
106
+ import anthropic.types.messages.batch_create_params
107
+ import gllm_inference.schema.MessageRole
108
+ import langchain_core.language_models
109
+ import langchain_core.messages
110
+ import gllm_inference.exceptions._get_exception_key
111
+ import litellm
112
+ import inspect
113
+ import time
114
+ import jsonschema
115
+ import gllm_inference.lm_invoker.batch.BatchOperations
116
+ import gllm_inference.schema.MessageContent
117
+ import __future__
118
+ import gllm_inference.schema.ActivityEvent
119
+ import gllm_inference.schema.CodeEvent
120
+ import gllm_inference.schema.CodeExecResult
121
+ import gllm_inference.schema.MCPCall
122
+ import gllm_inference.schema.MCPCallActivity
123
+ import gllm_inference.schema.MCPListToolsActivity
124
+ import gllm_inference.schema.MCPServer
125
+ import gllm_inference.schema.WebSearchActivity
126
+ import logging
127
+ import portkey_ai
128
+ import xai_sdk
129
+ import xai_sdk.chat
130
+ import xai_sdk.search
131
+ import xai_sdk.proto
132
+ import xai_sdk.proto.v5
133
+ import xai_sdk.proto.v5.chat_pb2
134
+ import jinja2
135
+ import jinja2.sandbox
136
+ import gllm_inference.schema.JinjaEnvType
137
+ import gllm_inference.prompt_builder.format_strategy.JinjaFormatStrategy
138
+ import gllm_inference.prompt_builder.format_strategy.StringFormatStrategy
139
+ import transformers
140
+ import gllm_inference.prompt_formatter.HuggingFacePromptFormatter
141
+ import traceback
142
+ import gllm_inference.realtime_chat.input_streamer.KeyboardInputStreamer
143
+ import gllm_inference.realtime_chat.output_streamer.ConsoleOutputStreamer
144
+ import google.genai.live
145
+ import gllm_core.utils.logger_manager
146
+ import mimetypes
147
+ import pathlib
148
+ import pathlib.Path
149
+ import filetype
150
+ import magic
151
+ import requests
152
+ import binascii
153
+ import fnmatch
154
+ import importlib
@@ -0,0 +1,138 @@
1
+ Metadata-Version: 2.2
2
+ Name: gllm-inference-binary
3
+ Version: 0.5.55
4
+ Summary: A library containing components related to model inferences in Gen AI applications.
5
+ Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, Resti Febrina <resti.febrina@gdplabs.id>
6
+ Requires-Python: <3.14,>=3.11
7
+ Description-Content-Type: text/markdown
8
+ Requires-Dist: poetry<3.0.0,>=2.1.3
9
+ Requires-Dist: gllm-core-binary<0.4.0,>=0.3.23
10
+ Requires-Dist: aiohttp<4.0.0,>=3.12.14
11
+ Requires-Dist: filetype<2.0.0,>=1.2.0
12
+ Requires-Dist: httpx<0.29.0,>=0.28.0
13
+ Requires-Dist: jinja2<4.0.0,>=3.1.4
14
+ Requires-Dist: jsonschema<5.0.0,>=4.24.0
15
+ Requires-Dist: langchain<0.4.0,>=0.3.0
16
+ Requires-Dist: pandas<3.0.0,>=2.2.3
17
+ Requires-Dist: protobuf<7.0.0,>=5.29.4
18
+ Requires-Dist: python-magic<0.5.0,>=0.4.27; sys_platform != "win32"
19
+ Requires-Dist: python-magic-bin<0.5.0,>=0.4.14; sys_platform == "win32"
20
+ Requires-Dist: sentencepiece<0.3.0,>=0.2.0
21
+ Provides-Extra: dev
22
+ Requires-Dist: coverage<8.0.0,>=7.4.4; extra == "dev"
23
+ Requires-Dist: mypy<2.0.0,>=1.15.0; extra == "dev"
24
+ Requires-Dist: pre-commit<4.0.0,>=3.7.0; extra == "dev"
25
+ Requires-Dist: pytest<9.0.0,>=8.1.1; extra == "dev"
26
+ Requires-Dist: pytest-asyncio<0.24.0,>=0.23.6; extra == "dev"
27
+ Requires-Dist: pytest-cov<6.0.0,>=5.0.0; extra == "dev"
28
+ Requires-Dist: ruff<0.7.0,>=0.6.7; extra == "dev"
29
+ Provides-Extra: anthropic
30
+ Requires-Dist: anthropic<0.61.0,>=0.60.0; extra == "anthropic"
31
+ Provides-Extra: bedrock
32
+ Requires-Dist: aioboto3<16.0.0,>=15.0.0; extra == "bedrock"
33
+ Provides-Extra: cohere
34
+ Requires-Dist: cohere<6.0.0,>=5.18.0; extra == "cohere"
35
+ Provides-Extra: datasaur
36
+ Requires-Dist: openai<2.0.0,>=1.98.0; extra == "datasaur"
37
+ Provides-Extra: google
38
+ Requires-Dist: google-genai<=1.36,>=1.23; extra == "google"
39
+ Provides-Extra: huggingface
40
+ Requires-Dist: huggingface-hub<0.31.0,>=0.30.0; extra == "huggingface"
41
+ Requires-Dist: transformers<5.0.0,>=4.52.0; extra == "huggingface"
42
+ Provides-Extra: litellm
43
+ Requires-Dist: litellm<2.0.0,>=1.69.2; extra == "litellm"
44
+ Provides-Extra: openai
45
+ Requires-Dist: openai<2.0.0,>=1.98.0; extra == "openai"
46
+ Provides-Extra: portkey-ai
47
+ Requires-Dist: portkey-ai<2.0.0,>=1.14.4; extra == "portkey-ai"
48
+ Provides-Extra: twelvelabs
49
+ Requires-Dist: twelvelabs<0.5.0,>=0.4.4; extra == "twelvelabs"
50
+ Provides-Extra: voyage
51
+ Requires-Dist: voyageai<0.4.0,>=0.3.0; python_version < "3.13" and extra == "voyage"
52
+ Provides-Extra: xai
53
+ Requires-Dist: xai_sdk<2.0.0,>=1.0.0; extra == "xai"
54
+
55
+ # GLLM Inference
56
+
57
+ ## Description
58
+
59
+ A library containing components related to model inferences in Gen AI applications.
60
+
61
+ ## Installation
62
+
63
+ ### Prerequisites
64
+ 1. Python 3.11+ - [Install here](https://www.python.org/downloads/)
65
+ 2. Pip (if using Pip) - [Install here](https://pip.pypa.io/en/stable/installation/)
66
+ 3. Poetry (automatically installed via Makefile) - [Install here](https://python-poetry.org/docs/#installation)
67
+ 4. Git (if using Git) - [Install here](https://git-scm.com/downloads)
68
+ 5. gcloud CLI (for authentication) - [Install here](https://cloud.google.com/sdk/docs/install)
69
+ 6. For git installation, access to the [GDP Labs SDK github repository](https://github.com/GDP-ADMIN/gl-sdk)
70
+
71
+ ### 1. Installation from Artifact Registry
72
+ Choose one of the following methods to install the package:
73
+
74
+ #### Using pip
75
+ ```bash
76
+ pip install gllm-inference-binary
77
+ ```
78
+
79
+ #### Using Poetry
80
+ ```bash
81
+ poetry add gllm-inference-binary
82
+ ```
83
+
84
+ ### 2. Development Installation (Git)
85
+ For development purposes, you can install directly from the Git repository:
86
+ ```bash
87
+ poetry add "git+ssh://git@github.com/GDP-ADMIN/gen-ai-internal.git#subdirectory=libs/gllm-inference"
88
+ ```
89
+
90
+ Available extras:
91
+ - `anthropic`: Install Anthropic models dependencies
92
+ - `google-genai`: Install Google Generative AI models dependencies
93
+ - `google-vertexai`: Install Google Vertex AI models dependencies
94
+ - `huggingface`: Install HuggingFace models dependencies
95
+ - `openai`: Install OpenAI models dependencies
96
+ - `twelvelabs`: Install TwelveLabs models dependencies
97
+
98
+ ## Local Development Setup
99
+
100
+ ### Quick Setup (Recommended)
101
+ For local development with editable gllm packages, use the provided Makefile:
102
+
103
+ ```bash
104
+ # Complete setup: installs Poetry, configures auth, installs packages, sets up pre-commit
105
+ make setup
106
+ ```
107
+
108
+ The following are the available Makefile targets:
109
+
110
+ 1. `make setup` - Complete development setup (recommended for new developers)
111
+ 2. `make install-poetry` - Install or upgrade Poetry to the latest version
112
+ 3. `make auth` - Configure authentication for internal repositories
113
+ 4. `make install` - Install all dependencies
114
+ 5. `make install-pre-commit` - Set up pre-commit hooks
115
+ 6. `make update` - Update dependencies
116
+ ### Manual Development Setup (Legacy)
117
+ If you prefer to manage dependencies manually:
118
+
119
+ 1. Go to root folder of `gllm-inference` module, e.g. `cd libs/gllm-inference`.
120
+ 2. Run `poetry shell` to create a virtual environment.
121
+ 3. Run `poetry lock` to create a lock file if you haven't done it yet.
122
+ 4. Run `poetry install` to install the `gllm-inference` requirements for the first time.
123
+ 5. Run `poetry update` if you update any dependency module version at `pyproject.toml`.
124
+
125
+
126
+ ## Contributing
127
+ Please refer to this [Python Style Guide](https://docs.google.com/document/d/1uRggCrHnVfDPBnG641FyQBwUwLoFw0kTzNqRm92vUwM/edit?usp=sharing)
128
+ to get information about code style, documentation standard, and SCA that you need to use when contributing to this project
129
+
130
+ ### Getting Started with Development
131
+ 1. Clone the repository and navigate to the gllm-inference directory
132
+ 2. Run `make setup` to set up your development environment
133
+ 3. Run `which python` to get the path to be referenced at Visual Studio Code interpreter path (`Ctrl`+`Shift`+`P` or `Cmd`+`Shift`+`P`)
134
+ 4. Try running the unit test to see if it's working:
135
+ ```bash
136
+ poetry run pytest -s tests/unit_tests/
137
+ ```
138
+ 5. When you want to update the dependencies, run `make update`
@@ -0,0 +1,137 @@
1
+ gllm_inference.cpython-313-darwin.so,sha256=HXwpg2rVBMxFD0jE_gpu9Hrq3xPHXBEafPwU4JGBOuY,5057880
2
+ gllm_inference.pyi,sha256=OzEB4lPytQEVVGAnTVtmkYXAlqaH-jnpZAXphCx9aV4,5116
3
+ gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
4
+ gllm_inference/constants.pyi,sha256=tBFhwE1at2gXMJ1bBM32eVIRgCJlB1uzg7ItXGx3RQE,316
5
+ gllm_inference/builder/__init__.pyi,sha256=usz2lvfwO4Yk-ZGKXbCWG1cEr3nlQXxMNDNC-2yc1NM,500
6
+ gllm_inference/builder/_build_invoker.pyi,sha256=v__-YT0jGmLqEsgl7Abk9we-wYWcyVFvlRN_Uu4vVak,848
7
+ gllm_inference/builder/build_em_invoker.pyi,sha256=0IVcRGaciPBAFcgvGRC4-Kje_PPXOSug77LqxRc-x_U,5749
8
+ gllm_inference/builder/build_lm_invoker.pyi,sha256=TBAr7Sk9Jgckdlfj69pYOzZFMsooL-FpfbPYt_kCcXU,8852
9
+ gllm_inference/builder/build_lm_request_processor.pyi,sha256=KbQkcPa8C-yzyelht4mWLP8kDmh17itAT3tn8ZJB6pg,4144
10
+ gllm_inference/builder/build_output_parser.pyi,sha256=_Lrq-bh1oPsb_Nwkkr_zyEUwIOMysRFZkvEtEM29LZM,936
11
+ gllm_inference/catalog/__init__.pyi,sha256=JBkPGTyiiZ30GECzJBW-mW8LekWyY2qyzal3eW7ynaM,287
12
+ gllm_inference/catalog/catalog.pyi,sha256=a4RNG1lKv51GxQpOqh47tz-PAROMPaeP2o5XNLBSZaU,4790
13
+ gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=WW1j8jWujnatF0c9rCk94CyobtFe3gOky2vjoGCV6nw,5424
14
+ gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=OU8k_4HbqjZEzHZlzSM3uzGQZJmM2uGD76Csqom0CEQ,3197
15
+ gllm_inference/em_invoker/__init__.pyi,sha256=FHFR4k8RqtdMNZkda6A89tvWFtIoxF-t_NahDGeBQ2k,1199
16
+ gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=fE-RZ10Leerbyqp8ISXeqdU_m0kaHdlHml_nQqwFjqY,4961
17
+ gllm_inference/em_invoker/bedrock_em_invoker.pyi,sha256=3GkVO9hvmYck4NJY5spAg6opSxr0v8OJ4qhMvz1rcAA,6365
18
+ gllm_inference/em_invoker/cohere_em_invoker.pyi,sha256=LSNVGx58JdSP2YB6CziOS_v5VuSLdARVpBBWhz8t5E4,6665
19
+ gllm_inference/em_invoker/em_invoker.pyi,sha256=KGjLiAWGIA3ziV50zMwSzx6lTVUbYspQCl4LFxqbDlY,5101
20
+ gllm_inference/em_invoker/google_em_invoker.pyi,sha256=OgIQhT2g6g09Hqr9mc2--MjV3Ti-XuHfqLTunvcqU1k,6824
21
+ gllm_inference/em_invoker/jina_em_invoker.pyi,sha256=WZJ8BeaqxrAGMrUtIEUy3itcOEzvRQYEwe0gazIv6ng,5651
22
+ gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=BhZjkYZoxQXPJjP0PgA8W0xrtwiqDkW5E6NpGit0h1E,3498
23
+ gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=z4_jEuaprna6AJF2wXH-JgmynpVFXeb2HQZr-0ATmMw,2882
24
+ gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=tb6cXTvAGr01MMSc96mYWDqK-EK_VCdDOQ-Be0_EZfM,6195
25
+ gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=mitrJZW3M43k306cHwJoDjBfN1nr20C4sZ42Xr6hfO0,5347
26
+ gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=gyYQA3rbekHJIvvY6TE4tTt8W-DbPqjiPW6FaJoQPHI,5517
27
+ gllm_inference/em_invoker/langchain/__init__.pyi,sha256=VYGKE5OgU0my1RlhgzkU_A7-GLGnUDDnNFuctuRwILE,148
28
+ gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi,sha256=VU3-Vhb9BCDhJo8PPdWHe2rBEOCs_HMXT6ZaWwjUzZE,3304
29
+ gllm_inference/em_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
30
+ gllm_inference/em_invoker/schema/bedrock.pyi,sha256=ePYwailnZAYbZKt9HRDi4iAISq003jpxQPxDCiFdEZg,615
31
+ gllm_inference/em_invoker/schema/cohere.pyi,sha256=UTbTtePRR1zJMsM09SiTZSZZP0IaUGaODvc7ZqH9S8c,547
32
+ gllm_inference/em_invoker/schema/google.pyi,sha256=ovDlvinu99QJhIxMkvVUoGBEFkkEoAZhadSuk0nI9N8,181
33
+ gllm_inference/em_invoker/schema/jina.pyi,sha256=vE1ySd8OTDM35saEZos7UCdPwHeX66iuHkZ3RchSZKA,741
34
+ gllm_inference/em_invoker/schema/langchain.pyi,sha256=edcUvc1IHoSMFwqV83uqWqd0U3fLhkyWQjVknvjHI8U,112
35
+ gllm_inference/em_invoker/schema/openai.pyi,sha256=Q_dsEcodkOXYXPdrkOkW0LnuLhfeq8tEbtZAGMz2ajA,139
36
+ gllm_inference/em_invoker/schema/openai_compatible.pyi,sha256=gmvGtsWoOMBelke_tZjC6dKimFBW9f4Vrgv0Ig0OM9Q,150
37
+ gllm_inference/em_invoker/schema/twelvelabs.pyi,sha256=F6wKHgG01bYskJpKoheBSpRpHUfFpteKn9sj9n5YfcU,372
38
+ gllm_inference/em_invoker/schema/voyage.pyi,sha256=HVpor0fqNy-IwapCICfsgFmqf1FJXCOMIxS2vOXhHd8,289
39
+ gllm_inference/exceptions/__init__.pyi,sha256=Upcuj7od2lkbdueQ0iMT2ktFYYi-KKTynTLAaxWDTjU,1214
40
+ gllm_inference/exceptions/error_parser.pyi,sha256=IOfa--NpLUW5E9Qq0mwWi6ZpTAbUyyNe6iAqunBNGLI,1999
41
+ gllm_inference/exceptions/exceptions.pyi,sha256=Bv996qLa_vju0Qjf4GewMxdkq8CV9LRZb0S6289DldA,5725
42
+ gllm_inference/exceptions/provider_error_map.pyi,sha256=XPLWU42-r8MHZgg5ZkE80Gdqg3p8Z_JHvq_Na03iTqY,1243
43
+ gllm_inference/lm_invoker/__init__.pyi,sha256=Ze9CxgGYguyz8BAU87_2JM-D4OZjlYAqktLI_B2tj_s,1488
44
+ gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=oU0dzg87OHTVPGhKBT8WdcdR0AzeJJNSSVvGKtY0UQU,15178
45
+ gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=drtMgbDzBQJrWXLiI2t5PNy7HtcW5Kuj0XR2b6rltjc,12936
46
+ gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=9Gz0U2c94UM9SOt-_e89_NqT_bDQ6wECRiJ9VTwsqfw,10739
47
+ gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=IqvDxBzwEf2z34FZcLKIH404y386Rnk5gsj2TcAm424,7878
48
+ gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=-trlpAmc5APGohAmVAsRytTVQ2ODRCmRtRm3EKiOR2Q,16359
49
+ gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=OzISl89C2s-qB6VxNlMgf5dFRC-ooj30YCFfsZzcX4s,11887
50
+ gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=0PZYitAljAjzyympqqNyN5fMyoakmqr1XIz1PE6NNc4,11176
51
+ gllm_inference/lm_invoker/lm_invoker.pyi,sha256=pJ0-s37NqTHdFD7IijvNzJnQ7JXgrGxsEaXuS8cxz3s,8487
52
+ gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=6-44DBp9yItKR0b1nyIFIS0KnAXctqFqcuHhwszDGyA,13720
53
+ gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=i5pMpZf4-r_7FQ1qfsqcjpc98sI-cPiqheuTfTEKxJs,4192
54
+ gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=AlOKeTpzOXYCwwhHewG0WR7KpIqcolWZ5ZhLRMUj_eM,21178
55
+ gllm_inference/lm_invoker/portkey_lm_invoker.pyi,sha256=ewxzRT-ekmvCFeb7ij840s4p18AO_LAKA-UP0ot12hs,14940
56
+ gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=ZLwgugdsGW9B3YA4YU_GyyAvYprJD_VhN6SBvhT-t8E,12865
57
+ gllm_inference/lm_invoker/batch/__init__.pyi,sha256=W4W-_yfk7lL20alREJai6GnwuQvdlKRfwQCX4mQK4XI,127
58
+ gllm_inference/lm_invoker/batch/batch_operations.pyi,sha256=Oo7hoyPSfPZdy1mXvSdvtRndvq-XTIbPIjEoGvJj5C0,5372
59
+ gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
60
+ gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=6lreMyHKRfZzX5NBYKnQf1Z6RzXBjTvqZj2VbMeaTLQ,1098
61
+ gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=FJLY-ZkkLUYDV48pfsLatnot4ev_xxz9xAayLK28CpU,1027
62
+ gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=WSuwOqL1j2ZioCZFC-gbB7vTRIZHQ3sU40c3ool5L6c,265
63
+ gllm_inference/lm_invoker/schema/google.pyi,sha256=AIsNgq0ZZuicHmx4bL7z6q-946T05nWts3HUeA8hhHQ,505
64
+ gllm_inference/lm_invoker/schema/langchain.pyi,sha256=rZcIxuvABI4pKfyVvkRBRqfJJogZ67EFPydpubHt49c,429
65
+ gllm_inference/lm_invoker/schema/openai.pyi,sha256=J_rT5Z3rx0hLIae-me1ENeemOESpavcRmYI5pgpkhhk,2222
66
+ gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=8byBRZ4xyTidIQJsZqiSjp5t1X875Obe-aEbT0yYfuA,1199
67
+ gllm_inference/lm_invoker/schema/portkey.pyi,sha256=NeRjHNd84HgE_ur2F3Cv6Jx30v6V7eQvI_iJiq4kuME,631
68
+ gllm_inference/lm_invoker/schema/xai.pyi,sha256=cWnbJmDtllqRH3NXpQbiXgkNBcUXr8ksDSDywcgJebE,632
69
+ gllm_inference/model/__init__.pyi,sha256=LTeBCSJJwCSd5Qrg7RZCXcp9fURNVNXFR5akk1ZZrTk,810
70
+ gllm_inference/model/em/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
71
+ gllm_inference/model/em/cohere_em.pyi,sha256=fArRlV08NwbsJ_h6vpWr94XxUVBtbqW1Jh8s42LRXCo,488
72
+ gllm_inference/model/em/google_em.pyi,sha256=ZPN5LmReO0bcTfnZixFooUTzgD-daNFPzfxzZ-5WzQQ,471
73
+ gllm_inference/model/em/jina_em.pyi,sha256=txEvDI61nhDRUMgvFzpoe-f0onpUAs1j9HPDN01IHxg,627
74
+ gllm_inference/model/em/openai_em.pyi,sha256=KcWpMmxNqS28r4zT4H2TIADHr7e7f3VSI1MPzjJXH9k,442
75
+ gllm_inference/model/em/twelvelabs_em.pyi,sha256=pf9YfTfTPAceBoe1mA5VgtCroHZi5k42mEz-mGSD5QM,400
76
+ gllm_inference/model/em/voyage_em.pyi,sha256=CEfXjLNZamfhsLyAxIkDXND2Jk4GzwXK5puK9yKJDyE,531
77
+ gllm_inference/model/lm/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
78
+ gllm_inference/model/lm/anthropic_lm.pyi,sha256=dWfG-M_gD644yJ-LK_T8HnAT649j3Vx7TVof03XQimE,611
79
+ gllm_inference/model/lm/google_lm.pyi,sha256=cMV5zYX8uwUF7pErv4pXnXD2G52umo3sxKwbSx7nFhQ,511
80
+ gllm_inference/model/lm/openai_lm.pyi,sha256=u11zvvIS7-XaHKZ33cZxGQmT6cZ4DqK9Do8l7gFOUTc,618
81
+ gllm_inference/model/lm/xai_lm.pyi,sha256=2ZEQ_--e_zsb23zZQ8bKdQShU7zChx5TrDKF8EpwEpU,506
82
+ gllm_inference/output_parser/__init__.pyi,sha256=WQOOgsYnPk8vd-SOhFMMaVTzy4gkYrOAyT5gnAxv0A0,129
83
+ gllm_inference/output_parser/json_output_parser.pyi,sha256=uulh91uQLMSb4ZXZhHYi9W9w7zGnmrOweEkL6wdDJN8,2933
84
+ gllm_inference/output_parser/output_parser.pyi,sha256=Yzk7F26pH8Uc7FQZo4G6l67YkfppefUvaV9cNK-HyDs,948
85
+ gllm_inference/prompt_builder/__init__.pyi,sha256=kshfBMvwIwiIvjxiGG5BrJZNvpPa8rhtkbHo5FPifBg,117
86
+ gllm_inference/prompt_builder/prompt_builder.pyi,sha256=VQaw8nE8SPK1nvVVB4YjeqZSazigNPuWvOEcgGji0W8,4557
87
+ gllm_inference/prompt_builder/format_strategy/__init__.pyi,sha256=QhORHac3ySOPmL9k9kmCKL70vtaUtwkZEtGoRNWNuA8,308
88
+ gllm_inference/prompt_builder/format_strategy/format_strategy.pyi,sha256=JSUl_7Ka08oDZPpslymkUa8pDzqNGIK2TlcVANspqrY,2273
89
+ gllm_inference/prompt_builder/format_strategy/jinja_format_strategy.pyi,sha256=IAezLUiKSJMaoDyleo8pFnFqq8rBM_Q-lNXuAGvwXhI,2225
90
+ gllm_inference/prompt_builder/format_strategy/string_format_strategy.pyi,sha256=E0r8x6NTVbPLUfbJBz75PW1n3Ong6bi1WNgxBD0FTM4,693
91
+ gllm_inference/prompt_formatter/__init__.pyi,sha256=rTsjfRsT-y00qH67fPewMNPMN1fAO2y7DM9scR1ccm0,740
92
+ gllm_inference/prompt_formatter/agnostic_prompt_formatter.pyi,sha256=c9mN4t8LXn79h8wq8DAeWYwMgmZGzXjP7EcjLpwfNZg,2018
93
+ gllm_inference/prompt_formatter/huggingface_prompt_formatter.pyi,sha256=AJ-D11HBhTKBA1B5s0km_K1R6o5HD1yvdbGoL7SpGhI,2729
94
+ gllm_inference/prompt_formatter/llama_prompt_formatter.pyi,sha256=nz24trAqHsUY_zHVYfdNzhBO_TCUtzRkg-QYYow4l2s,2848
95
+ gllm_inference/prompt_formatter/mistral_prompt_formatter.pyi,sha256=DgFitxfvCSJf8_6NUfSre-PXYMuPI4vX4AN6BPF1cBE,2650
96
+ gllm_inference/prompt_formatter/openai_prompt_formatter.pyi,sha256=QB7gHddipQuAolUKO01ApZeaRxBtRZzglS5B78mzsV8,1286
97
+ gllm_inference/prompt_formatter/prompt_formatter.pyi,sha256=UkcPi5ao98OGJyNRsqfhYTlMW-ZLNITaGZUTwzvBzqk,1146
98
+ gllm_inference/realtime_chat/__init__.pyi,sha256=Fhxr_sYOt_p43BT2uvJmCp5FMhfb3dDMoIZglq2y6G8,137
99
+ gllm_inference/realtime_chat/google_realtime_chat.pyi,sha256=S5FOCacuc7j4J2tjpUHCdp91YlalxIJ3V3Bh0t0GGus,10112
100
+ gllm_inference/realtime_chat/realtime_chat.pyi,sha256=qG0fXyhTPLw6wGpnDX6wQrLfs6RBaxCK2ASwIqOymp4,1375
101
+ gllm_inference/realtime_chat/input_streamer/__init__.pyi,sha256=REN1dEEZB-I2RX5QsBMxWcUiGm0-fWLevgj7IT18W6Y,317
102
+ gllm_inference/realtime_chat/input_streamer/input_streamer.pyi,sha256=DOVBe_lxievONS_t14opNNuv0zM3MWyBcCwmxzR4_BE,1259
103
+ gllm_inference/realtime_chat/input_streamer/keyboard_input_streamer.pyi,sha256=3L0y91efGmb4YFWigs2WGAURQ48EwNxLgcVMm15yrIg,1043
104
+ gllm_inference/realtime_chat/input_streamer/linux_mic_input_streamer.pyi,sha256=o8rOootmcKHmKRUCCHvfGlCcfeBow5UTbQHIw89XIqg,1337
105
+ gllm_inference/realtime_chat/output_streamer/__init__.pyi,sha256=FP5amnApQcIvW7YfQEnLzJBzEDz51npgocfCcRWDQXA,339
106
+ gllm_inference/realtime_chat/output_streamer/console_output_streamer.pyi,sha256=pO_llZK7JDjllTWEZjujEVo83Hvrv_8LwVqEkQiIOOo,674
107
+ gllm_inference/realtime_chat/output_streamer/linux_speaker_output_streamer.pyi,sha256=yie6VaOyozzE4bKtvWr-6hRiobK-8mshCqeUOhK1PVA,1487
108
+ gllm_inference/realtime_chat/output_streamer/output_streamer.pyi,sha256=GPAw1wPSrwHpJ2MeZJB3V9CCESTDZpyYYgsoy_LI2Nk,1083
109
+ gllm_inference/request_processor/__init__.pyi,sha256=hVnfdNZnkTBJHnmLtN3Na4ANP0yK6AstWdIizVr2Apo,227
110
+ gllm_inference/request_processor/lm_request_processor.pyi,sha256=VnYc8E3Iayyhw-rPnGPfTKuO3ohgFsS8HPrZJeyES5I,5889
111
+ gllm_inference/request_processor/uses_lm_mixin.pyi,sha256=Yu0XPNuHxq1tWBviHTPw1oThojneFwGHepvGjBXxKQA,6382
112
+ gllm_inference/schema/__init__.pyi,sha256=hgRrwTocQ8b5MDDosSQN8zEuarGckpVply1OwEbrd28,2404
113
+ gllm_inference/schema/activity.pyi,sha256=JnO2hqj91P5Tc6qb4pbkEMrHer2u5owiCvhl-igcQKQ,2303
114
+ gllm_inference/schema/attachment.pyi,sha256=EHV3uPB3lVZCMNxySVuIoB6ikAiY7THSNh77q3ZtkVY,3246
115
+ gllm_inference/schema/code_exec_result.pyi,sha256=ZTHh6JtRrPIdQ059P1UAiD2L-tAO1_S5YcMsAXfJ5A0,559
116
+ gllm_inference/schema/config.pyi,sha256=rAL_UeXyQeXVk1P2kqd8vFWOMwmKenfpQLtvMP74t9s,674
117
+ gllm_inference/schema/enums.pyi,sha256=jByrR0Y84-WZ3KDPUjuOyfecouUATyO-A8rdehKPjgs,2065
118
+ gllm_inference/schema/events.pyi,sha256=XQEy5SqYoutq2DguwCCTGi5DHrlnDai6nJElAYsj1gk,4638
119
+ gllm_inference/schema/lm_input.pyi,sha256=A5pjz1id6tP9XRNhzQrbmzd66C_q3gzo0UP8rCemz6Q,193
120
+ gllm_inference/schema/lm_output.pyi,sha256=hQNfXbGBZAgLxHf_gOD_aifn0_9CbCzYS2HiMECqQvc,7409
121
+ gllm_inference/schema/mcp.pyi,sha256=Vwu8E2BDl6FvvnI42gIyY3Oki1BdwRE3Uh3aV0rmhQU,1014
122
+ gllm_inference/schema/message.pyi,sha256=VP9YppKj2mo1esl9cy6qQO9m2mMHUjTmfGDdyUor880,2220
123
+ gllm_inference/schema/model_id.pyi,sha256=BHzi8PdYsV_rPGkBcKvZJDCf62ToPjmRTcXcTgKfoTg,5894
124
+ gllm_inference/schema/reasoning.pyi,sha256=SlTuiDw87GdnAn-I6YOPIJRhEBiwQljM46JohG05guQ,562
125
+ gllm_inference/schema/token_usage.pyi,sha256=1GTQVORV0dBNmD_jix8aVaUqxMKFF04KpLP7y2urqbk,2950
126
+ gllm_inference/schema/tool_call.pyi,sha256=zQaVxCnkVxOfOEhBidqohU85gb4PRwnwBiygKaunamk,389
127
+ gllm_inference/schema/tool_result.pyi,sha256=cAG7TVtB4IWJPt8XBBbB92cuY1ZsX9M276bN9aqjcvM,276
128
+ gllm_inference/schema/type_alias.pyi,sha256=q_C1x8hUUoerGEQFj99g3yaZ1ILigGQSEo-FheqS_lM,529
129
+ gllm_inference/utils/__init__.pyi,sha256=mDJ2gLSeQzm-79Tov5-dhrMNaDqgcD1VVzDYAWvIRqA,391
130
+ gllm_inference/utils/io_utils.pyi,sha256=7kUTacHAVRYoemFUOjCH7-Qmw-YsQGd6rGYxjf_qmtw,1094
131
+ gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
132
+ gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
133
+ gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
134
+ gllm_inference_binary-0.5.55.dist-info/METADATA,sha256=x6uemrRDozu49LQH2KLiFERGFI1w19V2WGoTLi1CqVk,5807
135
+ gllm_inference_binary-0.5.55.dist-info/WHEEL,sha256=t5_2FPbm0tV7_MDNYR2NrQvum0LVPlNiR9hm8r8eRow,104
136
+ gllm_inference_binary-0.5.55.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
137
+ gllm_inference_binary-0.5.55.dist-info/RECORD,,
@@ -0,0 +1,5 @@
1
+ Wheel-Version: 1.0
2
+ Generator: Nuitka (2.6.9)
3
+ Root-Is-Purelib: false
4
+ Tag: cp313-cp313-macosx_13_0_arm64
5
+
@@ -0,0 +1 @@
1
+ gllm_inference