gllm-inference-binary 0.5.65__cp313-cp313-macosx_13_0_arm64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (140) hide show
  1. gllm_inference/__init__.pyi +0 -0
  2. gllm_inference/builder/__init__.pyi +6 -0
  3. gllm_inference/builder/_build_invoker.pyi +28 -0
  4. gllm_inference/builder/build_em_invoker.pyi +130 -0
  5. gllm_inference/builder/build_lm_invoker.pyi +222 -0
  6. gllm_inference/builder/build_lm_request_processor.pyi +88 -0
  7. gllm_inference/builder/build_output_parser.pyi +29 -0
  8. gllm_inference/catalog/__init__.pyi +4 -0
  9. gllm_inference/catalog/catalog.pyi +121 -0
  10. gllm_inference/catalog/lm_request_processor_catalog.pyi +112 -0
  11. gllm_inference/catalog/prompt_builder_catalog.pyi +82 -0
  12. gllm_inference/constants.pyi +12 -0
  13. gllm_inference/em_invoker/__init__.pyi +12 -0
  14. gllm_inference/em_invoker/azure_openai_em_invoker.pyi +88 -0
  15. gllm_inference/em_invoker/bedrock_em_invoker.pyi +118 -0
  16. gllm_inference/em_invoker/cohere_em_invoker.pyi +127 -0
  17. gllm_inference/em_invoker/em_invoker.pyi +90 -0
  18. gllm_inference/em_invoker/google_em_invoker.pyi +129 -0
  19. gllm_inference/em_invoker/jina_em_invoker.pyi +103 -0
  20. gllm_inference/em_invoker/langchain/__init__.pyi +3 -0
  21. gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi +84 -0
  22. gllm_inference/em_invoker/langchain_em_invoker.pyi +46 -0
  23. gllm_inference/em_invoker/openai_compatible_em_invoker.pyi +41 -0
  24. gllm_inference/em_invoker/openai_em_invoker.pyi +118 -0
  25. gllm_inference/em_invoker/schema/__init__.pyi +0 -0
  26. gllm_inference/em_invoker/schema/bedrock.pyi +29 -0
  27. gllm_inference/em_invoker/schema/cohere.pyi +20 -0
  28. gllm_inference/em_invoker/schema/google.pyi +9 -0
  29. gllm_inference/em_invoker/schema/jina.pyi +29 -0
  30. gllm_inference/em_invoker/schema/langchain.pyi +5 -0
  31. gllm_inference/em_invoker/schema/openai.pyi +7 -0
  32. gllm_inference/em_invoker/schema/openai_compatible.pyi +7 -0
  33. gllm_inference/em_invoker/schema/twelvelabs.pyi +17 -0
  34. gllm_inference/em_invoker/schema/voyage.pyi +15 -0
  35. gllm_inference/em_invoker/twelevelabs_em_invoker.pyi +101 -0
  36. gllm_inference/em_invoker/voyage_em_invoker.pyi +104 -0
  37. gllm_inference/exceptions/__init__.pyi +4 -0
  38. gllm_inference/exceptions/error_parser.pyi +41 -0
  39. gllm_inference/exceptions/exceptions.pyi +132 -0
  40. gllm_inference/exceptions/provider_error_map.pyi +24 -0
  41. gllm_inference/lm_invoker/__init__.pyi +15 -0
  42. gllm_inference/lm_invoker/anthropic_lm_invoker.pyi +320 -0
  43. gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi +237 -0
  44. gllm_inference/lm_invoker/batch/__init__.pyi +3 -0
  45. gllm_inference/lm_invoker/batch/batch_operations.pyi +128 -0
  46. gllm_inference/lm_invoker/bedrock_lm_invoker.pyi +212 -0
  47. gllm_inference/lm_invoker/datasaur_lm_invoker.pyi +157 -0
  48. gllm_inference/lm_invoker/google_lm_invoker.pyi +421 -0
  49. gllm_inference/lm_invoker/langchain_lm_invoker.pyi +239 -0
  50. gllm_inference/lm_invoker/litellm_lm_invoker.pyi +224 -0
  51. gllm_inference/lm_invoker/lm_invoker.pyi +183 -0
  52. gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi +252 -0
  53. gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi +52 -0
  54. gllm_inference/lm_invoker/openai_lm_invoker.pyi +437 -0
  55. gllm_inference/lm_invoker/portkey_lm_invoker.pyi +296 -0
  56. gllm_inference/lm_invoker/schema/__init__.pyi +0 -0
  57. gllm_inference/lm_invoker/schema/anthropic.pyi +56 -0
  58. gllm_inference/lm_invoker/schema/bedrock.pyi +53 -0
  59. gllm_inference/lm_invoker/schema/datasaur.pyi +14 -0
  60. gllm_inference/lm_invoker/schema/google.pyi +36 -0
  61. gllm_inference/lm_invoker/schema/langchain.pyi +23 -0
  62. gllm_inference/lm_invoker/schema/openai.pyi +109 -0
  63. gllm_inference/lm_invoker/schema/openai_chat_completions.pyi +62 -0
  64. gllm_inference/lm_invoker/schema/portkey.pyi +31 -0
  65. gllm_inference/lm_invoker/schema/xai.pyi +31 -0
  66. gllm_inference/lm_invoker/sea_lion_lm_invoker.pyi +48 -0
  67. gllm_inference/lm_invoker/xai_lm_invoker.pyi +252 -0
  68. gllm_inference/model/__init__.pyi +13 -0
  69. gllm_inference/model/em/__init__.pyi +0 -0
  70. gllm_inference/model/em/cohere_em.pyi +17 -0
  71. gllm_inference/model/em/google_em.pyi +16 -0
  72. gllm_inference/model/em/jina_em.pyi +22 -0
  73. gllm_inference/model/em/openai_em.pyi +15 -0
  74. gllm_inference/model/em/twelvelabs_em.pyi +13 -0
  75. gllm_inference/model/em/voyage_em.pyi +20 -0
  76. gllm_inference/model/lm/__init__.pyi +0 -0
  77. gllm_inference/model/lm/anthropic_lm.pyi +22 -0
  78. gllm_inference/model/lm/google_lm.pyi +18 -0
  79. gllm_inference/model/lm/openai_lm.pyi +27 -0
  80. gllm_inference/model/lm/sea_lion_lm.pyi +16 -0
  81. gllm_inference/model/lm/xai_lm.pyi +19 -0
  82. gllm_inference/output_parser/__init__.pyi +3 -0
  83. gllm_inference/output_parser/json_output_parser.pyi +60 -0
  84. gllm_inference/output_parser/output_parser.pyi +27 -0
  85. gllm_inference/prompt_builder/__init__.pyi +3 -0
  86. gllm_inference/prompt_builder/format_strategy/__init__.pyi +4 -0
  87. gllm_inference/prompt_builder/format_strategy/format_strategy.pyi +55 -0
  88. gllm_inference/prompt_builder/format_strategy/jinja_format_strategy.pyi +45 -0
  89. gllm_inference/prompt_builder/format_strategy/string_format_strategy.pyi +20 -0
  90. gllm_inference/prompt_builder/prompt_builder.pyi +73 -0
  91. gllm_inference/prompt_formatter/__init__.pyi +7 -0
  92. gllm_inference/prompt_formatter/agnostic_prompt_formatter.pyi +49 -0
  93. gllm_inference/prompt_formatter/huggingface_prompt_formatter.pyi +55 -0
  94. gllm_inference/prompt_formatter/llama_prompt_formatter.pyi +59 -0
  95. gllm_inference/prompt_formatter/mistral_prompt_formatter.pyi +53 -0
  96. gllm_inference/prompt_formatter/openai_prompt_formatter.pyi +35 -0
  97. gllm_inference/prompt_formatter/prompt_formatter.pyi +30 -0
  98. gllm_inference/realtime_chat/__init__.pyi +3 -0
  99. gllm_inference/realtime_chat/google_realtime_chat.pyi +205 -0
  100. gllm_inference/realtime_chat/input_streamer/__init__.pyi +4 -0
  101. gllm_inference/realtime_chat/input_streamer/input_streamer.pyi +36 -0
  102. gllm_inference/realtime_chat/input_streamer/keyboard_input_streamer.pyi +27 -0
  103. gllm_inference/realtime_chat/input_streamer/linux_mic_input_streamer.pyi +36 -0
  104. gllm_inference/realtime_chat/output_streamer/__init__.pyi +4 -0
  105. gllm_inference/realtime_chat/output_streamer/console_output_streamer.pyi +21 -0
  106. gllm_inference/realtime_chat/output_streamer/linux_speaker_output_streamer.pyi +42 -0
  107. gllm_inference/realtime_chat/output_streamer/output_streamer.pyi +33 -0
  108. gllm_inference/realtime_chat/realtime_chat.pyi +28 -0
  109. gllm_inference/request_processor/__init__.pyi +4 -0
  110. gllm_inference/request_processor/lm_request_processor.pyi +101 -0
  111. gllm_inference/request_processor/uses_lm_mixin.pyi +130 -0
  112. gllm_inference/schema/__init__.pyi +19 -0
  113. gllm_inference/schema/activity.pyi +64 -0
  114. gllm_inference/schema/attachment.pyi +102 -0
  115. gllm_inference/schema/code_exec_result.pyi +14 -0
  116. gllm_inference/schema/config.pyi +15 -0
  117. gllm_inference/schema/enums.pyi +82 -0
  118. gllm_inference/schema/events.pyi +105 -0
  119. gllm_inference/schema/formatter.pyi +31 -0
  120. gllm_inference/schema/lm_input.pyi +4 -0
  121. gllm_inference/schema/lm_output.pyi +266 -0
  122. gllm_inference/schema/mcp.pyi +31 -0
  123. gllm_inference/schema/message.pyi +52 -0
  124. gllm_inference/schema/model_id.pyi +176 -0
  125. gllm_inference/schema/reasoning.pyi +15 -0
  126. gllm_inference/schema/token_usage.pyi +75 -0
  127. gllm_inference/schema/tool_call.pyi +14 -0
  128. gllm_inference/schema/tool_result.pyi +11 -0
  129. gllm_inference/schema/type_alias.pyi +11 -0
  130. gllm_inference/utils/__init__.pyi +5 -0
  131. gllm_inference/utils/io_utils.pyi +26 -0
  132. gllm_inference/utils/langchain.pyi +30 -0
  133. gllm_inference/utils/validation.pyi +15 -0
  134. gllm_inference.build/.gitignore +1 -0
  135. gllm_inference.cpython-313-darwin.so +0 -0
  136. gllm_inference.pyi +156 -0
  137. gllm_inference_binary-0.5.65.dist-info/METADATA +138 -0
  138. gllm_inference_binary-0.5.65.dist-info/RECORD +140 -0
  139. gllm_inference_binary-0.5.65.dist-info/WHEEL +5 -0
  140. gllm_inference_binary-0.5.65.dist-info/top_level.txt +1 -0
@@ -0,0 +1,26 @@
1
+ from _typeshed import Incomplete
2
+
3
+ logger: Incomplete
4
+ DEFAULT_BASE64_ALLOWED_MIMETYPES: Incomplete
5
+
6
+ def base64_to_bytes(value: str, *, allowed_mimetypes: tuple[str, ...] | None = ...) -> str | bytes:
7
+ '''Decode a base64 string to bytes based on allowed MIME type.
8
+
9
+ The conversion steps are as follows:
10
+ 1. The function first attempts to decode the given string from base64.
11
+ 2. If decoding succeeds, it checks the MIME type of the decoded content.
12
+ 3. When the MIME type matches one of the allowed patterns (e.g., ``"image/*"``),
13
+ the raw bytes are returned. Otherwise, the original string is returned unchanged.
14
+
15
+ Args:
16
+ value (str): Input data to decode.
17
+ allowed_mimetypes (tuple[str, ...], optional): MIME type prefixes that are allowed
18
+ to be decoded into bytes. Defaults to ("image/*", "audio/*", "video/*").
19
+
20
+ Returns:
21
+ str | bytes: Base64-encoded string or raw bytes if MIME type is allowed;
22
+ otherwise returns original string.
23
+
24
+ Raises:
25
+ ValueError: If the input is not a string.
26
+ '''
@@ -0,0 +1,30 @@
1
+ from _typeshed import Incomplete
2
+ from langchain_core.embeddings import Embeddings as Embeddings
3
+ from langchain_core.language_models import BaseChatModel as BaseChatModel
4
+ from typing import Any
5
+
6
+ MODEL_NAME_KEYS: Incomplete
7
+
8
+ def load_langchain_model(model_class_path: str, model_name: str, model_kwargs: dict[str, Any]) -> BaseChatModel | Embeddings:
9
+ '''Loads the LangChain\'s model instance.
10
+
11
+ Args:
12
+ model_class_path (str): The path to the LangChain\'s class, e.g. "langchain_openai.ChatOpenAI".
13
+ model_name (str): The model name.
14
+ model_kwargs (dict[str, Any]): The additional keyword arguments.
15
+
16
+ Returns:
17
+ BaseChatModel | Embeddings: The LangChain\'s model instance.
18
+ '''
19
+ def parse_model_data(model: BaseChatModel | Embeddings) -> dict[str, str]:
20
+ """Parses the model data from LangChain's BaseChatModel or Embeddings instance.
21
+
22
+ Args:
23
+ model (BaseChatModel | Embeddings): The LangChain's BaseChatModel or Embeddings instance.
24
+
25
+ Returns:
26
+ dict[str, str]: The dictionary containing the model name and path.
27
+
28
+ Raises:
29
+ ValueError: If the model name is not found in the model data.
30
+ """
@@ -0,0 +1,15 @@
1
+ from _typeshed import Incomplete
2
+ from enum import StrEnum
3
+
4
+ logger: Incomplete
5
+
6
+ def validate_string_enum(enum_type: type[StrEnum], value: str) -> None:
7
+ """Validates that the provided value is a valid string enum value.
8
+
9
+ Args:
10
+ enum_type (type[StrEnum]): The type of the string enum.
11
+ value (str): The value to validate.
12
+
13
+ Raises:
14
+ ValueError: If the provided value is not a valid string enum value.
15
+ """
@@ -0,0 +1 @@
1
+ *
Binary file
gllm_inference.pyi ADDED
@@ -0,0 +1,156 @@
1
+ # This file was generated by Nuitka
2
+
3
+ # Stubs included by default
4
+
5
+
6
+ __name__ = ...
7
+
8
+
9
+
10
+ # Modules used internally, to allow implicit dependencies to be seen:
11
+ import os
12
+ import typing
13
+ import gllm_core
14
+ import gllm_core.utils
15
+ import gllm_inference.em_invoker.AzureOpenAIEMInvoker
16
+ import gllm_inference.em_invoker.BedrockEMInvoker
17
+ import gllm_inference.em_invoker.CohereEMInvoker
18
+ import gllm_inference.em_invoker.GoogleEMInvoker
19
+ import gllm_inference.em_invoker.JinaEMInvoker
20
+ import gllm_inference.em_invoker.LangChainEMInvoker
21
+ import gllm_inference.em_invoker.OpenAICompatibleEMInvoker
22
+ import gllm_inference.em_invoker.OpenAIEMInvoker
23
+ import gllm_inference.em_invoker.TwelveLabsEMInvoker
24
+ import gllm_inference.em_invoker.VoyageEMInvoker
25
+ import gllm_inference.lm_invoker.AnthropicLMInvoker
26
+ import gllm_inference.lm_invoker.AzureOpenAILMInvoker
27
+ import gllm_inference.lm_invoker.BedrockLMInvoker
28
+ import gllm_inference.lm_invoker.DatasaurLMInvoker
29
+ import gllm_inference.lm_invoker.GoogleLMInvoker
30
+ import gllm_inference.lm_invoker.LangChainLMInvoker
31
+ import gllm_inference.lm_invoker.LiteLLMLMInvoker
32
+ import gllm_inference.lm_invoker.OpenAIChatCompletionsLMInvoker
33
+ import gllm_inference.lm_invoker.OpenAICompatibleLMInvoker
34
+ import gllm_inference.lm_invoker.OpenAILMInvoker
35
+ import gllm_inference.lm_invoker.PortkeyLMInvoker
36
+ import gllm_inference.lm_invoker.SeaLionLMInvoker
37
+ import gllm_inference.lm_invoker.XAILMInvoker
38
+ import gllm_inference.prompt_builder.PromptBuilder
39
+ import gllm_inference.output_parser.JSONOutputParser
40
+ import json
41
+ import abc
42
+ import pandas
43
+ import pydantic
44
+ import re
45
+ import gllm_core.utils.retry
46
+ import gllm_inference.request_processor.LMRequestProcessor
47
+ import gllm_core.utils.imports
48
+ import gllm_inference.schema.ModelId
49
+ import gllm_inference.schema.ModelProvider
50
+ import gllm_inference.schema.TruncationConfig
51
+ import asyncio
52
+ import base64
53
+ import enum
54
+ import gllm_inference.exceptions.BaseInvokerError
55
+ import gllm_inference.exceptions.convert_http_status_to_base_invoker_error
56
+ import gllm_inference.schema.Attachment
57
+ import gllm_inference.schema.AttachmentType
58
+ import gllm_inference.schema.EMContent
59
+ import gllm_inference.schema.Vector
60
+ import aioboto3
61
+ import cohere
62
+ import asyncio.CancelledError
63
+ import gllm_inference.exceptions.convert_to_base_invoker_error
64
+ import gllm_inference.schema.TruncateSide
65
+ import google
66
+ import google.auth
67
+ import google.genai
68
+ import google.genai.types
69
+ import httpx
70
+ import gllm_inference.exceptions.ProviderInternalError
71
+ import gllm_core.utils.concurrency
72
+ import langchain_core
73
+ import langchain_core.embeddings
74
+ import gllm_inference.exceptions.InvokerRuntimeError
75
+ import gllm_inference.exceptions.build_debug_info
76
+ import gllm_inference.utils.load_langchain_model
77
+ import gllm_inference.utils.parse_model_data
78
+ import openai
79
+ import io
80
+ import twelvelabs
81
+ import sys
82
+ import voyageai
83
+ import voyageai.client_async
84
+ import http
85
+ import http.HTTPStatus
86
+ import __future__
87
+ import uuid
88
+ import gllm_core.constants
89
+ import gllm_core.event
90
+ import gllm_core.schema
91
+ import gllm_core.schema.tool
92
+ import langchain_core.tools
93
+ import gllm_inference.schema.BatchStatus
94
+ import gllm_inference.schema.LMInput
95
+ import gllm_inference.schema.LMOutput
96
+ import gllm_inference.schema.Message
97
+ import gllm_inference.schema.Reasoning
98
+ import gllm_inference.schema.ResponseSchema
99
+ import gllm_inference.schema.ThinkingEvent
100
+ import gllm_inference.schema.TokenUsage
101
+ import gllm_inference.schema.ToolCall
102
+ import gllm_inference.schema.ToolResult
103
+ import anthropic
104
+ import anthropic.types
105
+ import anthropic.types.message_create_params
106
+ import anthropic.types.messages
107
+ import anthropic.types.messages.batch_create_params
108
+ import gllm_inference.schema.MessageRole
109
+ import langchain_core.language_models
110
+ import langchain_core.messages
111
+ import gllm_inference.exceptions._get_exception_key
112
+ import litellm
113
+ import inspect
114
+ import time
115
+ import jsonschema
116
+ import gllm_inference.lm_invoker.batch.BatchOperations
117
+ import gllm_inference.schema.MessageContent
118
+ import gllm_inference.schema.ActivityEvent
119
+ import gllm_inference.schema.CodeEvent
120
+ import gllm_inference.schema.CodeExecResult
121
+ import gllm_inference.schema.MCPCall
122
+ import gllm_inference.schema.MCPCallActivity
123
+ import gllm_inference.schema.MCPListToolsActivity
124
+ import gllm_inference.schema.MCPServer
125
+ import gllm_inference.schema.WebSearchActivity
126
+ import logging
127
+ import portkey_ai
128
+ import xai_sdk
129
+ import xai_sdk.chat
130
+ import xai_sdk.search
131
+ import xai_sdk.proto
132
+ import xai_sdk.proto.v5
133
+ import xai_sdk.proto.v5.chat_pb2
134
+ import jinja2
135
+ import jinja2.sandbox
136
+ import gllm_inference.schema.JinjaEnvType
137
+ import gllm_inference.prompt_builder.format_strategy.JinjaFormatStrategy
138
+ import gllm_inference.prompt_builder.format_strategy.StringFormatStrategy
139
+ import gllm_inference.schema.HistoryFormatter
140
+ import transformers
141
+ import gllm_inference.prompt_formatter.HuggingFacePromptFormatter
142
+ import traceback
143
+ import gllm_inference.realtime_chat.input_streamer.KeyboardInputStreamer
144
+ import gllm_inference.realtime_chat.output_streamer.ConsoleOutputStreamer
145
+ import google.genai.live
146
+ import gllm_core.utils.logger_manager
147
+ import mimetypes
148
+ import pathlib
149
+ import pathlib.Path
150
+ import filetype
151
+ import magic
152
+ import requests
153
+ import gllm_core.schema.chunk
154
+ import binascii
155
+ import fnmatch
156
+ import importlib
@@ -0,0 +1,138 @@
1
+ Metadata-Version: 2.2
2
+ Name: gllm-inference-binary
3
+ Version: 0.5.65
4
+ Summary: A library containing components related to model inferences in Gen AI applications.
5
+ Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, "Delfia N. A. Putri" <delfia.n.a.putri@gdplabs.id>
6
+ Requires-Python: <3.14,>=3.11
7
+ Description-Content-Type: text/markdown
8
+ Requires-Dist: poetry<3.0.0,>=2.1.3
9
+ Requires-Dist: gllm-core-binary<0.4.0,>=0.3.23
10
+ Requires-Dist: aiohttp<4.0.0,>=3.12.14
11
+ Requires-Dist: filetype<2.0.0,>=1.2.0
12
+ Requires-Dist: httpx<0.29.0,>=0.28.0
13
+ Requires-Dist: jinja2<4.0.0,>=3.1.4
14
+ Requires-Dist: jsonschema<5.0.0,>=4.24.0
15
+ Requires-Dist: langchain<0.4.0,>=0.3.0
16
+ Requires-Dist: pandas<3.0.0,>=2.2.3
17
+ Requires-Dist: protobuf<7.0.0,>=5.29.4
18
+ Requires-Dist: python-magic<0.5.0,>=0.4.27; sys_platform != "win32"
19
+ Requires-Dist: python-magic-bin<0.5.0,>=0.4.14; sys_platform == "win32"
20
+ Requires-Dist: sentencepiece<0.3.0,>=0.2.0
21
+ Provides-Extra: dev
22
+ Requires-Dist: coverage<8.0.0,>=7.4.4; extra == "dev"
23
+ Requires-Dist: mypy<2.0.0,>=1.15.0; extra == "dev"
24
+ Requires-Dist: pre-commit<4.0.0,>=3.7.0; extra == "dev"
25
+ Requires-Dist: pytest<9.0.0,>=8.1.1; extra == "dev"
26
+ Requires-Dist: pytest-asyncio<0.24.0,>=0.23.6; extra == "dev"
27
+ Requires-Dist: pytest-cov<6.0.0,>=5.0.0; extra == "dev"
28
+ Requires-Dist: ruff<0.7.0,>=0.6.7; extra == "dev"
29
+ Provides-Extra: anthropic
30
+ Requires-Dist: anthropic<0.61.0,>=0.60.0; extra == "anthropic"
31
+ Provides-Extra: bedrock
32
+ Requires-Dist: aioboto3<16.0.0,>=15.0.0; extra == "bedrock"
33
+ Provides-Extra: cohere
34
+ Requires-Dist: cohere<6.0.0,>=5.18.0; extra == "cohere"
35
+ Provides-Extra: datasaur
36
+ Requires-Dist: openai<3.0.0,>=2.7.0; extra == "datasaur"
37
+ Provides-Extra: google
38
+ Requires-Dist: google-genai<=1.36,>=1.23; extra == "google"
39
+ Provides-Extra: huggingface
40
+ Requires-Dist: huggingface-hub<0.31.0,>=0.30.0; extra == "huggingface"
41
+ Requires-Dist: transformers<5.0.0,>=4.52.0; extra == "huggingface"
42
+ Provides-Extra: litellm
43
+ Requires-Dist: litellm<2.0.0,>=1.69.2; extra == "litellm"
44
+ Provides-Extra: openai
45
+ Requires-Dist: openai<3.0.0,>=2.7.0; extra == "openai"
46
+ Provides-Extra: portkey-ai
47
+ Requires-Dist: portkey-ai<2.0.0,>=1.14.4; extra == "portkey-ai"
48
+ Provides-Extra: twelvelabs
49
+ Requires-Dist: twelvelabs<0.5.0,>=0.4.4; extra == "twelvelabs"
50
+ Provides-Extra: voyage
51
+ Requires-Dist: voyageai<0.4.0,>=0.3.0; python_version < "3.13" and extra == "voyage"
52
+ Provides-Extra: xai
53
+ Requires-Dist: xai_sdk<2.0.0,>=1.0.0; extra == "xai"
54
+
55
+ # GLLM Inference
56
+
57
+ ## Description
58
+
59
+ A library containing components related to model inferences in Gen AI applications.
60
+
61
+ ## Installation
62
+
63
+ ### Prerequisites
64
+ 1. Python 3.11+ - [Install here](https://www.python.org/downloads/)
65
+ 2. Pip (if using Pip) - [Install here](https://pip.pypa.io/en/stable/installation/)
66
+ 3. Poetry (automatically installed via Makefile) - [Install here](https://python-poetry.org/docs/#installation)
67
+ 4. Git (if using Git) - [Install here](https://git-scm.com/downloads)
68
+ 5. gcloud CLI (for authentication) - [Install here](https://cloud.google.com/sdk/docs/install)
69
+ 6. For git installation, access to the [GDP Labs SDK github repository](https://github.com/GDP-ADMIN/gl-sdk)
70
+
71
+ ### 1. Installation from Artifact Registry
72
+ Choose one of the following methods to install the package:
73
+
74
+ #### Using pip
75
+ ```bash
76
+ pip install gllm-inference-binary
77
+ ```
78
+
79
+ #### Using Poetry
80
+ ```bash
81
+ poetry add gllm-inference-binary
82
+ ```
83
+
84
+ ### 2. Development Installation (Git)
85
+ For development purposes, you can install directly from the Git repository:
86
+ ```bash
87
+ poetry add "git+ssh://git@github.com/GDP-ADMIN/gen-ai-internal.git#subdirectory=libs/gllm-inference"
88
+ ```
89
+
90
+ Available extras:
91
+ - `anthropic`: Install Anthropic models dependencies
92
+ - `google-genai`: Install Google Generative AI models dependencies
93
+ - `google-vertexai`: Install Google Vertex AI models dependencies
94
+ - `huggingface`: Install HuggingFace models dependencies
95
+ - `openai`: Install OpenAI models dependencies
96
+ - `twelvelabs`: Install TwelveLabs models dependencies
97
+
98
+ ## Local Development Setup
99
+
100
+ ### Quick Setup (Recommended)
101
+ For local development with editable gllm packages, use the provided Makefile:
102
+
103
+ ```bash
104
+ # Complete setup: installs Poetry, configures auth, installs packages, sets up pre-commit
105
+ make setup
106
+ ```
107
+
108
+ The following are the available Makefile targets:
109
+
110
+ 1. `make setup` - Complete development setup (recommended for new developers)
111
+ 2. `make install-poetry` - Install or upgrade Poetry to the latest version
112
+ 3. `make auth` - Configure authentication for internal repositories
113
+ 4. `make install` - Install all dependencies
114
+ 5. `make install-pre-commit` - Set up pre-commit hooks
115
+ 6. `make update` - Update dependencies
116
+ ### Manual Development Setup (Legacy)
117
+ If you prefer to manage dependencies manually:
118
+
119
+ 1. Go to root folder of `gllm-inference` module, e.g. `cd libs/gllm-inference`.
120
+ 2. Run `poetry shell` to create a virtual environment.
121
+ 3. Run `poetry lock` to create a lock file if you haven't done it yet.
122
+ 4. Run `poetry install` to install the `gllm-inference` requirements for the first time.
123
+ 5. Run `poetry update` if you update any dependency module version at `pyproject.toml`.
124
+
125
+
126
+ ## Contributing
127
+ Please refer to this [Python Style Guide](https://docs.google.com/document/d/1uRggCrHnVfDPBnG641FyQBwUwLoFw0kTzNqRm92vUwM/edit?usp=sharing)
128
+ to get information about code style, documentation standard, and SCA that you need to use when contributing to this project
129
+
130
+ ### Getting Started with Development
131
+ 1. Clone the repository and navigate to the gllm-inference directory
132
+ 2. Run `make setup` to set up your development environment
133
+ 3. Run `which python` to get the path to be referenced at Visual Studio Code interpreter path (`Ctrl`+`Shift`+`P` or `Cmd`+`Shift`+`P`)
134
+ 4. Try running the unit test to see if it's working:
135
+ ```bash
136
+ poetry run pytest -s tests/unit_tests/
137
+ ```
138
+ 5. When you want to update the dependencies, run `make update`
@@ -0,0 +1,140 @@
1
+ gllm_inference.cpython-313-darwin.so,sha256=4S0I8bDh8uKRGWvxzGVAKfYrVQQcQtFy-Ibh9VDvPiE,5241544
2
+ gllm_inference.pyi,sha256=3UoxsyZnC_6PhNzIKIp37rL6CNEw-Ve0RJHiG9jLOBY,5193
3
+ gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
4
+ gllm_inference/constants.pyi,sha256=tBFhwE1at2gXMJ1bBM32eVIRgCJlB1uzg7ItXGx3RQE,316
5
+ gllm_inference/builder/__init__.pyi,sha256=usz2lvfwO4Yk-ZGKXbCWG1cEr3nlQXxMNDNC-2yc1NM,500
6
+ gllm_inference/builder/_build_invoker.pyi,sha256=v__-YT0jGmLqEsgl7Abk9we-wYWcyVFvlRN_Uu4vVak,848
7
+ gllm_inference/builder/build_em_invoker.pyi,sha256=0IVcRGaciPBAFcgvGRC4-Kje_PPXOSug77LqxRc-x_U,5749
8
+ gllm_inference/builder/build_lm_invoker.pyi,sha256=o4dIL90wDu94y_fIeUs2ei5Qkk1zZo5_nYBqL2EiHAE,9195
9
+ gllm_inference/builder/build_lm_request_processor.pyi,sha256=KbQkcPa8C-yzyelht4mWLP8kDmh17itAT3tn8ZJB6pg,4144
10
+ gllm_inference/builder/build_output_parser.pyi,sha256=_Lrq-bh1oPsb_Nwkkr_zyEUwIOMysRFZkvEtEM29LZM,936
11
+ gllm_inference/catalog/__init__.pyi,sha256=JBkPGTyiiZ30GECzJBW-mW8LekWyY2qyzal3eW7ynaM,287
12
+ gllm_inference/catalog/catalog.pyi,sha256=a4RNG1lKv51GxQpOqh47tz-PAROMPaeP2o5XNLBSZaU,4790
13
+ gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=WW1j8jWujnatF0c9rCk94CyobtFe3gOky2vjoGCV6nw,5424
14
+ gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=OU8k_4HbqjZEzHZlzSM3uzGQZJmM2uGD76Csqom0CEQ,3197
15
+ gllm_inference/em_invoker/__init__.pyi,sha256=FHFR4k8RqtdMNZkda6A89tvWFtIoxF-t_NahDGeBQ2k,1199
16
+ gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=fE-RZ10Leerbyqp8ISXeqdU_m0kaHdlHml_nQqwFjqY,4961
17
+ gllm_inference/em_invoker/bedrock_em_invoker.pyi,sha256=3GkVO9hvmYck4NJY5spAg6opSxr0v8OJ4qhMvz1rcAA,6365
18
+ gllm_inference/em_invoker/cohere_em_invoker.pyi,sha256=FqDCVs6uxKmQuetGkeXOahK6TNGcLpNj7sEnOM8XCFY,6581
19
+ gllm_inference/em_invoker/em_invoker.pyi,sha256=KGjLiAWGIA3ziV50zMwSzx6lTVUbYspQCl4LFxqbDlY,5101
20
+ gllm_inference/em_invoker/google_em_invoker.pyi,sha256=OgIQhT2g6g09Hqr9mc2--MjV3Ti-XuHfqLTunvcqU1k,6824
21
+ gllm_inference/em_invoker/jina_em_invoker.pyi,sha256=WZJ8BeaqxrAGMrUtIEUy3itcOEzvRQYEwe0gazIv6ng,5651
22
+ gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=BhZjkYZoxQXPJjP0PgA8W0xrtwiqDkW5E6NpGit0h1E,3498
23
+ gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=z4_jEuaprna6AJF2wXH-JgmynpVFXeb2HQZr-0ATmMw,2882
24
+ gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=tb6cXTvAGr01MMSc96mYWDqK-EK_VCdDOQ-Be0_EZfM,6195
25
+ gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=mitrJZW3M43k306cHwJoDjBfN1nr20C4sZ42Xr6hfO0,5347
26
+ gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=gyYQA3rbekHJIvvY6TE4tTt8W-DbPqjiPW6FaJoQPHI,5517
27
+ gllm_inference/em_invoker/langchain/__init__.pyi,sha256=VYGKE5OgU0my1RlhgzkU_A7-GLGnUDDnNFuctuRwILE,148
28
+ gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi,sha256=VU3-Vhb9BCDhJo8PPdWHe2rBEOCs_HMXT6ZaWwjUzZE,3304
29
+ gllm_inference/em_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
30
+ gllm_inference/em_invoker/schema/bedrock.pyi,sha256=ePYwailnZAYbZKt9HRDi4iAISq003jpxQPxDCiFdEZg,615
31
+ gllm_inference/em_invoker/schema/cohere.pyi,sha256=UTbTtePRR1zJMsM09SiTZSZZP0IaUGaODvc7ZqH9S8c,547
32
+ gllm_inference/em_invoker/schema/google.pyi,sha256=ovDlvinu99QJhIxMkvVUoGBEFkkEoAZhadSuk0nI9N8,181
33
+ gllm_inference/em_invoker/schema/jina.pyi,sha256=vE1ySd8OTDM35saEZos7UCdPwHeX66iuHkZ3RchSZKA,741
34
+ gllm_inference/em_invoker/schema/langchain.pyi,sha256=edcUvc1IHoSMFwqV83uqWqd0U3fLhkyWQjVknvjHI8U,112
35
+ gllm_inference/em_invoker/schema/openai.pyi,sha256=Q_dsEcodkOXYXPdrkOkW0LnuLhfeq8tEbtZAGMz2ajA,139
36
+ gllm_inference/em_invoker/schema/openai_compatible.pyi,sha256=gmvGtsWoOMBelke_tZjC6dKimFBW9f4Vrgv0Ig0OM9Q,150
37
+ gllm_inference/em_invoker/schema/twelvelabs.pyi,sha256=F6wKHgG01bYskJpKoheBSpRpHUfFpteKn9sj9n5YfcU,372
38
+ gllm_inference/em_invoker/schema/voyage.pyi,sha256=HVpor0fqNy-IwapCICfsgFmqf1FJXCOMIxS2vOXhHd8,289
39
+ gllm_inference/exceptions/__init__.pyi,sha256=Upcuj7od2lkbdueQ0iMT2ktFYYi-KKTynTLAaxWDTjU,1214
40
+ gllm_inference/exceptions/error_parser.pyi,sha256=IOfa--NpLUW5E9Qq0mwWi6ZpTAbUyyNe6iAqunBNGLI,1999
41
+ gllm_inference/exceptions/exceptions.pyi,sha256=Bv996qLa_vju0Qjf4GewMxdkq8CV9LRZb0S6289DldA,5725
42
+ gllm_inference/exceptions/provider_error_map.pyi,sha256=XPLWU42-r8MHZgg5ZkE80Gdqg3p8Z_JHvq_Na03iTqY,1243
43
+ gllm_inference/lm_invoker/__init__.pyi,sha256=dvwZQhO-7bHDOouXBXxujux9QQbu10ux-0vAsJI1DeI,1603
44
+ gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=_by_rLLdBpnpIyniBVCOEhjur5HzdBJ1UYoXdpxb8Lw,15409
45
+ gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=drtMgbDzBQJrWXLiI2t5PNy7HtcW5Kuj0XR2b6rltjc,12936
46
+ gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=9Gz0U2c94UM9SOt-_e89_NqT_bDQ6wECRiJ9VTwsqfw,10739
47
+ gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=IqvDxBzwEf2z34FZcLKIH404y386Rnk5gsj2TcAm424,7878
48
+ gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=DrMIhGhWolSBH26jTkx9zaXTVcRBqxBwNWgzI3InODE,20063
49
+ gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=OzISl89C2s-qB6VxNlMgf5dFRC-ooj30YCFfsZzcX4s,11887
50
+ gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=0PZYitAljAjzyympqqNyN5fMyoakmqr1XIz1PE6NNc4,11176
51
+ gllm_inference/lm_invoker/lm_invoker.pyi,sha256=tgyv7A8K_36mYxL_Fkb7DFCZfJT_Hy08CKSiSJ88f0M,9370
52
+ gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=47lMyuzWKn57Ndt124-zzl9kKHa7IYzDu51iRdUPg6c,13636
53
+ gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=i5pMpZf4-r_7FQ1qfsqcjpc98sI-cPiqheuTfTEKxJs,4192
54
+ gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=oKVs5UnJJxjWaO7Rd0mmwPfBqEFeZRI09mmRdONs3js,22618
55
+ gllm_inference/lm_invoker/portkey_lm_invoker.pyi,sha256=ewxzRT-ekmvCFeb7ij840s4p18AO_LAKA-UP0ot12hs,14940
56
+ gllm_inference/lm_invoker/sea_lion_lm_invoker.pyi,sha256=ElV7iKYWnI3J1CUYuHtvOTsJByMY_l2WF4Rc7IJsBjw,3485
57
+ gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=SVIsRGcqbRnR9sqoLYWwigoEumDib5m4cTaTJT98Uz4,12765
58
+ gllm_inference/lm_invoker/batch/__init__.pyi,sha256=W4W-_yfk7lL20alREJai6GnwuQvdlKRfwQCX4mQK4XI,127
59
+ gllm_inference/lm_invoker/batch/batch_operations.pyi,sha256=cEdRVHAb5uiIPnlQudtaNzQ9kszI7cH6nD9mpMYQLco,5445
60
+ gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
61
+ gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=6lreMyHKRfZzX5NBYKnQf1Z6RzXBjTvqZj2VbMeaTLQ,1098
62
+ gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=FJLY-ZkkLUYDV48pfsLatnot4ev_xxz9xAayLK28CpU,1027
63
+ gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=WSuwOqL1j2ZioCZFC-gbB7vTRIZHQ3sU40c3ool5L6c,265
64
+ gllm_inference/lm_invoker/schema/google.pyi,sha256=MYjznjkKfNdh9XwTIrrK29tS4pkGPEU7WebVfCvFLGw,791
65
+ gllm_inference/lm_invoker/schema/langchain.pyi,sha256=rZcIxuvABI4pKfyVvkRBRqfJJogZ67EFPydpubHt49c,429
66
+ gllm_inference/lm_invoker/schema/openai.pyi,sha256=y__q2_2ZcfqCkWBw3uA6iCEQKDCm7IJoLVo3GS-ck9c,2310
67
+ gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=8byBRZ4xyTidIQJsZqiSjp5t1X875Obe-aEbT0yYfuA,1199
68
+ gllm_inference/lm_invoker/schema/portkey.pyi,sha256=NeRjHNd84HgE_ur2F3Cv6Jx30v6V7eQvI_iJiq4kuME,631
69
+ gllm_inference/lm_invoker/schema/xai.pyi,sha256=cWnbJmDtllqRH3NXpQbiXgkNBcUXr8ksDSDywcgJebE,632
70
+ gllm_inference/model/__init__.pyi,sha256=1gesoNUUHutpN2-DYmVrG5LZhpeSW96ciON3SjVEqqM,894
71
+ gllm_inference/model/em/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
72
+ gllm_inference/model/em/cohere_em.pyi,sha256=fArRlV08NwbsJ_h6vpWr94XxUVBtbqW1Jh8s42LRXCo,488
73
+ gllm_inference/model/em/google_em.pyi,sha256=ZPN5LmReO0bcTfnZixFooUTzgD-daNFPzfxzZ-5WzQQ,471
74
+ gllm_inference/model/em/jina_em.pyi,sha256=txEvDI61nhDRUMgvFzpoe-f0onpUAs1j9HPDN01IHxg,627
75
+ gllm_inference/model/em/openai_em.pyi,sha256=KcWpMmxNqS28r4zT4H2TIADHr7e7f3VSI1MPzjJXH9k,442
76
+ gllm_inference/model/em/twelvelabs_em.pyi,sha256=pf9YfTfTPAceBoe1mA5VgtCroHZi5k42mEz-mGSD5QM,400
77
+ gllm_inference/model/em/voyage_em.pyi,sha256=CEfXjLNZamfhsLyAxIkDXND2Jk4GzwXK5puK9yKJDyE,531
78
+ gllm_inference/model/lm/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
79
+ gllm_inference/model/lm/anthropic_lm.pyi,sha256=dWfG-M_gD644yJ-LK_T8HnAT649j3Vx7TVof03XQimE,611
80
+ gllm_inference/model/lm/google_lm.pyi,sha256=cMV5zYX8uwUF7pErv4pXnXD2G52umo3sxKwbSx7nFhQ,511
81
+ gllm_inference/model/lm/openai_lm.pyi,sha256=u11zvvIS7-XaHKZ33cZxGQmT6cZ4DqK9Do8l7gFOUTc,618
82
+ gllm_inference/model/lm/sea_lion_lm.pyi,sha256=k0xG5JGrecBDlNKWwNZb4BG0Ath_tGfy4fudr51-10w,492
83
+ gllm_inference/model/lm/xai_lm.pyi,sha256=2ZEQ_--e_zsb23zZQ8bKdQShU7zChx5TrDKF8EpwEpU,506
84
+ gllm_inference/output_parser/__init__.pyi,sha256=WQOOgsYnPk8vd-SOhFMMaVTzy4gkYrOAyT5gnAxv0A0,129
85
+ gllm_inference/output_parser/json_output_parser.pyi,sha256=uulh91uQLMSb4ZXZhHYi9W9w7zGnmrOweEkL6wdDJN8,2933
86
+ gllm_inference/output_parser/output_parser.pyi,sha256=Yzk7F26pH8Uc7FQZo4G6l67YkfppefUvaV9cNK-HyDs,948
87
+ gllm_inference/prompt_builder/__init__.pyi,sha256=kshfBMvwIwiIvjxiGG5BrJZNvpPa8rhtkbHo5FPifBg,117
88
+ gllm_inference/prompt_builder/prompt_builder.pyi,sha256=qJ6L81KMLdZvS2MmOGGjrJ9btmWwVPoRVl3NQxHq3fw,5010
89
+ gllm_inference/prompt_builder/format_strategy/__init__.pyi,sha256=QhORHac3ySOPmL9k9kmCKL70vtaUtwkZEtGoRNWNuA8,308
90
+ gllm_inference/prompt_builder/format_strategy/format_strategy.pyi,sha256=JSUl_7Ka08oDZPpslymkUa8pDzqNGIK2TlcVANspqrY,2273
91
+ gllm_inference/prompt_builder/format_strategy/jinja_format_strategy.pyi,sha256=IAezLUiKSJMaoDyleo8pFnFqq8rBM_Q-lNXuAGvwXhI,2225
92
+ gllm_inference/prompt_builder/format_strategy/string_format_strategy.pyi,sha256=E0r8x6NTVbPLUfbJBz75PW1n3Ong6bi1WNgxBD0FTM4,693
93
+ gllm_inference/prompt_formatter/__init__.pyi,sha256=rTsjfRsT-y00qH67fPewMNPMN1fAO2y7DM9scR1ccm0,740
94
+ gllm_inference/prompt_formatter/agnostic_prompt_formatter.pyi,sha256=c9mN4t8LXn79h8wq8DAeWYwMgmZGzXjP7EcjLpwfNZg,2018
95
+ gllm_inference/prompt_formatter/huggingface_prompt_formatter.pyi,sha256=AJ-D11HBhTKBA1B5s0km_K1R6o5HD1yvdbGoL7SpGhI,2729
96
+ gllm_inference/prompt_formatter/llama_prompt_formatter.pyi,sha256=nz24trAqHsUY_zHVYfdNzhBO_TCUtzRkg-QYYow4l2s,2848
97
+ gllm_inference/prompt_formatter/mistral_prompt_formatter.pyi,sha256=DgFitxfvCSJf8_6NUfSre-PXYMuPI4vX4AN6BPF1cBE,2650
98
+ gllm_inference/prompt_formatter/openai_prompt_formatter.pyi,sha256=QB7gHddipQuAolUKO01ApZeaRxBtRZzglS5B78mzsV8,1286
99
+ gllm_inference/prompt_formatter/prompt_formatter.pyi,sha256=UkcPi5ao98OGJyNRsqfhYTlMW-ZLNITaGZUTwzvBzqk,1146
100
+ gllm_inference/realtime_chat/__init__.pyi,sha256=Fhxr_sYOt_p43BT2uvJmCp5FMhfb3dDMoIZglq2y6G8,137
101
+ gllm_inference/realtime_chat/google_realtime_chat.pyi,sha256=S5FOCacuc7j4J2tjpUHCdp91YlalxIJ3V3Bh0t0GGus,10112
102
+ gllm_inference/realtime_chat/realtime_chat.pyi,sha256=qG0fXyhTPLw6wGpnDX6wQrLfs6RBaxCK2ASwIqOymp4,1375
103
+ gllm_inference/realtime_chat/input_streamer/__init__.pyi,sha256=REN1dEEZB-I2RX5QsBMxWcUiGm0-fWLevgj7IT18W6Y,317
104
+ gllm_inference/realtime_chat/input_streamer/input_streamer.pyi,sha256=DOVBe_lxievONS_t14opNNuv0zM3MWyBcCwmxzR4_BE,1259
105
+ gllm_inference/realtime_chat/input_streamer/keyboard_input_streamer.pyi,sha256=3L0y91efGmb4YFWigs2WGAURQ48EwNxLgcVMm15yrIg,1043
106
+ gllm_inference/realtime_chat/input_streamer/linux_mic_input_streamer.pyi,sha256=o8rOootmcKHmKRUCCHvfGlCcfeBow5UTbQHIw89XIqg,1337
107
+ gllm_inference/realtime_chat/output_streamer/__init__.pyi,sha256=FP5amnApQcIvW7YfQEnLzJBzEDz51npgocfCcRWDQXA,339
108
+ gllm_inference/realtime_chat/output_streamer/console_output_streamer.pyi,sha256=pO_llZK7JDjllTWEZjujEVo83Hvrv_8LwVqEkQiIOOo,674
109
+ gllm_inference/realtime_chat/output_streamer/linux_speaker_output_streamer.pyi,sha256=yie6VaOyozzE4bKtvWr-6hRiobK-8mshCqeUOhK1PVA,1487
110
+ gllm_inference/realtime_chat/output_streamer/output_streamer.pyi,sha256=GPAw1wPSrwHpJ2MeZJB3V9CCESTDZpyYYgsoy_LI2Nk,1083
111
+ gllm_inference/request_processor/__init__.pyi,sha256=hVnfdNZnkTBJHnmLtN3Na4ANP0yK6AstWdIizVr2Apo,227
112
+ gllm_inference/request_processor/lm_request_processor.pyi,sha256=VnYc8E3Iayyhw-rPnGPfTKuO3ohgFsS8HPrZJeyES5I,5889
113
+ gllm_inference/request_processor/uses_lm_mixin.pyi,sha256=Yu0XPNuHxq1tWBviHTPw1oThojneFwGHepvGjBXxKQA,6382
114
+ gllm_inference/schema/__init__.pyi,sha256=OOZE9H4DgUqOsk5_01hXb5opkAwBmp1TvhqcV2SrqJY,2505
115
+ gllm_inference/schema/activity.pyi,sha256=JnO2hqj91P5Tc6qb4pbkEMrHer2u5owiCvhl-igcQKQ,2303
116
+ gllm_inference/schema/attachment.pyi,sha256=oCopoxiPgGSkCRdPsqmjcMofTawfbdCDxaPdo6mits0,4509
117
+ gllm_inference/schema/code_exec_result.pyi,sha256=ZTHh6JtRrPIdQ059P1UAiD2L-tAO1_S5YcMsAXfJ5A0,559
118
+ gllm_inference/schema/config.pyi,sha256=rAL_UeXyQeXVk1P2kqd8vFWOMwmKenfpQLtvMP74t9s,674
119
+ gllm_inference/schema/enums.pyi,sha256=rLTlnhtdTIKT_Q8p-ukKpCMhRk9NbqcR_ylZ0-8UivA,2111
120
+ gllm_inference/schema/events.pyi,sha256=XQEy5SqYoutq2DguwCCTGi5DHrlnDai6nJElAYsj1gk,4638
121
+ gllm_inference/schema/formatter.pyi,sha256=qPQ1oSnmSgr7yBsBBMe-aehLLk9lKI9OPZ8Og-EbzdI,1281
122
+ gllm_inference/schema/lm_input.pyi,sha256=A5pjz1id6tP9XRNhzQrbmzd66C_q3gzo0UP8rCemz6Q,193
123
+ gllm_inference/schema/lm_output.pyi,sha256=ec5ErE68PVthNHhHlaVHsiAN3nn03yBqRR7HVabsf68,11401
124
+ gllm_inference/schema/mcp.pyi,sha256=Vwu8E2BDl6FvvnI42gIyY3Oki1BdwRE3Uh3aV0rmhQU,1014
125
+ gllm_inference/schema/message.pyi,sha256=VP9YppKj2mo1esl9cy6qQO9m2mMHUjTmfGDdyUor880,2220
126
+ gllm_inference/schema/model_id.pyi,sha256=MuH0KyFjI1uC9v7PoIU6Uuk6wPdpmczVrHZj0r5EcZk,5842
127
+ gllm_inference/schema/reasoning.pyi,sha256=SlTuiDw87GdnAn-I6YOPIJRhEBiwQljM46JohG05guQ,562
128
+ gllm_inference/schema/token_usage.pyi,sha256=1GTQVORV0dBNmD_jix8aVaUqxMKFF04KpLP7y2urqbk,2950
129
+ gllm_inference/schema/tool_call.pyi,sha256=zQaVxCnkVxOfOEhBidqohU85gb4PRwnwBiygKaunamk,389
130
+ gllm_inference/schema/tool_result.pyi,sha256=cAG7TVtB4IWJPt8XBBbB92cuY1ZsX9M276bN9aqjcvM,276
131
+ gllm_inference/schema/type_alias.pyi,sha256=q_C1x8hUUoerGEQFj99g3yaZ1ILigGQSEo-FheqS_lM,529
132
+ gllm_inference/utils/__init__.pyi,sha256=mDJ2gLSeQzm-79Tov5-dhrMNaDqgcD1VVzDYAWvIRqA,391
133
+ gllm_inference/utils/io_utils.pyi,sha256=7kUTacHAVRYoemFUOjCH7-Qmw-YsQGd6rGYxjf_qmtw,1094
134
+ gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
135
+ gllm_inference/utils/validation.pyi,sha256=W9RQddN90F8SJMu_HXEQyQTDMBaRL-bo7fOosZWK7oY,438
136
+ gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
137
+ gllm_inference_binary-0.5.65.dist-info/METADATA,sha256=ZoIOAqMSWLFjLzw3e3FTD-H3GxGsndl9lxlK4MbwsgA,5815
138
+ gllm_inference_binary-0.5.65.dist-info/WHEEL,sha256=t5_2FPbm0tV7_MDNYR2NrQvum0LVPlNiR9hm8r8eRow,104
139
+ gllm_inference_binary-0.5.65.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
140
+ gllm_inference_binary-0.5.65.dist-info/RECORD,,
@@ -0,0 +1,5 @@
1
+ Wheel-Version: 1.0
2
+ Generator: Nuitka (2.6.9)
3
+ Root-Is-Purelib: false
4
+ Tag: cp313-cp313-macosx_13_0_arm64
5
+
@@ -0,0 +1 @@
1
+ gllm_inference