gllm-inference-binary 0.5.51b3__cp313-cp313-macosx_11_0_arm64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (137) hide show
  1. gllm_inference/__init__.pyi +0 -0
  2. gllm_inference/builder/__init__.pyi +6 -0
  3. gllm_inference/builder/_build_invoker.pyi +28 -0
  4. gllm_inference/builder/build_em_invoker.pyi +130 -0
  5. gllm_inference/builder/build_lm_invoker.pyi +213 -0
  6. gllm_inference/builder/build_lm_request_processor.pyi +88 -0
  7. gllm_inference/builder/build_output_parser.pyi +29 -0
  8. gllm_inference/catalog/__init__.pyi +4 -0
  9. gllm_inference/catalog/catalog.pyi +121 -0
  10. gllm_inference/catalog/lm_request_processor_catalog.pyi +112 -0
  11. gllm_inference/catalog/prompt_builder_catalog.pyi +82 -0
  12. gllm_inference/constants.pyi +13 -0
  13. gllm_inference/em_invoker/__init__.pyi +12 -0
  14. gllm_inference/em_invoker/azure_openai_em_invoker.pyi +88 -0
  15. gllm_inference/em_invoker/bedrock_em_invoker.pyi +118 -0
  16. gllm_inference/em_invoker/cohere_em_invoker.pyi +128 -0
  17. gllm_inference/em_invoker/em_invoker.pyi +90 -0
  18. gllm_inference/em_invoker/google_em_invoker.pyi +129 -0
  19. gllm_inference/em_invoker/jina_em_invoker.pyi +103 -0
  20. gllm_inference/em_invoker/langchain/__init__.pyi +3 -0
  21. gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi +84 -0
  22. gllm_inference/em_invoker/langchain_em_invoker.pyi +46 -0
  23. gllm_inference/em_invoker/openai_compatible_em_invoker.pyi +41 -0
  24. gllm_inference/em_invoker/openai_em_invoker.pyi +118 -0
  25. gllm_inference/em_invoker/schema/__init__.pyi +0 -0
  26. gllm_inference/em_invoker/schema/bedrock.pyi +29 -0
  27. gllm_inference/em_invoker/schema/cohere.pyi +20 -0
  28. gllm_inference/em_invoker/schema/google.pyi +9 -0
  29. gllm_inference/em_invoker/schema/jina.pyi +29 -0
  30. gllm_inference/em_invoker/schema/langchain.pyi +5 -0
  31. gllm_inference/em_invoker/schema/openai.pyi +7 -0
  32. gllm_inference/em_invoker/schema/openai_compatible.pyi +7 -0
  33. gllm_inference/em_invoker/schema/twelvelabs.pyi +17 -0
  34. gllm_inference/em_invoker/schema/voyage.pyi +15 -0
  35. gllm_inference/em_invoker/twelevelabs_em_invoker.pyi +101 -0
  36. gllm_inference/em_invoker/voyage_em_invoker.pyi +104 -0
  37. gllm_inference/exceptions/__init__.pyi +4 -0
  38. gllm_inference/exceptions/error_parser.pyi +41 -0
  39. gllm_inference/exceptions/exceptions.pyi +132 -0
  40. gllm_inference/exceptions/provider_error_map.pyi +24 -0
  41. gllm_inference/lm_invoker/__init__.pyi +14 -0
  42. gllm_inference/lm_invoker/anthropic_lm_invoker.pyi +334 -0
  43. gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi +254 -0
  44. gllm_inference/lm_invoker/batch/__init__.pyi +3 -0
  45. gllm_inference/lm_invoker/batch/batch_operations.pyi +127 -0
  46. gllm_inference/lm_invoker/bedrock_lm_invoker.pyi +226 -0
  47. gllm_inference/lm_invoker/datasaur_lm_invoker.pyi +157 -0
  48. gllm_inference/lm_invoker/google_lm_invoker.pyi +336 -0
  49. gllm_inference/lm_invoker/langchain_lm_invoker.pyi +251 -0
  50. gllm_inference/lm_invoker/litellm_lm_invoker.pyi +244 -0
  51. gllm_inference/lm_invoker/lm_invoker.pyi +164 -0
  52. gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi +272 -0
  53. gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi +52 -0
  54. gllm_inference/lm_invoker/openai_lm_invoker.pyi +433 -0
  55. gllm_inference/lm_invoker/portkey_lm_invoker.pyi +260 -0
  56. gllm_inference/lm_invoker/schema/__init__.pyi +0 -0
  57. gllm_inference/lm_invoker/schema/anthropic.pyi +56 -0
  58. gllm_inference/lm_invoker/schema/bedrock.pyi +53 -0
  59. gllm_inference/lm_invoker/schema/datasaur.pyi +14 -0
  60. gllm_inference/lm_invoker/schema/google.pyi +24 -0
  61. gllm_inference/lm_invoker/schema/langchain.pyi +23 -0
  62. gllm_inference/lm_invoker/schema/openai.pyi +106 -0
  63. gllm_inference/lm_invoker/schema/openai_chat_completions.pyi +62 -0
  64. gllm_inference/lm_invoker/schema/portkey.pyi +31 -0
  65. gllm_inference/lm_invoker/schema/xai.pyi +31 -0
  66. gllm_inference/lm_invoker/xai_lm_invoker.pyi +289 -0
  67. gllm_inference/model/__init__.pyi +12 -0
  68. gllm_inference/model/em/__init__.pyi +0 -0
  69. gllm_inference/model/em/cohere_em.pyi +17 -0
  70. gllm_inference/model/em/google_em.pyi +16 -0
  71. gllm_inference/model/em/jina_em.pyi +22 -0
  72. gllm_inference/model/em/openai_em.pyi +15 -0
  73. gllm_inference/model/em/twelvelabs_em.pyi +13 -0
  74. gllm_inference/model/em/voyage_em.pyi +20 -0
  75. gllm_inference/model/lm/__init__.pyi +0 -0
  76. gllm_inference/model/lm/anthropic_lm.pyi +22 -0
  77. gllm_inference/model/lm/google_lm.pyi +18 -0
  78. gllm_inference/model/lm/openai_lm.pyi +27 -0
  79. gllm_inference/model/lm/xai_lm.pyi +19 -0
  80. gllm_inference/output_parser/__init__.pyi +3 -0
  81. gllm_inference/output_parser/json_output_parser.pyi +60 -0
  82. gllm_inference/output_parser/output_parser.pyi +27 -0
  83. gllm_inference/prompt_builder/__init__.pyi +3 -0
  84. gllm_inference/prompt_builder/format_strategy/__init__.pyi +4 -0
  85. gllm_inference/prompt_builder/format_strategy/format_strategy.pyi +55 -0
  86. gllm_inference/prompt_builder/format_strategy/jinja_format_strategy.pyi +45 -0
  87. gllm_inference/prompt_builder/format_strategy/string_format_strategy.pyi +20 -0
  88. gllm_inference/prompt_builder/prompt_builder.pyi +69 -0
  89. gllm_inference/prompt_formatter/__init__.pyi +7 -0
  90. gllm_inference/prompt_formatter/agnostic_prompt_formatter.pyi +49 -0
  91. gllm_inference/prompt_formatter/huggingface_prompt_formatter.pyi +55 -0
  92. gllm_inference/prompt_formatter/llama_prompt_formatter.pyi +59 -0
  93. gllm_inference/prompt_formatter/mistral_prompt_formatter.pyi +53 -0
  94. gllm_inference/prompt_formatter/openai_prompt_formatter.pyi +35 -0
  95. gllm_inference/prompt_formatter/prompt_formatter.pyi +30 -0
  96. gllm_inference/realtime_chat/__init__.pyi +3 -0
  97. gllm_inference/realtime_chat/google_realtime_chat.pyi +205 -0
  98. gllm_inference/realtime_chat/input_streamer/__init__.pyi +4 -0
  99. gllm_inference/realtime_chat/input_streamer/input_streamer.pyi +36 -0
  100. gllm_inference/realtime_chat/input_streamer/keyboard_input_streamer.pyi +27 -0
  101. gllm_inference/realtime_chat/input_streamer/linux_mic_input_streamer.pyi +36 -0
  102. gllm_inference/realtime_chat/output_streamer/__init__.pyi +4 -0
  103. gllm_inference/realtime_chat/output_streamer/console_output_streamer.pyi +21 -0
  104. gllm_inference/realtime_chat/output_streamer/linux_speaker_output_streamer.pyi +42 -0
  105. gllm_inference/realtime_chat/output_streamer/output_streamer.pyi +33 -0
  106. gllm_inference/realtime_chat/realtime_chat.pyi +28 -0
  107. gllm_inference/request_processor/__init__.pyi +4 -0
  108. gllm_inference/request_processor/lm_request_processor.pyi +101 -0
  109. gllm_inference/request_processor/uses_lm_mixin.pyi +130 -0
  110. gllm_inference/schema/__init__.pyi +18 -0
  111. gllm_inference/schema/activity.pyi +64 -0
  112. gllm_inference/schema/attachment.pyi +88 -0
  113. gllm_inference/schema/code_exec_result.pyi +14 -0
  114. gllm_inference/schema/config.pyi +15 -0
  115. gllm_inference/schema/enums.pyi +69 -0
  116. gllm_inference/schema/events.pyi +105 -0
  117. gllm_inference/schema/lm_input.pyi +4 -0
  118. gllm_inference/schema/lm_output.pyi +44 -0
  119. gllm_inference/schema/mcp.pyi +31 -0
  120. gllm_inference/schema/message.pyi +52 -0
  121. gllm_inference/schema/model_id.pyi +176 -0
  122. gllm_inference/schema/reasoning.pyi +15 -0
  123. gllm_inference/schema/token_usage.pyi +75 -0
  124. gllm_inference/schema/tool_call.pyi +14 -0
  125. gllm_inference/schema/tool_result.pyi +11 -0
  126. gllm_inference/schema/type_alias.pyi +11 -0
  127. gllm_inference/utils/__init__.pyi +5 -0
  128. gllm_inference/utils/io_utils.pyi +26 -0
  129. gllm_inference/utils/langchain.pyi +30 -0
  130. gllm_inference/utils/validation.pyi +12 -0
  131. gllm_inference.build/.gitignore +1 -0
  132. gllm_inference.cpython-313-darwin.so +0 -0
  133. gllm_inference.pyi +157 -0
  134. gllm_inference_binary-0.5.51b3.dist-info/METADATA +138 -0
  135. gllm_inference_binary-0.5.51b3.dist-info/RECORD +137 -0
  136. gllm_inference_binary-0.5.51b3.dist-info/WHEEL +5 -0
  137. gllm_inference_binary-0.5.51b3.dist-info/top_level.txt +1 -0
@@ -0,0 +1,30 @@
1
+ from _typeshed import Incomplete
2
+ from langchain_core.embeddings import Embeddings as Embeddings
3
+ from langchain_core.language_models import BaseChatModel as BaseChatModel
4
+ from typing import Any
5
+
6
+ MODEL_NAME_KEYS: Incomplete
7
+
8
+ def load_langchain_model(model_class_path: str, model_name: str, model_kwargs: dict[str, Any]) -> BaseChatModel | Embeddings:
9
+ '''Loads the LangChain\'s model instance.
10
+
11
+ Args:
12
+ model_class_path (str): The path to the LangChain\'s class, e.g. "langchain_openai.ChatOpenAI".
13
+ model_name (str): The model name.
14
+ model_kwargs (dict[str, Any]): The additional keyword arguments.
15
+
16
+ Returns:
17
+ BaseChatModel | Embeddings: The LangChain\'s model instance.
18
+ '''
19
+ def parse_model_data(model: BaseChatModel | Embeddings) -> dict[str, str]:
20
+ """Parses the model data from LangChain's BaseChatModel or Embeddings instance.
21
+
22
+ Args:
23
+ model (BaseChatModel | Embeddings): The LangChain's BaseChatModel or Embeddings instance.
24
+
25
+ Returns:
26
+ dict[str, str]: The dictionary containing the model name and path.
27
+
28
+ Raises:
29
+ ValueError: If the model name is not found in the model data.
30
+ """
@@ -0,0 +1,12 @@
1
+ from enum import StrEnum
2
+
3
+ def validate_string_enum(enum_type: type[StrEnum], value: str) -> None:
4
+ """Validates that the provided value is a valid string enum value.
5
+
6
+ Args:
7
+ enum_type (type[StrEnum]): The type of the string enum.
8
+ value (str): The value to validate.
9
+
10
+ Raises:
11
+ ValueError: If the provided value is not a valid string enum value.
12
+ """
@@ -0,0 +1 @@
1
+ *
Binary file
gllm_inference.pyi ADDED
@@ -0,0 +1,157 @@
1
+ # This file was generated by Nuitka
2
+
3
+ # Stubs included by default
4
+
5
+
6
+ __name__ = ...
7
+
8
+
9
+
10
+ # Modules used internally, to allow implicit dependencies to be seen:
11
+ import os
12
+ import typing
13
+ import gllm_core
14
+ import gllm_core.utils
15
+ import gllm_inference.em_invoker.AzureOpenAIEMInvoker
16
+ import gllm_inference.em_invoker.BedrockEMInvoker
17
+ import gllm_inference.em_invoker.CohereEMInvoker
18
+ import gllm_inference.em_invoker.GoogleEMInvoker
19
+ import gllm_inference.em_invoker.JinaEMInvoker
20
+ import gllm_inference.em_invoker.LangChainEMInvoker
21
+ import gllm_inference.em_invoker.OpenAICompatibleEMInvoker
22
+ import gllm_inference.em_invoker.OpenAIEMInvoker
23
+ import gllm_inference.em_invoker.TwelveLabsEMInvoker
24
+ import gllm_inference.em_invoker.VoyageEMInvoker
25
+ import gllm_inference.lm_invoker.AnthropicLMInvoker
26
+ import gllm_inference.lm_invoker.AzureOpenAILMInvoker
27
+ import gllm_inference.lm_invoker.BedrockLMInvoker
28
+ import gllm_inference.lm_invoker.DatasaurLMInvoker
29
+ import gllm_inference.lm_invoker.GoogleLMInvoker
30
+ import gllm_inference.lm_invoker.LangChainLMInvoker
31
+ import gllm_inference.lm_invoker.LiteLLMLMInvoker
32
+ import gllm_inference.lm_invoker.OpenAIChatCompletionsLMInvoker
33
+ import gllm_inference.lm_invoker.OpenAICompatibleLMInvoker
34
+ import gllm_inference.lm_invoker.OpenAILMInvoker
35
+ import gllm_inference.lm_invoker.PortkeyLMInvoker
36
+ import gllm_inference.lm_invoker.XAILMInvoker
37
+ import gllm_inference.prompt_builder.PromptBuilder
38
+ import gllm_inference.output_parser.JSONOutputParser
39
+ import json
40
+ import abc
41
+ import pandas
42
+ import pydantic
43
+ import re
44
+ import gllm_core.utils.retry
45
+ import gllm_inference.request_processor.LMRequestProcessor
46
+ import gllm_core.utils.imports
47
+ import gllm_inference.schema.ModelId
48
+ import gllm_inference.schema.ModelProvider
49
+ import gllm_inference.schema.TruncationConfig
50
+ import asyncio
51
+ import base64
52
+ import enum
53
+ import gllm_inference.exceptions.BaseInvokerError
54
+ import gllm_inference.exceptions.convert_http_status_to_base_invoker_error
55
+ import gllm_inference.schema.Attachment
56
+ import gllm_inference.schema.AttachmentType
57
+ import gllm_inference.schema.EMContent
58
+ import gllm_inference.schema.Vector
59
+ import aioboto3
60
+ import gllm_inference.utils.validate_string_enum
61
+ import cohere
62
+ import asyncio.CancelledError
63
+ import gllm_inference.exceptions.convert_to_base_invoker_error
64
+ import gllm_inference.schema.TruncateSide
65
+ import google
66
+ import google.auth
67
+ import google.genai
68
+ import google.genai.types
69
+ import httpx
70
+ import gllm_inference.exceptions.ProviderInternalError
71
+ import concurrent
72
+ import concurrent.futures
73
+ import concurrent.futures.ThreadPoolExecutor
74
+ import langchain_core
75
+ import langchain_core.embeddings
76
+ import gllm_inference.exceptions.InvokerRuntimeError
77
+ import gllm_inference.exceptions.build_debug_info
78
+ import gllm_inference.utils.load_langchain_model
79
+ import gllm_inference.utils.parse_model_data
80
+ import openai
81
+ import io
82
+ import twelvelabs
83
+ import sys
84
+ import voyageai
85
+ import voyageai.client_async
86
+ import http
87
+ import http.HTTPStatus
88
+ import uuid
89
+ import gllm_core.constants
90
+ import gllm_core.event
91
+ import gllm_core.schema
92
+ import gllm_core.schema.tool
93
+ import langchain_core.tools
94
+ import gllm_inference.schema.BatchStatus
95
+ import gllm_inference.schema.LMInput
96
+ import gllm_inference.schema.LMOutput
97
+ import gllm_inference.schema.Message
98
+ import gllm_inference.schema.Reasoning
99
+ import gllm_inference.schema.ResponseSchema
100
+ import gllm_inference.schema.ThinkingEvent
101
+ import gllm_inference.schema.TokenUsage
102
+ import gllm_inference.schema.ToolCall
103
+ import gllm_inference.schema.ToolResult
104
+ import anthropic
105
+ import anthropic.types
106
+ import anthropic.types.message_create_params
107
+ import anthropic.types.messages
108
+ import anthropic.types.messages.batch_create_params
109
+ import gllm_inference.schema.MessageRole
110
+ import langchain_core.language_models
111
+ import langchain_core.messages
112
+ import gllm_inference.exceptions._get_exception_key
113
+ import litellm
114
+ import inspect
115
+ import time
116
+ import jsonschema
117
+ import gllm_inference.lm_invoker.batch.BatchOperations
118
+ import gllm_inference.schema.LMEventType
119
+ import gllm_inference.schema.MessageContent
120
+ import __future__
121
+ import gllm_inference.schema.ActivityEvent
122
+ import gllm_inference.schema.CodeEvent
123
+ import gllm_inference.schema.CodeExecResult
124
+ import gllm_inference.schema.MCPCall
125
+ import gllm_inference.schema.MCPCallActivity
126
+ import gllm_inference.schema.MCPListToolsActivity
127
+ import gllm_inference.schema.MCPServer
128
+ import gllm_inference.schema.WebSearchActivity
129
+ import logging
130
+ import portkey_ai
131
+ import xai_sdk
132
+ import xai_sdk.chat
133
+ import xai_sdk.search
134
+ import xai_sdk.proto
135
+ import xai_sdk.proto.v5
136
+ import xai_sdk.proto.v5.chat_pb2
137
+ import jinja2
138
+ import jinja2.sandbox
139
+ import gllm_inference.schema.JinjaEnvType
140
+ import gllm_inference.prompt_builder.format_strategy.JinjaFormatStrategy
141
+ import gllm_inference.prompt_builder.format_strategy.StringFormatStrategy
142
+ import transformers
143
+ import gllm_inference.prompt_formatter.HuggingFacePromptFormatter
144
+ import traceback
145
+ import gllm_inference.realtime_chat.input_streamer.KeyboardInputStreamer
146
+ import gllm_inference.realtime_chat.output_streamer.ConsoleOutputStreamer
147
+ import google.genai.live
148
+ import gllm_core.utils.logger_manager
149
+ import mimetypes
150
+ import pathlib
151
+ import pathlib.Path
152
+ import filetype
153
+ import magic
154
+ import requests
155
+ import binascii
156
+ import fnmatch
157
+ import importlib
@@ -0,0 +1,138 @@
1
+ Metadata-Version: 2.2
2
+ Name: gllm-inference-binary
3
+ Version: 0.5.51b3
4
+ Summary: A library containing components related to model inferences in Gen AI applications.
5
+ Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, Resti Febrina <resti.febrina@gdplabs.id>
6
+ Requires-Python: <3.14,>=3.11
7
+ Description-Content-Type: text/markdown
8
+ Requires-Dist: poetry<3.0.0,>=2.1.3
9
+ Requires-Dist: gllm-core-binary==0.3.23b2
10
+ Requires-Dist: aiohttp<4.0.0,>=3.12.14
11
+ Requires-Dist: filetype<2.0.0,>=1.2.0
12
+ Requires-Dist: httpx<0.29.0,>=0.28.0
13
+ Requires-Dist: jinja2<4.0.0,>=3.1.4
14
+ Requires-Dist: jsonschema<5.0.0,>=4.24.0
15
+ Requires-Dist: langchain<0.4.0,>=0.3.0
16
+ Requires-Dist: pandas<3.0.0,>=2.2.3
17
+ Requires-Dist: protobuf<7.0.0,>=5.29.4
18
+ Requires-Dist: python-magic<0.5.0,>=0.4.27; sys_platform != "win32"
19
+ Requires-Dist: python-magic-bin<0.5.0,>=0.4.14; sys_platform == "win32"
20
+ Requires-Dist: sentencepiece<0.3.0,>=0.2.0
21
+ Provides-Extra: dev
22
+ Requires-Dist: coverage<8.0.0,>=7.4.4; extra == "dev"
23
+ Requires-Dist: mypy<2.0.0,>=1.15.0; extra == "dev"
24
+ Requires-Dist: pre-commit<4.0.0,>=3.7.0; extra == "dev"
25
+ Requires-Dist: pytest<9.0.0,>=8.1.1; extra == "dev"
26
+ Requires-Dist: pytest-asyncio<0.24.0,>=0.23.6; extra == "dev"
27
+ Requires-Dist: pytest-cov<6.0.0,>=5.0.0; extra == "dev"
28
+ Requires-Dist: ruff<0.7.0,>=0.6.7; extra == "dev"
29
+ Provides-Extra: anthropic
30
+ Requires-Dist: anthropic<0.61.0,>=0.60.0; extra == "anthropic"
31
+ Provides-Extra: bedrock
32
+ Requires-Dist: aioboto3<16.0.0,>=15.0.0; extra == "bedrock"
33
+ Provides-Extra: cohere
34
+ Requires-Dist: cohere<6.0.0,>=5.18.0; extra == "cohere"
35
+ Provides-Extra: datasaur
36
+ Requires-Dist: openai<2.0.0,>=1.98.0; extra == "datasaur"
37
+ Provides-Extra: google
38
+ Requires-Dist: google-genai<=1.36,>=1.23; extra == "google"
39
+ Provides-Extra: huggingface
40
+ Requires-Dist: huggingface-hub<0.31.0,>=0.30.0; extra == "huggingface"
41
+ Requires-Dist: transformers<5.0.0,>=4.52.0; extra == "huggingface"
42
+ Provides-Extra: litellm
43
+ Requires-Dist: litellm<2.0.0,>=1.69.2; extra == "litellm"
44
+ Provides-Extra: openai
45
+ Requires-Dist: openai<2.0.0,>=1.98.0; extra == "openai"
46
+ Provides-Extra: portkey-ai
47
+ Requires-Dist: portkey-ai<2.0.0,>=1.14.4; extra == "portkey-ai"
48
+ Provides-Extra: twelvelabs
49
+ Requires-Dist: twelvelabs<0.5.0,>=0.4.4; extra == "twelvelabs"
50
+ Provides-Extra: voyage
51
+ Requires-Dist: voyageai<0.4.0,>=0.3.0; python_version < "3.13" and extra == "voyage"
52
+ Provides-Extra: xai
53
+ Requires-Dist: xai_sdk<2.0.0,>=1.0.0; extra == "xai"
54
+
55
+ # GLLM Inference
56
+
57
+ ## Description
58
+
59
+ A library containing components related to model inferences in Gen AI applications.
60
+
61
+ ## Installation
62
+
63
+ ### Prerequisites
64
+ 1. Python 3.11+ - [Install here](https://www.python.org/downloads/)
65
+ 2. Pip (if using Pip) - [Install here](https://pip.pypa.io/en/stable/installation/)
66
+ 3. Poetry (automatically installed via Makefile) - [Install here](https://python-poetry.org/docs/#installation)
67
+ 4. Git (if using Git) - [Install here](https://git-scm.com/downloads)
68
+ 5. gcloud CLI (for authentication) - [Install here](https://cloud.google.com/sdk/docs/install)
69
+ 6. For git installation, access to the [GDP Labs SDK github repository](https://github.com/GDP-ADMIN/gl-sdk)
70
+
71
+ ### 1. Installation from Artifact Registry
72
+ Choose one of the following methods to install the package:
73
+
74
+ #### Using pip
75
+ ```bash
76
+ pip install gllm-inference-binary
77
+ ```
78
+
79
+ #### Using Poetry
80
+ ```bash
81
+ poetry add gllm-inference-binary
82
+ ```
83
+
84
+ ### 2. Development Installation (Git)
85
+ For development purposes, you can install directly from the Git repository:
86
+ ```bash
87
+ poetry add "git+ssh://git@github.com/GDP-ADMIN/gen-ai-internal.git#subdirectory=libs/gllm-inference"
88
+ ```
89
+
90
+ Available extras:
91
+ - `anthropic`: Install Anthropic models dependencies
92
+ - `google-genai`: Install Google Generative AI models dependencies
93
+ - `google-vertexai`: Install Google Vertex AI models dependencies
94
+ - `huggingface`: Install HuggingFace models dependencies
95
+ - `openai`: Install OpenAI models dependencies
96
+ - `twelvelabs`: Install TwelveLabs models dependencies
97
+
98
+ ## Local Development Setup
99
+
100
+ ### Quick Setup (Recommended)
101
+ For local development with editable gllm packages, use the provided Makefile:
102
+
103
+ ```bash
104
+ # Complete setup: installs Poetry, configures auth, installs packages, sets up pre-commit
105
+ make setup
106
+ ```
107
+
108
+ The following are the available Makefile targets:
109
+
110
+ 1. `make setup` - Complete development setup (recommended for new developers)
111
+ 2. `make install-poetry` - Install or upgrade Poetry to the latest version
112
+ 3. `make auth` - Configure authentication for internal repositories
113
+ 4. `make install` - Install all dependencies
114
+ 5. `make install-pre-commit` - Set up pre-commit hooks
115
+ 6. `make update` - Update dependencies
116
+ ### Manual Development Setup (Legacy)
117
+ If you prefer to manage dependencies manually:
118
+
119
+ 1. Go to root folder of `gllm-inference` module, e.g. `cd libs/gllm-inference`.
120
+ 2. Run `poetry shell` to create a virtual environment.
121
+ 3. Run `poetry lock` to create a lock file if you haven't done it yet.
122
+ 4. Run `poetry install` to install the `gllm-inference` requirements for the first time.
123
+ 5. Run `poetry update` if you update any dependency module version at `pyproject.toml`.
124
+
125
+
126
+ ## Contributing
127
+ Please refer to this [Python Style Guide](https://docs.google.com/document/d/1uRggCrHnVfDPBnG641FyQBwUwLoFw0kTzNqRm92vUwM/edit?usp=sharing)
128
+ to get information about code style, documentation standard, and SCA that you need to use when contributing to this project
129
+
130
+ ### Getting Started with Development
131
+ 1. Clone the repository and navigate to the gllm-inference directory
132
+ 2. Run `make setup` to set up your development environment
133
+ 3. Run `which python` to get the path to be referenced at Visual Studio Code interpreter path (`Ctrl`+`Shift`+`P` or `Cmd`+`Shift`+`P`)
134
+ 4. Try running the unit test to see if it's working:
135
+ ```bash
136
+ poetry run pytest -s tests/unit_tests/
137
+ ```
138
+ 5. When you want to update the dependencies, run `make update`
@@ -0,0 +1,137 @@
1
+ gllm_inference.cpython-313-darwin.so,sha256=e0m_8i2W9bQrHKjotHdO9i17dUgPRrd4ZaXup8BzYNY,4945704
2
+ gllm_inference.pyi,sha256=3TbylLc0CkZzlLT7WqOloWNruBH0tojkkQJ_krV9mQs,5211
3
+ gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
4
+ gllm_inference/constants.pyi,sha256=viU-ACRbVSGvsCJ0FQmuR1yhyl-BzoHDVIWo5cwHmF0,337
5
+ gllm_inference/builder/__init__.pyi,sha256=usz2lvfwO4Yk-ZGKXbCWG1cEr3nlQXxMNDNC-2yc1NM,500
6
+ gllm_inference/builder/_build_invoker.pyi,sha256=v__-YT0jGmLqEsgl7Abk9we-wYWcyVFvlRN_Uu4vVak,848
7
+ gllm_inference/builder/build_em_invoker.pyi,sha256=0IVcRGaciPBAFcgvGRC4-Kje_PPXOSug77LqxRc-x_U,5749
8
+ gllm_inference/builder/build_lm_invoker.pyi,sha256=TBAr7Sk9Jgckdlfj69pYOzZFMsooL-FpfbPYt_kCcXU,8852
9
+ gllm_inference/builder/build_lm_request_processor.pyi,sha256=KbQkcPa8C-yzyelht4mWLP8kDmh17itAT3tn8ZJB6pg,4144
10
+ gllm_inference/builder/build_output_parser.pyi,sha256=_Lrq-bh1oPsb_Nwkkr_zyEUwIOMysRFZkvEtEM29LZM,936
11
+ gllm_inference/catalog/__init__.pyi,sha256=JBkPGTyiiZ30GECzJBW-mW8LekWyY2qyzal3eW7ynaM,287
12
+ gllm_inference/catalog/catalog.pyi,sha256=a4RNG1lKv51GxQpOqh47tz-PAROMPaeP2o5XNLBSZaU,4790
13
+ gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=WW1j8jWujnatF0c9rCk94CyobtFe3gOky2vjoGCV6nw,5424
14
+ gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=OU8k_4HbqjZEzHZlzSM3uzGQZJmM2uGD76Csqom0CEQ,3197
15
+ gllm_inference/em_invoker/__init__.pyi,sha256=FHFR4k8RqtdMNZkda6A89tvWFtIoxF-t_NahDGeBQ2k,1199
16
+ gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=fE-RZ10Leerbyqp8ISXeqdU_m0kaHdlHml_nQqwFjqY,4961
17
+ gllm_inference/em_invoker/bedrock_em_invoker.pyi,sha256=3GkVO9hvmYck4NJY5spAg6opSxr0v8OJ4qhMvz1rcAA,6365
18
+ gllm_inference/em_invoker/cohere_em_invoker.pyi,sha256=LSNVGx58JdSP2YB6CziOS_v5VuSLdARVpBBWhz8t5E4,6665
19
+ gllm_inference/em_invoker/em_invoker.pyi,sha256=KGjLiAWGIA3ziV50zMwSzx6lTVUbYspQCl4LFxqbDlY,5101
20
+ gllm_inference/em_invoker/google_em_invoker.pyi,sha256=OgIQhT2g6g09Hqr9mc2--MjV3Ti-XuHfqLTunvcqU1k,6824
21
+ gllm_inference/em_invoker/jina_em_invoker.pyi,sha256=WZJ8BeaqxrAGMrUtIEUy3itcOEzvRQYEwe0gazIv6ng,5651
22
+ gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=BhZjkYZoxQXPJjP0PgA8W0xrtwiqDkW5E6NpGit0h1E,3498
23
+ gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=z4_jEuaprna6AJF2wXH-JgmynpVFXeb2HQZr-0ATmMw,2882
24
+ gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=tb6cXTvAGr01MMSc96mYWDqK-EK_VCdDOQ-Be0_EZfM,6195
25
+ gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=mitrJZW3M43k306cHwJoDjBfN1nr20C4sZ42Xr6hfO0,5347
26
+ gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=gyYQA3rbekHJIvvY6TE4tTt8W-DbPqjiPW6FaJoQPHI,5517
27
+ gllm_inference/em_invoker/langchain/__init__.pyi,sha256=VYGKE5OgU0my1RlhgzkU_A7-GLGnUDDnNFuctuRwILE,148
28
+ gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi,sha256=VU3-Vhb9BCDhJo8PPdWHe2rBEOCs_HMXT6ZaWwjUzZE,3304
29
+ gllm_inference/em_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
30
+ gllm_inference/em_invoker/schema/bedrock.pyi,sha256=ePYwailnZAYbZKt9HRDi4iAISq003jpxQPxDCiFdEZg,615
31
+ gllm_inference/em_invoker/schema/cohere.pyi,sha256=UTbTtePRR1zJMsM09SiTZSZZP0IaUGaODvc7ZqH9S8c,547
32
+ gllm_inference/em_invoker/schema/google.pyi,sha256=ovDlvinu99QJhIxMkvVUoGBEFkkEoAZhadSuk0nI9N8,181
33
+ gllm_inference/em_invoker/schema/jina.pyi,sha256=vE1ySd8OTDM35saEZos7UCdPwHeX66iuHkZ3RchSZKA,741
34
+ gllm_inference/em_invoker/schema/langchain.pyi,sha256=edcUvc1IHoSMFwqV83uqWqd0U3fLhkyWQjVknvjHI8U,112
35
+ gllm_inference/em_invoker/schema/openai.pyi,sha256=Q_dsEcodkOXYXPdrkOkW0LnuLhfeq8tEbtZAGMz2ajA,139
36
+ gllm_inference/em_invoker/schema/openai_compatible.pyi,sha256=gmvGtsWoOMBelke_tZjC6dKimFBW9f4Vrgv0Ig0OM9Q,150
37
+ gllm_inference/em_invoker/schema/twelvelabs.pyi,sha256=F6wKHgG01bYskJpKoheBSpRpHUfFpteKn9sj9n5YfcU,372
38
+ gllm_inference/em_invoker/schema/voyage.pyi,sha256=HVpor0fqNy-IwapCICfsgFmqf1FJXCOMIxS2vOXhHd8,289
39
+ gllm_inference/exceptions/__init__.pyi,sha256=Upcuj7od2lkbdueQ0iMT2ktFYYi-KKTynTLAaxWDTjU,1214
40
+ gllm_inference/exceptions/error_parser.pyi,sha256=IOfa--NpLUW5E9Qq0mwWi6ZpTAbUyyNe6iAqunBNGLI,1999
41
+ gllm_inference/exceptions/exceptions.pyi,sha256=Bv996qLa_vju0Qjf4GewMxdkq8CV9LRZb0S6289DldA,5725
42
+ gllm_inference/exceptions/provider_error_map.pyi,sha256=XPLWU42-r8MHZgg5ZkE80Gdqg3p8Z_JHvq_Na03iTqY,1243
43
+ gllm_inference/lm_invoker/__init__.pyi,sha256=Ze9CxgGYguyz8BAU87_2JM-D4OZjlYAqktLI_B2tj_s,1488
44
+ gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=tsO9WJRj7l4auLc2vxftMX-niWJ9dlYwUIeiTL2FtIo,16392
45
+ gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=6aqQdwVNLO9bDNrqcO8s75IJAN2hLSGvhO6VeDyz44w,14347
46
+ gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=BHv503cfL9d1SjIbxgPXSYjX-ZUrxCbUbnnONRAL0MI,11625
47
+ gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=6DTnHsCFtGkwqVIlvmIZqKQqMVld_JVl-MFaR0WAO8g,8185
48
+ gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=vbnh9PxtEtK740QEgqk6a7sBmtuvOPQ5_1Z3UDSonR4,17338
49
+ gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=NViXl_eGfFGVwLKjJ5OFdV78oay85YLgBZ8QVJGbaLA,12491
50
+ gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=FHAxghha7Jfjjnty0gwgmcWRdEke_tr-54PkKc9BRUU,12404
51
+ gllm_inference/lm_invoker/lm_invoker.pyi,sha256=YfwlV5uJwwoC0uJ54qsF2PZ5eoX85kUWV14nEiBvXTM,8501
52
+ gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=zdHK-YmfplLmvwi1VASm2zvqqyLMoVSMHh2ggqaHIDA,14905
53
+ gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=i5pMpZf4-r_7FQ1qfsqcjpc98sI-cPiqheuTfTEKxJs,4192
54
+ gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=YplNfGRnLAv9puh_9hhCiQ3HbVD3XIZAkyvBR06qCpg,23019
55
+ gllm_inference/lm_invoker/portkey_lm_invoker.pyi,sha256=yI6pN_a5J2uZA3Q9LxniXy6vg8g8ySUzhvpK09PTDgQ,13205
56
+ gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=eovp5rwHykcsoT3-TDo5YU6Kt-azfH_IIIFqmkAsD74,14343
57
+ gllm_inference/lm_invoker/batch/__init__.pyi,sha256=W4W-_yfk7lL20alREJai6GnwuQvdlKRfwQCX4mQK4XI,127
58
+ gllm_inference/lm_invoker/batch/batch_operations.pyi,sha256=Oo7hoyPSfPZdy1mXvSdvtRndvq-XTIbPIjEoGvJj5C0,5372
59
+ gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
60
+ gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=6lreMyHKRfZzX5NBYKnQf1Z6RzXBjTvqZj2VbMeaTLQ,1098
61
+ gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=FJLY-ZkkLUYDV48pfsLatnot4ev_xxz9xAayLK28CpU,1027
62
+ gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=WSuwOqL1j2ZioCZFC-gbB7vTRIZHQ3sU40c3ool5L6c,265
63
+ gllm_inference/lm_invoker/schema/google.pyi,sha256=AIsNgq0ZZuicHmx4bL7z6q-946T05nWts3HUeA8hhHQ,505
64
+ gllm_inference/lm_invoker/schema/langchain.pyi,sha256=rZcIxuvABI4pKfyVvkRBRqfJJogZ67EFPydpubHt49c,429
65
+ gllm_inference/lm_invoker/schema/openai.pyi,sha256=J_rT5Z3rx0hLIae-me1ENeemOESpavcRmYI5pgpkhhk,2222
66
+ gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=8byBRZ4xyTidIQJsZqiSjp5t1X875Obe-aEbT0yYfuA,1199
67
+ gllm_inference/lm_invoker/schema/portkey.pyi,sha256=NeRjHNd84HgE_ur2F3Cv6Jx30v6V7eQvI_iJiq4kuME,631
68
+ gllm_inference/lm_invoker/schema/xai.pyi,sha256=cWnbJmDtllqRH3NXpQbiXgkNBcUXr8ksDSDywcgJebE,632
69
+ gllm_inference/model/__init__.pyi,sha256=LTeBCSJJwCSd5Qrg7RZCXcp9fURNVNXFR5akk1ZZrTk,810
70
+ gllm_inference/model/em/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
71
+ gllm_inference/model/em/cohere_em.pyi,sha256=fArRlV08NwbsJ_h6vpWr94XxUVBtbqW1Jh8s42LRXCo,488
72
+ gllm_inference/model/em/google_em.pyi,sha256=ZPN5LmReO0bcTfnZixFooUTzgD-daNFPzfxzZ-5WzQQ,471
73
+ gllm_inference/model/em/jina_em.pyi,sha256=txEvDI61nhDRUMgvFzpoe-f0onpUAs1j9HPDN01IHxg,627
74
+ gllm_inference/model/em/openai_em.pyi,sha256=KcWpMmxNqS28r4zT4H2TIADHr7e7f3VSI1MPzjJXH9k,442
75
+ gllm_inference/model/em/twelvelabs_em.pyi,sha256=pf9YfTfTPAceBoe1mA5VgtCroHZi5k42mEz-mGSD5QM,400
76
+ gllm_inference/model/em/voyage_em.pyi,sha256=CEfXjLNZamfhsLyAxIkDXND2Jk4GzwXK5puK9yKJDyE,531
77
+ gllm_inference/model/lm/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
78
+ gllm_inference/model/lm/anthropic_lm.pyi,sha256=dWfG-M_gD644yJ-LK_T8HnAT649j3Vx7TVof03XQimE,611
79
+ gllm_inference/model/lm/google_lm.pyi,sha256=cMV5zYX8uwUF7pErv4pXnXD2G52umo3sxKwbSx7nFhQ,511
80
+ gllm_inference/model/lm/openai_lm.pyi,sha256=u11zvvIS7-XaHKZ33cZxGQmT6cZ4DqK9Do8l7gFOUTc,618
81
+ gllm_inference/model/lm/xai_lm.pyi,sha256=2ZEQ_--e_zsb23zZQ8bKdQShU7zChx5TrDKF8EpwEpU,506
82
+ gllm_inference/output_parser/__init__.pyi,sha256=WQOOgsYnPk8vd-SOhFMMaVTzy4gkYrOAyT5gnAxv0A0,129
83
+ gllm_inference/output_parser/json_output_parser.pyi,sha256=uulh91uQLMSb4ZXZhHYi9W9w7zGnmrOweEkL6wdDJN8,2933
84
+ gllm_inference/output_parser/output_parser.pyi,sha256=Yzk7F26pH8Uc7FQZo4G6l67YkfppefUvaV9cNK-HyDs,948
85
+ gllm_inference/prompt_builder/__init__.pyi,sha256=kshfBMvwIwiIvjxiGG5BrJZNvpPa8rhtkbHo5FPifBg,117
86
+ gllm_inference/prompt_builder/prompt_builder.pyi,sha256=VQaw8nE8SPK1nvVVB4YjeqZSazigNPuWvOEcgGji0W8,4557
87
+ gllm_inference/prompt_builder/format_strategy/__init__.pyi,sha256=QhORHac3ySOPmL9k9kmCKL70vtaUtwkZEtGoRNWNuA8,308
88
+ gllm_inference/prompt_builder/format_strategy/format_strategy.pyi,sha256=JSUl_7Ka08oDZPpslymkUa8pDzqNGIK2TlcVANspqrY,2273
89
+ gllm_inference/prompt_builder/format_strategy/jinja_format_strategy.pyi,sha256=IAezLUiKSJMaoDyleo8pFnFqq8rBM_Q-lNXuAGvwXhI,2225
90
+ gllm_inference/prompt_builder/format_strategy/string_format_strategy.pyi,sha256=E0r8x6NTVbPLUfbJBz75PW1n3Ong6bi1WNgxBD0FTM4,693
91
+ gllm_inference/prompt_formatter/__init__.pyi,sha256=rTsjfRsT-y00qH67fPewMNPMN1fAO2y7DM9scR1ccm0,740
92
+ gllm_inference/prompt_formatter/agnostic_prompt_formatter.pyi,sha256=c9mN4t8LXn79h8wq8DAeWYwMgmZGzXjP7EcjLpwfNZg,2018
93
+ gllm_inference/prompt_formatter/huggingface_prompt_formatter.pyi,sha256=AJ-D11HBhTKBA1B5s0km_K1R6o5HD1yvdbGoL7SpGhI,2729
94
+ gllm_inference/prompt_formatter/llama_prompt_formatter.pyi,sha256=nz24trAqHsUY_zHVYfdNzhBO_TCUtzRkg-QYYow4l2s,2848
95
+ gllm_inference/prompt_formatter/mistral_prompt_formatter.pyi,sha256=DgFitxfvCSJf8_6NUfSre-PXYMuPI4vX4AN6BPF1cBE,2650
96
+ gllm_inference/prompt_formatter/openai_prompt_formatter.pyi,sha256=QB7gHddipQuAolUKO01ApZeaRxBtRZzglS5B78mzsV8,1286
97
+ gllm_inference/prompt_formatter/prompt_formatter.pyi,sha256=UkcPi5ao98OGJyNRsqfhYTlMW-ZLNITaGZUTwzvBzqk,1146
98
+ gllm_inference/realtime_chat/__init__.pyi,sha256=Fhxr_sYOt_p43BT2uvJmCp5FMhfb3dDMoIZglq2y6G8,137
99
+ gllm_inference/realtime_chat/google_realtime_chat.pyi,sha256=S5FOCacuc7j4J2tjpUHCdp91YlalxIJ3V3Bh0t0GGus,10112
100
+ gllm_inference/realtime_chat/realtime_chat.pyi,sha256=qG0fXyhTPLw6wGpnDX6wQrLfs6RBaxCK2ASwIqOymp4,1375
101
+ gllm_inference/realtime_chat/input_streamer/__init__.pyi,sha256=REN1dEEZB-I2RX5QsBMxWcUiGm0-fWLevgj7IT18W6Y,317
102
+ gllm_inference/realtime_chat/input_streamer/input_streamer.pyi,sha256=DOVBe_lxievONS_t14opNNuv0zM3MWyBcCwmxzR4_BE,1259
103
+ gllm_inference/realtime_chat/input_streamer/keyboard_input_streamer.pyi,sha256=3L0y91efGmb4YFWigs2WGAURQ48EwNxLgcVMm15yrIg,1043
104
+ gllm_inference/realtime_chat/input_streamer/linux_mic_input_streamer.pyi,sha256=o8rOootmcKHmKRUCCHvfGlCcfeBow5UTbQHIw89XIqg,1337
105
+ gllm_inference/realtime_chat/output_streamer/__init__.pyi,sha256=FP5amnApQcIvW7YfQEnLzJBzEDz51npgocfCcRWDQXA,339
106
+ gllm_inference/realtime_chat/output_streamer/console_output_streamer.pyi,sha256=pO_llZK7JDjllTWEZjujEVo83Hvrv_8LwVqEkQiIOOo,674
107
+ gllm_inference/realtime_chat/output_streamer/linux_speaker_output_streamer.pyi,sha256=yie6VaOyozzE4bKtvWr-6hRiobK-8mshCqeUOhK1PVA,1487
108
+ gllm_inference/realtime_chat/output_streamer/output_streamer.pyi,sha256=GPAw1wPSrwHpJ2MeZJB3V9CCESTDZpyYYgsoy_LI2Nk,1083
109
+ gllm_inference/request_processor/__init__.pyi,sha256=hVnfdNZnkTBJHnmLtN3Na4ANP0yK6AstWdIizVr2Apo,227
110
+ gllm_inference/request_processor/lm_request_processor.pyi,sha256=VnYc8E3Iayyhw-rPnGPfTKuO3ohgFsS8HPrZJeyES5I,5889
111
+ gllm_inference/request_processor/uses_lm_mixin.pyi,sha256=Yu0XPNuHxq1tWBviHTPw1oThojneFwGHepvGjBXxKQA,6382
112
+ gllm_inference/schema/__init__.pyi,sha256=kTFb0oGdne0yMXgx71pmIyG_H07gOLf42mtN5n0Vajs,2266
113
+ gllm_inference/schema/activity.pyi,sha256=JnO2hqj91P5Tc6qb4pbkEMrHer2u5owiCvhl-igcQKQ,2303
114
+ gllm_inference/schema/attachment.pyi,sha256=jApuzjOHJDCz4lr4MlHzBgIndh559nbWu2Xp1fk3hso,3297
115
+ gllm_inference/schema/code_exec_result.pyi,sha256=ZTHh6JtRrPIdQ059P1UAiD2L-tAO1_S5YcMsAXfJ5A0,559
116
+ gllm_inference/schema/config.pyi,sha256=rAL_UeXyQeXVk1P2kqd8vFWOMwmKenfpQLtvMP74t9s,674
117
+ gllm_inference/schema/enums.pyi,sha256=-A7BuMVfUiyOhXdZLqagrdm8MsxRS_HV4PGC6khWMlc,1751
118
+ gllm_inference/schema/events.pyi,sha256=VObVT6B5rvXRYReiVL04vKSxlEHZIjIOVTa4iS87s3w,4705
119
+ gllm_inference/schema/lm_input.pyi,sha256=A5pjz1id6tP9XRNhzQrbmzd66C_q3gzo0UP8rCemz6Q,193
120
+ gllm_inference/schema/lm_output.pyi,sha256=1SZi6vIWvmrZlVQ59WeQUKO5VhKrLHsSRDYslEH9d7o,2435
121
+ gllm_inference/schema/mcp.pyi,sha256=Vwu8E2BDl6FvvnI42gIyY3Oki1BdwRE3Uh3aV0rmhQU,1014
122
+ gllm_inference/schema/message.pyi,sha256=VP9YppKj2mo1esl9cy6qQO9m2mMHUjTmfGDdyUor880,2220
123
+ gllm_inference/schema/model_id.pyi,sha256=BHzi8PdYsV_rPGkBcKvZJDCf62ToPjmRTcXcTgKfoTg,5894
124
+ gllm_inference/schema/reasoning.pyi,sha256=SlTuiDw87GdnAn-I6YOPIJRhEBiwQljM46JohG05guQ,562
125
+ gllm_inference/schema/token_usage.pyi,sha256=1GTQVORV0dBNmD_jix8aVaUqxMKFF04KpLP7y2urqbk,2950
126
+ gllm_inference/schema/tool_call.pyi,sha256=zQaVxCnkVxOfOEhBidqohU85gb4PRwnwBiygKaunamk,389
127
+ gllm_inference/schema/tool_result.pyi,sha256=cAG7TVtB4IWJPt8XBBbB92cuY1ZsX9M276bN9aqjcvM,276
128
+ gllm_inference/schema/type_alias.pyi,sha256=q_C1x8hUUoerGEQFj99g3yaZ1ILigGQSEo-FheqS_lM,529
129
+ gllm_inference/utils/__init__.pyi,sha256=mDJ2gLSeQzm-79Tov5-dhrMNaDqgcD1VVzDYAWvIRqA,391
130
+ gllm_inference/utils/io_utils.pyi,sha256=7kUTacHAVRYoemFUOjCH7-Qmw-YsQGd6rGYxjf_qmtw,1094
131
+ gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
132
+ gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
133
+ gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
134
+ gllm_inference_binary-0.5.51b3.dist-info/METADATA,sha256=3W0iVCB7pQeVBoru96XE5Y-xzRereb8lcDN7PoxknvU,5804
135
+ gllm_inference_binary-0.5.51b3.dist-info/WHEEL,sha256=LySLkZbKo_9Y43IIH4b2lKMJ4X6HWu97Axseguj1GhY,104
136
+ gllm_inference_binary-0.5.51b3.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
137
+ gllm_inference_binary-0.5.51b3.dist-info/RECORD,,
@@ -0,0 +1,5 @@
1
+ Wheel-Version: 1.0
2
+ Generator: Nuitka (2.6.9)
3
+ Root-Is-Purelib: false
4
+ Tag: cp313-cp313-macosx_11_0_arm64
5
+
@@ -0,0 +1 @@
1
+ gllm_inference