gllm-inference-binary 0.5.51b3__cp313-cp313-macosx_11_0_arm64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- gllm_inference/__init__.pyi +0 -0
- gllm_inference/builder/__init__.pyi +6 -0
- gllm_inference/builder/_build_invoker.pyi +28 -0
- gllm_inference/builder/build_em_invoker.pyi +130 -0
- gllm_inference/builder/build_lm_invoker.pyi +213 -0
- gllm_inference/builder/build_lm_request_processor.pyi +88 -0
- gllm_inference/builder/build_output_parser.pyi +29 -0
- gllm_inference/catalog/__init__.pyi +4 -0
- gllm_inference/catalog/catalog.pyi +121 -0
- gllm_inference/catalog/lm_request_processor_catalog.pyi +112 -0
- gllm_inference/catalog/prompt_builder_catalog.pyi +82 -0
- gllm_inference/constants.pyi +13 -0
- gllm_inference/em_invoker/__init__.pyi +12 -0
- gllm_inference/em_invoker/azure_openai_em_invoker.pyi +88 -0
- gllm_inference/em_invoker/bedrock_em_invoker.pyi +118 -0
- gllm_inference/em_invoker/cohere_em_invoker.pyi +128 -0
- gllm_inference/em_invoker/em_invoker.pyi +90 -0
- gllm_inference/em_invoker/google_em_invoker.pyi +129 -0
- gllm_inference/em_invoker/jina_em_invoker.pyi +103 -0
- gllm_inference/em_invoker/langchain/__init__.pyi +3 -0
- gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi +84 -0
- gllm_inference/em_invoker/langchain_em_invoker.pyi +46 -0
- gllm_inference/em_invoker/openai_compatible_em_invoker.pyi +41 -0
- gllm_inference/em_invoker/openai_em_invoker.pyi +118 -0
- gllm_inference/em_invoker/schema/__init__.pyi +0 -0
- gllm_inference/em_invoker/schema/bedrock.pyi +29 -0
- gllm_inference/em_invoker/schema/cohere.pyi +20 -0
- gllm_inference/em_invoker/schema/google.pyi +9 -0
- gllm_inference/em_invoker/schema/jina.pyi +29 -0
- gllm_inference/em_invoker/schema/langchain.pyi +5 -0
- gllm_inference/em_invoker/schema/openai.pyi +7 -0
- gllm_inference/em_invoker/schema/openai_compatible.pyi +7 -0
- gllm_inference/em_invoker/schema/twelvelabs.pyi +17 -0
- gllm_inference/em_invoker/schema/voyage.pyi +15 -0
- gllm_inference/em_invoker/twelevelabs_em_invoker.pyi +101 -0
- gllm_inference/em_invoker/voyage_em_invoker.pyi +104 -0
- gllm_inference/exceptions/__init__.pyi +4 -0
- gllm_inference/exceptions/error_parser.pyi +41 -0
- gllm_inference/exceptions/exceptions.pyi +132 -0
- gllm_inference/exceptions/provider_error_map.pyi +24 -0
- gllm_inference/lm_invoker/__init__.pyi +14 -0
- gllm_inference/lm_invoker/anthropic_lm_invoker.pyi +334 -0
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi +254 -0
- gllm_inference/lm_invoker/batch/__init__.pyi +3 -0
- gllm_inference/lm_invoker/batch/batch_operations.pyi +127 -0
- gllm_inference/lm_invoker/bedrock_lm_invoker.pyi +226 -0
- gllm_inference/lm_invoker/datasaur_lm_invoker.pyi +157 -0
- gllm_inference/lm_invoker/google_lm_invoker.pyi +336 -0
- gllm_inference/lm_invoker/langchain_lm_invoker.pyi +251 -0
- gllm_inference/lm_invoker/litellm_lm_invoker.pyi +244 -0
- gllm_inference/lm_invoker/lm_invoker.pyi +164 -0
- gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi +272 -0
- gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi +52 -0
- gllm_inference/lm_invoker/openai_lm_invoker.pyi +433 -0
- gllm_inference/lm_invoker/portkey_lm_invoker.pyi +260 -0
- gllm_inference/lm_invoker/schema/__init__.pyi +0 -0
- gllm_inference/lm_invoker/schema/anthropic.pyi +56 -0
- gllm_inference/lm_invoker/schema/bedrock.pyi +53 -0
- gllm_inference/lm_invoker/schema/datasaur.pyi +14 -0
- gllm_inference/lm_invoker/schema/google.pyi +24 -0
- gllm_inference/lm_invoker/schema/langchain.pyi +23 -0
- gllm_inference/lm_invoker/schema/openai.pyi +106 -0
- gllm_inference/lm_invoker/schema/openai_chat_completions.pyi +62 -0
- gllm_inference/lm_invoker/schema/portkey.pyi +31 -0
- gllm_inference/lm_invoker/schema/xai.pyi +31 -0
- gllm_inference/lm_invoker/xai_lm_invoker.pyi +289 -0
- gllm_inference/model/__init__.pyi +12 -0
- gllm_inference/model/em/__init__.pyi +0 -0
- gllm_inference/model/em/cohere_em.pyi +17 -0
- gllm_inference/model/em/google_em.pyi +16 -0
- gllm_inference/model/em/jina_em.pyi +22 -0
- gllm_inference/model/em/openai_em.pyi +15 -0
- gllm_inference/model/em/twelvelabs_em.pyi +13 -0
- gllm_inference/model/em/voyage_em.pyi +20 -0
- gllm_inference/model/lm/__init__.pyi +0 -0
- gllm_inference/model/lm/anthropic_lm.pyi +22 -0
- gllm_inference/model/lm/google_lm.pyi +18 -0
- gllm_inference/model/lm/openai_lm.pyi +27 -0
- gllm_inference/model/lm/xai_lm.pyi +19 -0
- gllm_inference/output_parser/__init__.pyi +3 -0
- gllm_inference/output_parser/json_output_parser.pyi +60 -0
- gllm_inference/output_parser/output_parser.pyi +27 -0
- gllm_inference/prompt_builder/__init__.pyi +3 -0
- gllm_inference/prompt_builder/format_strategy/__init__.pyi +4 -0
- gllm_inference/prompt_builder/format_strategy/format_strategy.pyi +55 -0
- gllm_inference/prompt_builder/format_strategy/jinja_format_strategy.pyi +45 -0
- gllm_inference/prompt_builder/format_strategy/string_format_strategy.pyi +20 -0
- gllm_inference/prompt_builder/prompt_builder.pyi +69 -0
- gllm_inference/prompt_formatter/__init__.pyi +7 -0
- gllm_inference/prompt_formatter/agnostic_prompt_formatter.pyi +49 -0
- gllm_inference/prompt_formatter/huggingface_prompt_formatter.pyi +55 -0
- gllm_inference/prompt_formatter/llama_prompt_formatter.pyi +59 -0
- gllm_inference/prompt_formatter/mistral_prompt_formatter.pyi +53 -0
- gllm_inference/prompt_formatter/openai_prompt_formatter.pyi +35 -0
- gllm_inference/prompt_formatter/prompt_formatter.pyi +30 -0
- gllm_inference/realtime_chat/__init__.pyi +3 -0
- gllm_inference/realtime_chat/google_realtime_chat.pyi +205 -0
- gllm_inference/realtime_chat/input_streamer/__init__.pyi +4 -0
- gllm_inference/realtime_chat/input_streamer/input_streamer.pyi +36 -0
- gllm_inference/realtime_chat/input_streamer/keyboard_input_streamer.pyi +27 -0
- gllm_inference/realtime_chat/input_streamer/linux_mic_input_streamer.pyi +36 -0
- gllm_inference/realtime_chat/output_streamer/__init__.pyi +4 -0
- gllm_inference/realtime_chat/output_streamer/console_output_streamer.pyi +21 -0
- gllm_inference/realtime_chat/output_streamer/linux_speaker_output_streamer.pyi +42 -0
- gllm_inference/realtime_chat/output_streamer/output_streamer.pyi +33 -0
- gllm_inference/realtime_chat/realtime_chat.pyi +28 -0
- gllm_inference/request_processor/__init__.pyi +4 -0
- gllm_inference/request_processor/lm_request_processor.pyi +101 -0
- gllm_inference/request_processor/uses_lm_mixin.pyi +130 -0
- gllm_inference/schema/__init__.pyi +18 -0
- gllm_inference/schema/activity.pyi +64 -0
- gllm_inference/schema/attachment.pyi +88 -0
- gllm_inference/schema/code_exec_result.pyi +14 -0
- gllm_inference/schema/config.pyi +15 -0
- gllm_inference/schema/enums.pyi +69 -0
- gllm_inference/schema/events.pyi +105 -0
- gllm_inference/schema/lm_input.pyi +4 -0
- gllm_inference/schema/lm_output.pyi +44 -0
- gllm_inference/schema/mcp.pyi +31 -0
- gllm_inference/schema/message.pyi +52 -0
- gllm_inference/schema/model_id.pyi +176 -0
- gllm_inference/schema/reasoning.pyi +15 -0
- gllm_inference/schema/token_usage.pyi +75 -0
- gllm_inference/schema/tool_call.pyi +14 -0
- gllm_inference/schema/tool_result.pyi +11 -0
- gllm_inference/schema/type_alias.pyi +11 -0
- gllm_inference/utils/__init__.pyi +5 -0
- gllm_inference/utils/io_utils.pyi +26 -0
- gllm_inference/utils/langchain.pyi +30 -0
- gllm_inference/utils/validation.pyi +12 -0
- gllm_inference.build/.gitignore +1 -0
- gllm_inference.cpython-313-darwin.so +0 -0
- gllm_inference.pyi +157 -0
- gllm_inference_binary-0.5.51b3.dist-info/METADATA +138 -0
- gllm_inference_binary-0.5.51b3.dist-info/RECORD +137 -0
- gllm_inference_binary-0.5.51b3.dist-info/WHEEL +5 -0
- gllm_inference_binary-0.5.51b3.dist-info/top_level.txt +1 -0
|
@@ -0,0 +1,30 @@
|
|
|
1
|
+
from _typeshed import Incomplete
|
|
2
|
+
from langchain_core.embeddings import Embeddings as Embeddings
|
|
3
|
+
from langchain_core.language_models import BaseChatModel as BaseChatModel
|
|
4
|
+
from typing import Any
|
|
5
|
+
|
|
6
|
+
MODEL_NAME_KEYS: Incomplete
|
|
7
|
+
|
|
8
|
+
def load_langchain_model(model_class_path: str, model_name: str, model_kwargs: dict[str, Any]) -> BaseChatModel | Embeddings:
|
|
9
|
+
'''Loads the LangChain\'s model instance.
|
|
10
|
+
|
|
11
|
+
Args:
|
|
12
|
+
model_class_path (str): The path to the LangChain\'s class, e.g. "langchain_openai.ChatOpenAI".
|
|
13
|
+
model_name (str): The model name.
|
|
14
|
+
model_kwargs (dict[str, Any]): The additional keyword arguments.
|
|
15
|
+
|
|
16
|
+
Returns:
|
|
17
|
+
BaseChatModel | Embeddings: The LangChain\'s model instance.
|
|
18
|
+
'''
|
|
19
|
+
def parse_model_data(model: BaseChatModel | Embeddings) -> dict[str, str]:
|
|
20
|
+
"""Parses the model data from LangChain's BaseChatModel or Embeddings instance.
|
|
21
|
+
|
|
22
|
+
Args:
|
|
23
|
+
model (BaseChatModel | Embeddings): The LangChain's BaseChatModel or Embeddings instance.
|
|
24
|
+
|
|
25
|
+
Returns:
|
|
26
|
+
dict[str, str]: The dictionary containing the model name and path.
|
|
27
|
+
|
|
28
|
+
Raises:
|
|
29
|
+
ValueError: If the model name is not found in the model data.
|
|
30
|
+
"""
|
|
@@ -0,0 +1,12 @@
|
|
|
1
|
+
from enum import StrEnum
|
|
2
|
+
|
|
3
|
+
def validate_string_enum(enum_type: type[StrEnum], value: str) -> None:
|
|
4
|
+
"""Validates that the provided value is a valid string enum value.
|
|
5
|
+
|
|
6
|
+
Args:
|
|
7
|
+
enum_type (type[StrEnum]): The type of the string enum.
|
|
8
|
+
value (str): The value to validate.
|
|
9
|
+
|
|
10
|
+
Raises:
|
|
11
|
+
ValueError: If the provided value is not a valid string enum value.
|
|
12
|
+
"""
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
*
|
|
Binary file
|
gllm_inference.pyi
ADDED
|
@@ -0,0 +1,157 @@
|
|
|
1
|
+
# This file was generated by Nuitka
|
|
2
|
+
|
|
3
|
+
# Stubs included by default
|
|
4
|
+
|
|
5
|
+
|
|
6
|
+
__name__ = ...
|
|
7
|
+
|
|
8
|
+
|
|
9
|
+
|
|
10
|
+
# Modules used internally, to allow implicit dependencies to be seen:
|
|
11
|
+
import os
|
|
12
|
+
import typing
|
|
13
|
+
import gllm_core
|
|
14
|
+
import gllm_core.utils
|
|
15
|
+
import gllm_inference.em_invoker.AzureOpenAIEMInvoker
|
|
16
|
+
import gllm_inference.em_invoker.BedrockEMInvoker
|
|
17
|
+
import gllm_inference.em_invoker.CohereEMInvoker
|
|
18
|
+
import gllm_inference.em_invoker.GoogleEMInvoker
|
|
19
|
+
import gllm_inference.em_invoker.JinaEMInvoker
|
|
20
|
+
import gllm_inference.em_invoker.LangChainEMInvoker
|
|
21
|
+
import gllm_inference.em_invoker.OpenAICompatibleEMInvoker
|
|
22
|
+
import gllm_inference.em_invoker.OpenAIEMInvoker
|
|
23
|
+
import gllm_inference.em_invoker.TwelveLabsEMInvoker
|
|
24
|
+
import gllm_inference.em_invoker.VoyageEMInvoker
|
|
25
|
+
import gllm_inference.lm_invoker.AnthropicLMInvoker
|
|
26
|
+
import gllm_inference.lm_invoker.AzureOpenAILMInvoker
|
|
27
|
+
import gllm_inference.lm_invoker.BedrockLMInvoker
|
|
28
|
+
import gllm_inference.lm_invoker.DatasaurLMInvoker
|
|
29
|
+
import gllm_inference.lm_invoker.GoogleLMInvoker
|
|
30
|
+
import gllm_inference.lm_invoker.LangChainLMInvoker
|
|
31
|
+
import gllm_inference.lm_invoker.LiteLLMLMInvoker
|
|
32
|
+
import gllm_inference.lm_invoker.OpenAIChatCompletionsLMInvoker
|
|
33
|
+
import gllm_inference.lm_invoker.OpenAICompatibleLMInvoker
|
|
34
|
+
import gllm_inference.lm_invoker.OpenAILMInvoker
|
|
35
|
+
import gllm_inference.lm_invoker.PortkeyLMInvoker
|
|
36
|
+
import gllm_inference.lm_invoker.XAILMInvoker
|
|
37
|
+
import gllm_inference.prompt_builder.PromptBuilder
|
|
38
|
+
import gllm_inference.output_parser.JSONOutputParser
|
|
39
|
+
import json
|
|
40
|
+
import abc
|
|
41
|
+
import pandas
|
|
42
|
+
import pydantic
|
|
43
|
+
import re
|
|
44
|
+
import gllm_core.utils.retry
|
|
45
|
+
import gllm_inference.request_processor.LMRequestProcessor
|
|
46
|
+
import gllm_core.utils.imports
|
|
47
|
+
import gllm_inference.schema.ModelId
|
|
48
|
+
import gllm_inference.schema.ModelProvider
|
|
49
|
+
import gllm_inference.schema.TruncationConfig
|
|
50
|
+
import asyncio
|
|
51
|
+
import base64
|
|
52
|
+
import enum
|
|
53
|
+
import gllm_inference.exceptions.BaseInvokerError
|
|
54
|
+
import gllm_inference.exceptions.convert_http_status_to_base_invoker_error
|
|
55
|
+
import gllm_inference.schema.Attachment
|
|
56
|
+
import gllm_inference.schema.AttachmentType
|
|
57
|
+
import gllm_inference.schema.EMContent
|
|
58
|
+
import gllm_inference.schema.Vector
|
|
59
|
+
import aioboto3
|
|
60
|
+
import gllm_inference.utils.validate_string_enum
|
|
61
|
+
import cohere
|
|
62
|
+
import asyncio.CancelledError
|
|
63
|
+
import gllm_inference.exceptions.convert_to_base_invoker_error
|
|
64
|
+
import gllm_inference.schema.TruncateSide
|
|
65
|
+
import google
|
|
66
|
+
import google.auth
|
|
67
|
+
import google.genai
|
|
68
|
+
import google.genai.types
|
|
69
|
+
import httpx
|
|
70
|
+
import gllm_inference.exceptions.ProviderInternalError
|
|
71
|
+
import concurrent
|
|
72
|
+
import concurrent.futures
|
|
73
|
+
import concurrent.futures.ThreadPoolExecutor
|
|
74
|
+
import langchain_core
|
|
75
|
+
import langchain_core.embeddings
|
|
76
|
+
import gllm_inference.exceptions.InvokerRuntimeError
|
|
77
|
+
import gllm_inference.exceptions.build_debug_info
|
|
78
|
+
import gllm_inference.utils.load_langchain_model
|
|
79
|
+
import gllm_inference.utils.parse_model_data
|
|
80
|
+
import openai
|
|
81
|
+
import io
|
|
82
|
+
import twelvelabs
|
|
83
|
+
import sys
|
|
84
|
+
import voyageai
|
|
85
|
+
import voyageai.client_async
|
|
86
|
+
import http
|
|
87
|
+
import http.HTTPStatus
|
|
88
|
+
import uuid
|
|
89
|
+
import gllm_core.constants
|
|
90
|
+
import gllm_core.event
|
|
91
|
+
import gllm_core.schema
|
|
92
|
+
import gllm_core.schema.tool
|
|
93
|
+
import langchain_core.tools
|
|
94
|
+
import gllm_inference.schema.BatchStatus
|
|
95
|
+
import gllm_inference.schema.LMInput
|
|
96
|
+
import gllm_inference.schema.LMOutput
|
|
97
|
+
import gllm_inference.schema.Message
|
|
98
|
+
import gllm_inference.schema.Reasoning
|
|
99
|
+
import gllm_inference.schema.ResponseSchema
|
|
100
|
+
import gllm_inference.schema.ThinkingEvent
|
|
101
|
+
import gllm_inference.schema.TokenUsage
|
|
102
|
+
import gllm_inference.schema.ToolCall
|
|
103
|
+
import gllm_inference.schema.ToolResult
|
|
104
|
+
import anthropic
|
|
105
|
+
import anthropic.types
|
|
106
|
+
import anthropic.types.message_create_params
|
|
107
|
+
import anthropic.types.messages
|
|
108
|
+
import anthropic.types.messages.batch_create_params
|
|
109
|
+
import gllm_inference.schema.MessageRole
|
|
110
|
+
import langchain_core.language_models
|
|
111
|
+
import langchain_core.messages
|
|
112
|
+
import gllm_inference.exceptions._get_exception_key
|
|
113
|
+
import litellm
|
|
114
|
+
import inspect
|
|
115
|
+
import time
|
|
116
|
+
import jsonschema
|
|
117
|
+
import gllm_inference.lm_invoker.batch.BatchOperations
|
|
118
|
+
import gllm_inference.schema.LMEventType
|
|
119
|
+
import gllm_inference.schema.MessageContent
|
|
120
|
+
import __future__
|
|
121
|
+
import gllm_inference.schema.ActivityEvent
|
|
122
|
+
import gllm_inference.schema.CodeEvent
|
|
123
|
+
import gllm_inference.schema.CodeExecResult
|
|
124
|
+
import gllm_inference.schema.MCPCall
|
|
125
|
+
import gllm_inference.schema.MCPCallActivity
|
|
126
|
+
import gllm_inference.schema.MCPListToolsActivity
|
|
127
|
+
import gllm_inference.schema.MCPServer
|
|
128
|
+
import gllm_inference.schema.WebSearchActivity
|
|
129
|
+
import logging
|
|
130
|
+
import portkey_ai
|
|
131
|
+
import xai_sdk
|
|
132
|
+
import xai_sdk.chat
|
|
133
|
+
import xai_sdk.search
|
|
134
|
+
import xai_sdk.proto
|
|
135
|
+
import xai_sdk.proto.v5
|
|
136
|
+
import xai_sdk.proto.v5.chat_pb2
|
|
137
|
+
import jinja2
|
|
138
|
+
import jinja2.sandbox
|
|
139
|
+
import gllm_inference.schema.JinjaEnvType
|
|
140
|
+
import gllm_inference.prompt_builder.format_strategy.JinjaFormatStrategy
|
|
141
|
+
import gllm_inference.prompt_builder.format_strategy.StringFormatStrategy
|
|
142
|
+
import transformers
|
|
143
|
+
import gllm_inference.prompt_formatter.HuggingFacePromptFormatter
|
|
144
|
+
import traceback
|
|
145
|
+
import gllm_inference.realtime_chat.input_streamer.KeyboardInputStreamer
|
|
146
|
+
import gllm_inference.realtime_chat.output_streamer.ConsoleOutputStreamer
|
|
147
|
+
import google.genai.live
|
|
148
|
+
import gllm_core.utils.logger_manager
|
|
149
|
+
import mimetypes
|
|
150
|
+
import pathlib
|
|
151
|
+
import pathlib.Path
|
|
152
|
+
import filetype
|
|
153
|
+
import magic
|
|
154
|
+
import requests
|
|
155
|
+
import binascii
|
|
156
|
+
import fnmatch
|
|
157
|
+
import importlib
|
|
@@ -0,0 +1,138 @@
|
|
|
1
|
+
Metadata-Version: 2.2
|
|
2
|
+
Name: gllm-inference-binary
|
|
3
|
+
Version: 0.5.51b3
|
|
4
|
+
Summary: A library containing components related to model inferences in Gen AI applications.
|
|
5
|
+
Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, Resti Febrina <resti.febrina@gdplabs.id>
|
|
6
|
+
Requires-Python: <3.14,>=3.11
|
|
7
|
+
Description-Content-Type: text/markdown
|
|
8
|
+
Requires-Dist: poetry<3.0.0,>=2.1.3
|
|
9
|
+
Requires-Dist: gllm-core-binary==0.3.23b2
|
|
10
|
+
Requires-Dist: aiohttp<4.0.0,>=3.12.14
|
|
11
|
+
Requires-Dist: filetype<2.0.0,>=1.2.0
|
|
12
|
+
Requires-Dist: httpx<0.29.0,>=0.28.0
|
|
13
|
+
Requires-Dist: jinja2<4.0.0,>=3.1.4
|
|
14
|
+
Requires-Dist: jsonschema<5.0.0,>=4.24.0
|
|
15
|
+
Requires-Dist: langchain<0.4.0,>=0.3.0
|
|
16
|
+
Requires-Dist: pandas<3.0.0,>=2.2.3
|
|
17
|
+
Requires-Dist: protobuf<7.0.0,>=5.29.4
|
|
18
|
+
Requires-Dist: python-magic<0.5.0,>=0.4.27; sys_platform != "win32"
|
|
19
|
+
Requires-Dist: python-magic-bin<0.5.0,>=0.4.14; sys_platform == "win32"
|
|
20
|
+
Requires-Dist: sentencepiece<0.3.0,>=0.2.0
|
|
21
|
+
Provides-Extra: dev
|
|
22
|
+
Requires-Dist: coverage<8.0.0,>=7.4.4; extra == "dev"
|
|
23
|
+
Requires-Dist: mypy<2.0.0,>=1.15.0; extra == "dev"
|
|
24
|
+
Requires-Dist: pre-commit<4.0.0,>=3.7.0; extra == "dev"
|
|
25
|
+
Requires-Dist: pytest<9.0.0,>=8.1.1; extra == "dev"
|
|
26
|
+
Requires-Dist: pytest-asyncio<0.24.0,>=0.23.6; extra == "dev"
|
|
27
|
+
Requires-Dist: pytest-cov<6.0.0,>=5.0.0; extra == "dev"
|
|
28
|
+
Requires-Dist: ruff<0.7.0,>=0.6.7; extra == "dev"
|
|
29
|
+
Provides-Extra: anthropic
|
|
30
|
+
Requires-Dist: anthropic<0.61.0,>=0.60.0; extra == "anthropic"
|
|
31
|
+
Provides-Extra: bedrock
|
|
32
|
+
Requires-Dist: aioboto3<16.0.0,>=15.0.0; extra == "bedrock"
|
|
33
|
+
Provides-Extra: cohere
|
|
34
|
+
Requires-Dist: cohere<6.0.0,>=5.18.0; extra == "cohere"
|
|
35
|
+
Provides-Extra: datasaur
|
|
36
|
+
Requires-Dist: openai<2.0.0,>=1.98.0; extra == "datasaur"
|
|
37
|
+
Provides-Extra: google
|
|
38
|
+
Requires-Dist: google-genai<=1.36,>=1.23; extra == "google"
|
|
39
|
+
Provides-Extra: huggingface
|
|
40
|
+
Requires-Dist: huggingface-hub<0.31.0,>=0.30.0; extra == "huggingface"
|
|
41
|
+
Requires-Dist: transformers<5.0.0,>=4.52.0; extra == "huggingface"
|
|
42
|
+
Provides-Extra: litellm
|
|
43
|
+
Requires-Dist: litellm<2.0.0,>=1.69.2; extra == "litellm"
|
|
44
|
+
Provides-Extra: openai
|
|
45
|
+
Requires-Dist: openai<2.0.0,>=1.98.0; extra == "openai"
|
|
46
|
+
Provides-Extra: portkey-ai
|
|
47
|
+
Requires-Dist: portkey-ai<2.0.0,>=1.14.4; extra == "portkey-ai"
|
|
48
|
+
Provides-Extra: twelvelabs
|
|
49
|
+
Requires-Dist: twelvelabs<0.5.0,>=0.4.4; extra == "twelvelabs"
|
|
50
|
+
Provides-Extra: voyage
|
|
51
|
+
Requires-Dist: voyageai<0.4.0,>=0.3.0; python_version < "3.13" and extra == "voyage"
|
|
52
|
+
Provides-Extra: xai
|
|
53
|
+
Requires-Dist: xai_sdk<2.0.0,>=1.0.0; extra == "xai"
|
|
54
|
+
|
|
55
|
+
# GLLM Inference
|
|
56
|
+
|
|
57
|
+
## Description
|
|
58
|
+
|
|
59
|
+
A library containing components related to model inferences in Gen AI applications.
|
|
60
|
+
|
|
61
|
+
## Installation
|
|
62
|
+
|
|
63
|
+
### Prerequisites
|
|
64
|
+
1. Python 3.11+ - [Install here](https://www.python.org/downloads/)
|
|
65
|
+
2. Pip (if using Pip) - [Install here](https://pip.pypa.io/en/stable/installation/)
|
|
66
|
+
3. Poetry (automatically installed via Makefile) - [Install here](https://python-poetry.org/docs/#installation)
|
|
67
|
+
4. Git (if using Git) - [Install here](https://git-scm.com/downloads)
|
|
68
|
+
5. gcloud CLI (for authentication) - [Install here](https://cloud.google.com/sdk/docs/install)
|
|
69
|
+
6. For git installation, access to the [GDP Labs SDK github repository](https://github.com/GDP-ADMIN/gl-sdk)
|
|
70
|
+
|
|
71
|
+
### 1. Installation from Artifact Registry
|
|
72
|
+
Choose one of the following methods to install the package:
|
|
73
|
+
|
|
74
|
+
#### Using pip
|
|
75
|
+
```bash
|
|
76
|
+
pip install gllm-inference-binary
|
|
77
|
+
```
|
|
78
|
+
|
|
79
|
+
#### Using Poetry
|
|
80
|
+
```bash
|
|
81
|
+
poetry add gllm-inference-binary
|
|
82
|
+
```
|
|
83
|
+
|
|
84
|
+
### 2. Development Installation (Git)
|
|
85
|
+
For development purposes, you can install directly from the Git repository:
|
|
86
|
+
```bash
|
|
87
|
+
poetry add "git+ssh://git@github.com/GDP-ADMIN/gen-ai-internal.git#subdirectory=libs/gllm-inference"
|
|
88
|
+
```
|
|
89
|
+
|
|
90
|
+
Available extras:
|
|
91
|
+
- `anthropic`: Install Anthropic models dependencies
|
|
92
|
+
- `google-genai`: Install Google Generative AI models dependencies
|
|
93
|
+
- `google-vertexai`: Install Google Vertex AI models dependencies
|
|
94
|
+
- `huggingface`: Install HuggingFace models dependencies
|
|
95
|
+
- `openai`: Install OpenAI models dependencies
|
|
96
|
+
- `twelvelabs`: Install TwelveLabs models dependencies
|
|
97
|
+
|
|
98
|
+
## Local Development Setup
|
|
99
|
+
|
|
100
|
+
### Quick Setup (Recommended)
|
|
101
|
+
For local development with editable gllm packages, use the provided Makefile:
|
|
102
|
+
|
|
103
|
+
```bash
|
|
104
|
+
# Complete setup: installs Poetry, configures auth, installs packages, sets up pre-commit
|
|
105
|
+
make setup
|
|
106
|
+
```
|
|
107
|
+
|
|
108
|
+
The following are the available Makefile targets:
|
|
109
|
+
|
|
110
|
+
1. `make setup` - Complete development setup (recommended for new developers)
|
|
111
|
+
2. `make install-poetry` - Install or upgrade Poetry to the latest version
|
|
112
|
+
3. `make auth` - Configure authentication for internal repositories
|
|
113
|
+
4. `make install` - Install all dependencies
|
|
114
|
+
5. `make install-pre-commit` - Set up pre-commit hooks
|
|
115
|
+
6. `make update` - Update dependencies
|
|
116
|
+
### Manual Development Setup (Legacy)
|
|
117
|
+
If you prefer to manage dependencies manually:
|
|
118
|
+
|
|
119
|
+
1. Go to root folder of `gllm-inference` module, e.g. `cd libs/gllm-inference`.
|
|
120
|
+
2. Run `poetry shell` to create a virtual environment.
|
|
121
|
+
3. Run `poetry lock` to create a lock file if you haven't done it yet.
|
|
122
|
+
4. Run `poetry install` to install the `gllm-inference` requirements for the first time.
|
|
123
|
+
5. Run `poetry update` if you update any dependency module version at `pyproject.toml`.
|
|
124
|
+
|
|
125
|
+
|
|
126
|
+
## Contributing
|
|
127
|
+
Please refer to this [Python Style Guide](https://docs.google.com/document/d/1uRggCrHnVfDPBnG641FyQBwUwLoFw0kTzNqRm92vUwM/edit?usp=sharing)
|
|
128
|
+
to get information about code style, documentation standard, and SCA that you need to use when contributing to this project
|
|
129
|
+
|
|
130
|
+
### Getting Started with Development
|
|
131
|
+
1. Clone the repository and navigate to the gllm-inference directory
|
|
132
|
+
2. Run `make setup` to set up your development environment
|
|
133
|
+
3. Run `which python` to get the path to be referenced at Visual Studio Code interpreter path (`Ctrl`+`Shift`+`P` or `Cmd`+`Shift`+`P`)
|
|
134
|
+
4. Try running the unit test to see if it's working:
|
|
135
|
+
```bash
|
|
136
|
+
poetry run pytest -s tests/unit_tests/
|
|
137
|
+
```
|
|
138
|
+
5. When you want to update the dependencies, run `make update`
|
|
@@ -0,0 +1,137 @@
|
|
|
1
|
+
gllm_inference.cpython-313-darwin.so,sha256=e0m_8i2W9bQrHKjotHdO9i17dUgPRrd4ZaXup8BzYNY,4945704
|
|
2
|
+
gllm_inference.pyi,sha256=3TbylLc0CkZzlLT7WqOloWNruBH0tojkkQJ_krV9mQs,5211
|
|
3
|
+
gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
4
|
+
gllm_inference/constants.pyi,sha256=viU-ACRbVSGvsCJ0FQmuR1yhyl-BzoHDVIWo5cwHmF0,337
|
|
5
|
+
gllm_inference/builder/__init__.pyi,sha256=usz2lvfwO4Yk-ZGKXbCWG1cEr3nlQXxMNDNC-2yc1NM,500
|
|
6
|
+
gllm_inference/builder/_build_invoker.pyi,sha256=v__-YT0jGmLqEsgl7Abk9we-wYWcyVFvlRN_Uu4vVak,848
|
|
7
|
+
gllm_inference/builder/build_em_invoker.pyi,sha256=0IVcRGaciPBAFcgvGRC4-Kje_PPXOSug77LqxRc-x_U,5749
|
|
8
|
+
gllm_inference/builder/build_lm_invoker.pyi,sha256=TBAr7Sk9Jgckdlfj69pYOzZFMsooL-FpfbPYt_kCcXU,8852
|
|
9
|
+
gllm_inference/builder/build_lm_request_processor.pyi,sha256=KbQkcPa8C-yzyelht4mWLP8kDmh17itAT3tn8ZJB6pg,4144
|
|
10
|
+
gllm_inference/builder/build_output_parser.pyi,sha256=_Lrq-bh1oPsb_Nwkkr_zyEUwIOMysRFZkvEtEM29LZM,936
|
|
11
|
+
gllm_inference/catalog/__init__.pyi,sha256=JBkPGTyiiZ30GECzJBW-mW8LekWyY2qyzal3eW7ynaM,287
|
|
12
|
+
gllm_inference/catalog/catalog.pyi,sha256=a4RNG1lKv51GxQpOqh47tz-PAROMPaeP2o5XNLBSZaU,4790
|
|
13
|
+
gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=WW1j8jWujnatF0c9rCk94CyobtFe3gOky2vjoGCV6nw,5424
|
|
14
|
+
gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=OU8k_4HbqjZEzHZlzSM3uzGQZJmM2uGD76Csqom0CEQ,3197
|
|
15
|
+
gllm_inference/em_invoker/__init__.pyi,sha256=FHFR4k8RqtdMNZkda6A89tvWFtIoxF-t_NahDGeBQ2k,1199
|
|
16
|
+
gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=fE-RZ10Leerbyqp8ISXeqdU_m0kaHdlHml_nQqwFjqY,4961
|
|
17
|
+
gllm_inference/em_invoker/bedrock_em_invoker.pyi,sha256=3GkVO9hvmYck4NJY5spAg6opSxr0v8OJ4qhMvz1rcAA,6365
|
|
18
|
+
gllm_inference/em_invoker/cohere_em_invoker.pyi,sha256=LSNVGx58JdSP2YB6CziOS_v5VuSLdARVpBBWhz8t5E4,6665
|
|
19
|
+
gllm_inference/em_invoker/em_invoker.pyi,sha256=KGjLiAWGIA3ziV50zMwSzx6lTVUbYspQCl4LFxqbDlY,5101
|
|
20
|
+
gllm_inference/em_invoker/google_em_invoker.pyi,sha256=OgIQhT2g6g09Hqr9mc2--MjV3Ti-XuHfqLTunvcqU1k,6824
|
|
21
|
+
gllm_inference/em_invoker/jina_em_invoker.pyi,sha256=WZJ8BeaqxrAGMrUtIEUy3itcOEzvRQYEwe0gazIv6ng,5651
|
|
22
|
+
gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=BhZjkYZoxQXPJjP0PgA8W0xrtwiqDkW5E6NpGit0h1E,3498
|
|
23
|
+
gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=z4_jEuaprna6AJF2wXH-JgmynpVFXeb2HQZr-0ATmMw,2882
|
|
24
|
+
gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=tb6cXTvAGr01MMSc96mYWDqK-EK_VCdDOQ-Be0_EZfM,6195
|
|
25
|
+
gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=mitrJZW3M43k306cHwJoDjBfN1nr20C4sZ42Xr6hfO0,5347
|
|
26
|
+
gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=gyYQA3rbekHJIvvY6TE4tTt8W-DbPqjiPW6FaJoQPHI,5517
|
|
27
|
+
gllm_inference/em_invoker/langchain/__init__.pyi,sha256=VYGKE5OgU0my1RlhgzkU_A7-GLGnUDDnNFuctuRwILE,148
|
|
28
|
+
gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi,sha256=VU3-Vhb9BCDhJo8PPdWHe2rBEOCs_HMXT6ZaWwjUzZE,3304
|
|
29
|
+
gllm_inference/em_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
30
|
+
gllm_inference/em_invoker/schema/bedrock.pyi,sha256=ePYwailnZAYbZKt9HRDi4iAISq003jpxQPxDCiFdEZg,615
|
|
31
|
+
gllm_inference/em_invoker/schema/cohere.pyi,sha256=UTbTtePRR1zJMsM09SiTZSZZP0IaUGaODvc7ZqH9S8c,547
|
|
32
|
+
gllm_inference/em_invoker/schema/google.pyi,sha256=ovDlvinu99QJhIxMkvVUoGBEFkkEoAZhadSuk0nI9N8,181
|
|
33
|
+
gllm_inference/em_invoker/schema/jina.pyi,sha256=vE1ySd8OTDM35saEZos7UCdPwHeX66iuHkZ3RchSZKA,741
|
|
34
|
+
gllm_inference/em_invoker/schema/langchain.pyi,sha256=edcUvc1IHoSMFwqV83uqWqd0U3fLhkyWQjVknvjHI8U,112
|
|
35
|
+
gllm_inference/em_invoker/schema/openai.pyi,sha256=Q_dsEcodkOXYXPdrkOkW0LnuLhfeq8tEbtZAGMz2ajA,139
|
|
36
|
+
gllm_inference/em_invoker/schema/openai_compatible.pyi,sha256=gmvGtsWoOMBelke_tZjC6dKimFBW9f4Vrgv0Ig0OM9Q,150
|
|
37
|
+
gllm_inference/em_invoker/schema/twelvelabs.pyi,sha256=F6wKHgG01bYskJpKoheBSpRpHUfFpteKn9sj9n5YfcU,372
|
|
38
|
+
gllm_inference/em_invoker/schema/voyage.pyi,sha256=HVpor0fqNy-IwapCICfsgFmqf1FJXCOMIxS2vOXhHd8,289
|
|
39
|
+
gllm_inference/exceptions/__init__.pyi,sha256=Upcuj7od2lkbdueQ0iMT2ktFYYi-KKTynTLAaxWDTjU,1214
|
|
40
|
+
gllm_inference/exceptions/error_parser.pyi,sha256=IOfa--NpLUW5E9Qq0mwWi6ZpTAbUyyNe6iAqunBNGLI,1999
|
|
41
|
+
gllm_inference/exceptions/exceptions.pyi,sha256=Bv996qLa_vju0Qjf4GewMxdkq8CV9LRZb0S6289DldA,5725
|
|
42
|
+
gllm_inference/exceptions/provider_error_map.pyi,sha256=XPLWU42-r8MHZgg5ZkE80Gdqg3p8Z_JHvq_Na03iTqY,1243
|
|
43
|
+
gllm_inference/lm_invoker/__init__.pyi,sha256=Ze9CxgGYguyz8BAU87_2JM-D4OZjlYAqktLI_B2tj_s,1488
|
|
44
|
+
gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=tsO9WJRj7l4auLc2vxftMX-niWJ9dlYwUIeiTL2FtIo,16392
|
|
45
|
+
gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=6aqQdwVNLO9bDNrqcO8s75IJAN2hLSGvhO6VeDyz44w,14347
|
|
46
|
+
gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=BHv503cfL9d1SjIbxgPXSYjX-ZUrxCbUbnnONRAL0MI,11625
|
|
47
|
+
gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=6DTnHsCFtGkwqVIlvmIZqKQqMVld_JVl-MFaR0WAO8g,8185
|
|
48
|
+
gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=vbnh9PxtEtK740QEgqk6a7sBmtuvOPQ5_1Z3UDSonR4,17338
|
|
49
|
+
gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=NViXl_eGfFGVwLKjJ5OFdV78oay85YLgBZ8QVJGbaLA,12491
|
|
50
|
+
gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=FHAxghha7Jfjjnty0gwgmcWRdEke_tr-54PkKc9BRUU,12404
|
|
51
|
+
gllm_inference/lm_invoker/lm_invoker.pyi,sha256=YfwlV5uJwwoC0uJ54qsF2PZ5eoX85kUWV14nEiBvXTM,8501
|
|
52
|
+
gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=zdHK-YmfplLmvwi1VASm2zvqqyLMoVSMHh2ggqaHIDA,14905
|
|
53
|
+
gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=i5pMpZf4-r_7FQ1qfsqcjpc98sI-cPiqheuTfTEKxJs,4192
|
|
54
|
+
gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=YplNfGRnLAv9puh_9hhCiQ3HbVD3XIZAkyvBR06qCpg,23019
|
|
55
|
+
gllm_inference/lm_invoker/portkey_lm_invoker.pyi,sha256=yI6pN_a5J2uZA3Q9LxniXy6vg8g8ySUzhvpK09PTDgQ,13205
|
|
56
|
+
gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=eovp5rwHykcsoT3-TDo5YU6Kt-azfH_IIIFqmkAsD74,14343
|
|
57
|
+
gllm_inference/lm_invoker/batch/__init__.pyi,sha256=W4W-_yfk7lL20alREJai6GnwuQvdlKRfwQCX4mQK4XI,127
|
|
58
|
+
gllm_inference/lm_invoker/batch/batch_operations.pyi,sha256=Oo7hoyPSfPZdy1mXvSdvtRndvq-XTIbPIjEoGvJj5C0,5372
|
|
59
|
+
gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
60
|
+
gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=6lreMyHKRfZzX5NBYKnQf1Z6RzXBjTvqZj2VbMeaTLQ,1098
|
|
61
|
+
gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=FJLY-ZkkLUYDV48pfsLatnot4ev_xxz9xAayLK28CpU,1027
|
|
62
|
+
gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=WSuwOqL1j2ZioCZFC-gbB7vTRIZHQ3sU40c3ool5L6c,265
|
|
63
|
+
gllm_inference/lm_invoker/schema/google.pyi,sha256=AIsNgq0ZZuicHmx4bL7z6q-946T05nWts3HUeA8hhHQ,505
|
|
64
|
+
gllm_inference/lm_invoker/schema/langchain.pyi,sha256=rZcIxuvABI4pKfyVvkRBRqfJJogZ67EFPydpubHt49c,429
|
|
65
|
+
gllm_inference/lm_invoker/schema/openai.pyi,sha256=J_rT5Z3rx0hLIae-me1ENeemOESpavcRmYI5pgpkhhk,2222
|
|
66
|
+
gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=8byBRZ4xyTidIQJsZqiSjp5t1X875Obe-aEbT0yYfuA,1199
|
|
67
|
+
gllm_inference/lm_invoker/schema/portkey.pyi,sha256=NeRjHNd84HgE_ur2F3Cv6Jx30v6V7eQvI_iJiq4kuME,631
|
|
68
|
+
gllm_inference/lm_invoker/schema/xai.pyi,sha256=cWnbJmDtllqRH3NXpQbiXgkNBcUXr8ksDSDywcgJebE,632
|
|
69
|
+
gllm_inference/model/__init__.pyi,sha256=LTeBCSJJwCSd5Qrg7RZCXcp9fURNVNXFR5akk1ZZrTk,810
|
|
70
|
+
gllm_inference/model/em/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
71
|
+
gllm_inference/model/em/cohere_em.pyi,sha256=fArRlV08NwbsJ_h6vpWr94XxUVBtbqW1Jh8s42LRXCo,488
|
|
72
|
+
gllm_inference/model/em/google_em.pyi,sha256=ZPN5LmReO0bcTfnZixFooUTzgD-daNFPzfxzZ-5WzQQ,471
|
|
73
|
+
gllm_inference/model/em/jina_em.pyi,sha256=txEvDI61nhDRUMgvFzpoe-f0onpUAs1j9HPDN01IHxg,627
|
|
74
|
+
gllm_inference/model/em/openai_em.pyi,sha256=KcWpMmxNqS28r4zT4H2TIADHr7e7f3VSI1MPzjJXH9k,442
|
|
75
|
+
gllm_inference/model/em/twelvelabs_em.pyi,sha256=pf9YfTfTPAceBoe1mA5VgtCroHZi5k42mEz-mGSD5QM,400
|
|
76
|
+
gllm_inference/model/em/voyage_em.pyi,sha256=CEfXjLNZamfhsLyAxIkDXND2Jk4GzwXK5puK9yKJDyE,531
|
|
77
|
+
gllm_inference/model/lm/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
78
|
+
gllm_inference/model/lm/anthropic_lm.pyi,sha256=dWfG-M_gD644yJ-LK_T8HnAT649j3Vx7TVof03XQimE,611
|
|
79
|
+
gllm_inference/model/lm/google_lm.pyi,sha256=cMV5zYX8uwUF7pErv4pXnXD2G52umo3sxKwbSx7nFhQ,511
|
|
80
|
+
gllm_inference/model/lm/openai_lm.pyi,sha256=u11zvvIS7-XaHKZ33cZxGQmT6cZ4DqK9Do8l7gFOUTc,618
|
|
81
|
+
gllm_inference/model/lm/xai_lm.pyi,sha256=2ZEQ_--e_zsb23zZQ8bKdQShU7zChx5TrDKF8EpwEpU,506
|
|
82
|
+
gllm_inference/output_parser/__init__.pyi,sha256=WQOOgsYnPk8vd-SOhFMMaVTzy4gkYrOAyT5gnAxv0A0,129
|
|
83
|
+
gllm_inference/output_parser/json_output_parser.pyi,sha256=uulh91uQLMSb4ZXZhHYi9W9w7zGnmrOweEkL6wdDJN8,2933
|
|
84
|
+
gllm_inference/output_parser/output_parser.pyi,sha256=Yzk7F26pH8Uc7FQZo4G6l67YkfppefUvaV9cNK-HyDs,948
|
|
85
|
+
gllm_inference/prompt_builder/__init__.pyi,sha256=kshfBMvwIwiIvjxiGG5BrJZNvpPa8rhtkbHo5FPifBg,117
|
|
86
|
+
gllm_inference/prompt_builder/prompt_builder.pyi,sha256=VQaw8nE8SPK1nvVVB4YjeqZSazigNPuWvOEcgGji0W8,4557
|
|
87
|
+
gllm_inference/prompt_builder/format_strategy/__init__.pyi,sha256=QhORHac3ySOPmL9k9kmCKL70vtaUtwkZEtGoRNWNuA8,308
|
|
88
|
+
gllm_inference/prompt_builder/format_strategy/format_strategy.pyi,sha256=JSUl_7Ka08oDZPpslymkUa8pDzqNGIK2TlcVANspqrY,2273
|
|
89
|
+
gllm_inference/prompt_builder/format_strategy/jinja_format_strategy.pyi,sha256=IAezLUiKSJMaoDyleo8pFnFqq8rBM_Q-lNXuAGvwXhI,2225
|
|
90
|
+
gllm_inference/prompt_builder/format_strategy/string_format_strategy.pyi,sha256=E0r8x6NTVbPLUfbJBz75PW1n3Ong6bi1WNgxBD0FTM4,693
|
|
91
|
+
gllm_inference/prompt_formatter/__init__.pyi,sha256=rTsjfRsT-y00qH67fPewMNPMN1fAO2y7DM9scR1ccm0,740
|
|
92
|
+
gllm_inference/prompt_formatter/agnostic_prompt_formatter.pyi,sha256=c9mN4t8LXn79h8wq8DAeWYwMgmZGzXjP7EcjLpwfNZg,2018
|
|
93
|
+
gllm_inference/prompt_formatter/huggingface_prompt_formatter.pyi,sha256=AJ-D11HBhTKBA1B5s0km_K1R6o5HD1yvdbGoL7SpGhI,2729
|
|
94
|
+
gllm_inference/prompt_formatter/llama_prompt_formatter.pyi,sha256=nz24trAqHsUY_zHVYfdNzhBO_TCUtzRkg-QYYow4l2s,2848
|
|
95
|
+
gllm_inference/prompt_formatter/mistral_prompt_formatter.pyi,sha256=DgFitxfvCSJf8_6NUfSre-PXYMuPI4vX4AN6BPF1cBE,2650
|
|
96
|
+
gllm_inference/prompt_formatter/openai_prompt_formatter.pyi,sha256=QB7gHddipQuAolUKO01ApZeaRxBtRZzglS5B78mzsV8,1286
|
|
97
|
+
gllm_inference/prompt_formatter/prompt_formatter.pyi,sha256=UkcPi5ao98OGJyNRsqfhYTlMW-ZLNITaGZUTwzvBzqk,1146
|
|
98
|
+
gllm_inference/realtime_chat/__init__.pyi,sha256=Fhxr_sYOt_p43BT2uvJmCp5FMhfb3dDMoIZglq2y6G8,137
|
|
99
|
+
gllm_inference/realtime_chat/google_realtime_chat.pyi,sha256=S5FOCacuc7j4J2tjpUHCdp91YlalxIJ3V3Bh0t0GGus,10112
|
|
100
|
+
gllm_inference/realtime_chat/realtime_chat.pyi,sha256=qG0fXyhTPLw6wGpnDX6wQrLfs6RBaxCK2ASwIqOymp4,1375
|
|
101
|
+
gllm_inference/realtime_chat/input_streamer/__init__.pyi,sha256=REN1dEEZB-I2RX5QsBMxWcUiGm0-fWLevgj7IT18W6Y,317
|
|
102
|
+
gllm_inference/realtime_chat/input_streamer/input_streamer.pyi,sha256=DOVBe_lxievONS_t14opNNuv0zM3MWyBcCwmxzR4_BE,1259
|
|
103
|
+
gllm_inference/realtime_chat/input_streamer/keyboard_input_streamer.pyi,sha256=3L0y91efGmb4YFWigs2WGAURQ48EwNxLgcVMm15yrIg,1043
|
|
104
|
+
gllm_inference/realtime_chat/input_streamer/linux_mic_input_streamer.pyi,sha256=o8rOootmcKHmKRUCCHvfGlCcfeBow5UTbQHIw89XIqg,1337
|
|
105
|
+
gllm_inference/realtime_chat/output_streamer/__init__.pyi,sha256=FP5amnApQcIvW7YfQEnLzJBzEDz51npgocfCcRWDQXA,339
|
|
106
|
+
gllm_inference/realtime_chat/output_streamer/console_output_streamer.pyi,sha256=pO_llZK7JDjllTWEZjujEVo83Hvrv_8LwVqEkQiIOOo,674
|
|
107
|
+
gllm_inference/realtime_chat/output_streamer/linux_speaker_output_streamer.pyi,sha256=yie6VaOyozzE4bKtvWr-6hRiobK-8mshCqeUOhK1PVA,1487
|
|
108
|
+
gllm_inference/realtime_chat/output_streamer/output_streamer.pyi,sha256=GPAw1wPSrwHpJ2MeZJB3V9CCESTDZpyYYgsoy_LI2Nk,1083
|
|
109
|
+
gllm_inference/request_processor/__init__.pyi,sha256=hVnfdNZnkTBJHnmLtN3Na4ANP0yK6AstWdIizVr2Apo,227
|
|
110
|
+
gllm_inference/request_processor/lm_request_processor.pyi,sha256=VnYc8E3Iayyhw-rPnGPfTKuO3ohgFsS8HPrZJeyES5I,5889
|
|
111
|
+
gllm_inference/request_processor/uses_lm_mixin.pyi,sha256=Yu0XPNuHxq1tWBviHTPw1oThojneFwGHepvGjBXxKQA,6382
|
|
112
|
+
gllm_inference/schema/__init__.pyi,sha256=kTFb0oGdne0yMXgx71pmIyG_H07gOLf42mtN5n0Vajs,2266
|
|
113
|
+
gllm_inference/schema/activity.pyi,sha256=JnO2hqj91P5Tc6qb4pbkEMrHer2u5owiCvhl-igcQKQ,2303
|
|
114
|
+
gllm_inference/schema/attachment.pyi,sha256=jApuzjOHJDCz4lr4MlHzBgIndh559nbWu2Xp1fk3hso,3297
|
|
115
|
+
gllm_inference/schema/code_exec_result.pyi,sha256=ZTHh6JtRrPIdQ059P1UAiD2L-tAO1_S5YcMsAXfJ5A0,559
|
|
116
|
+
gllm_inference/schema/config.pyi,sha256=rAL_UeXyQeXVk1P2kqd8vFWOMwmKenfpQLtvMP74t9s,674
|
|
117
|
+
gllm_inference/schema/enums.pyi,sha256=-A7BuMVfUiyOhXdZLqagrdm8MsxRS_HV4PGC6khWMlc,1751
|
|
118
|
+
gllm_inference/schema/events.pyi,sha256=VObVT6B5rvXRYReiVL04vKSxlEHZIjIOVTa4iS87s3w,4705
|
|
119
|
+
gllm_inference/schema/lm_input.pyi,sha256=A5pjz1id6tP9XRNhzQrbmzd66C_q3gzo0UP8rCemz6Q,193
|
|
120
|
+
gllm_inference/schema/lm_output.pyi,sha256=1SZi6vIWvmrZlVQ59WeQUKO5VhKrLHsSRDYslEH9d7o,2435
|
|
121
|
+
gllm_inference/schema/mcp.pyi,sha256=Vwu8E2BDl6FvvnI42gIyY3Oki1BdwRE3Uh3aV0rmhQU,1014
|
|
122
|
+
gllm_inference/schema/message.pyi,sha256=VP9YppKj2mo1esl9cy6qQO9m2mMHUjTmfGDdyUor880,2220
|
|
123
|
+
gllm_inference/schema/model_id.pyi,sha256=BHzi8PdYsV_rPGkBcKvZJDCf62ToPjmRTcXcTgKfoTg,5894
|
|
124
|
+
gllm_inference/schema/reasoning.pyi,sha256=SlTuiDw87GdnAn-I6YOPIJRhEBiwQljM46JohG05guQ,562
|
|
125
|
+
gllm_inference/schema/token_usage.pyi,sha256=1GTQVORV0dBNmD_jix8aVaUqxMKFF04KpLP7y2urqbk,2950
|
|
126
|
+
gllm_inference/schema/tool_call.pyi,sha256=zQaVxCnkVxOfOEhBidqohU85gb4PRwnwBiygKaunamk,389
|
|
127
|
+
gllm_inference/schema/tool_result.pyi,sha256=cAG7TVtB4IWJPt8XBBbB92cuY1ZsX9M276bN9aqjcvM,276
|
|
128
|
+
gllm_inference/schema/type_alias.pyi,sha256=q_C1x8hUUoerGEQFj99g3yaZ1ILigGQSEo-FheqS_lM,529
|
|
129
|
+
gllm_inference/utils/__init__.pyi,sha256=mDJ2gLSeQzm-79Tov5-dhrMNaDqgcD1VVzDYAWvIRqA,391
|
|
130
|
+
gllm_inference/utils/io_utils.pyi,sha256=7kUTacHAVRYoemFUOjCH7-Qmw-YsQGd6rGYxjf_qmtw,1094
|
|
131
|
+
gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
|
|
132
|
+
gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
|
|
133
|
+
gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
|
|
134
|
+
gllm_inference_binary-0.5.51b3.dist-info/METADATA,sha256=3W0iVCB7pQeVBoru96XE5Y-xzRereb8lcDN7PoxknvU,5804
|
|
135
|
+
gllm_inference_binary-0.5.51b3.dist-info/WHEEL,sha256=LySLkZbKo_9Y43IIH4b2lKMJ4X6HWu97Axseguj1GhY,104
|
|
136
|
+
gllm_inference_binary-0.5.51b3.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
|
|
137
|
+
gllm_inference_binary-0.5.51b3.dist-info/RECORD,,
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
gllm_inference
|