gllm-inference-binary 0.5.48__cp313-cp313-win_amd64.whl → 0.5.50__cp313-cp313-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of gllm-inference-binary might be problematic. Click here for more details.
- gllm_inference/builder/_build_invoker.pyi +28 -0
- gllm_inference/builder/build_em_invoker.pyi +1 -16
- gllm_inference/builder/build_lm_invoker.pyi +0 -20
- gllm_inference/lm_invoker/bedrock_lm_invoker.pyi +1 -0
- gllm_inference/schema/model_id.pyi +12 -4
- gllm_inference.cp313-win_amd64.pyd +0 -0
- {gllm_inference_binary-0.5.48.dist-info → gllm_inference_binary-0.5.50.dist-info}/METADATA +1 -1
- {gllm_inference_binary-0.5.48.dist-info → gllm_inference_binary-0.5.50.dist-info}/RECORD +10 -9
- {gllm_inference_binary-0.5.48.dist-info → gllm_inference_binary-0.5.50.dist-info}/WHEEL +0 -0
- {gllm_inference_binary-0.5.48.dist-info → gllm_inference_binary-0.5.50.dist-info}/top_level.txt +0 -0
|
@@ -0,0 +1,28 @@
|
|
|
1
|
+
from _typeshed import Incomplete
|
|
2
|
+
from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
|
|
3
|
+
from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
|
|
4
|
+
from gllm_inference.schema.model_id import ModelId as ModelId, ModelProvider as ModelProvider, PROVIDERS_OPTIONAL_PATH as PROVIDERS_OPTIONAL_PATH
|
|
5
|
+
|
|
6
|
+
logger: Incomplete
|
|
7
|
+
|
|
8
|
+
class Key:
|
|
9
|
+
"""Defines valid keys in the config."""
|
|
10
|
+
ACCESS_KEY_ID: str
|
|
11
|
+
API_KEY: str
|
|
12
|
+
AZURE_DEPLOYMENT: str
|
|
13
|
+
AZURE_ENDPOINT: str
|
|
14
|
+
BASE_URL: str
|
|
15
|
+
CONFIG: str
|
|
16
|
+
CUSTOM_HOST: str
|
|
17
|
+
CREDENTIALS_PATH: str
|
|
18
|
+
MODEL_ID: str
|
|
19
|
+
MODEL_KWARGS: str
|
|
20
|
+
MODEL_NAME: str
|
|
21
|
+
MODEL_CLASS_PATH: str
|
|
22
|
+
PORTKEY_API_KEY: str
|
|
23
|
+
PROVIDER: str
|
|
24
|
+
SECRET_ACCESS_KEY: str
|
|
25
|
+
|
|
26
|
+
PROVIDERS_REQUIRE_BASE_URL: Incomplete
|
|
27
|
+
MODEL_NAME_KEY_MAP: Incomplete
|
|
28
|
+
DEFAULT_MODEL_NAME_KEY: Incomplete
|
|
@@ -1,24 +1,9 @@
|
|
|
1
|
-
from _typeshed import Incomplete
|
|
2
1
|
from gllm_inference.em_invoker import AzureOpenAIEMInvoker as AzureOpenAIEMInvoker, BedrockEMInvoker as BedrockEMInvoker, CohereEMInvoker as CohereEMInvoker, GoogleEMInvoker as GoogleEMInvoker, JinaEMInvoker as JinaEMInvoker, LangChainEMInvoker as LangChainEMInvoker, OpenAICompatibleEMInvoker as OpenAICompatibleEMInvoker, OpenAIEMInvoker as OpenAIEMInvoker, TwelveLabsEMInvoker as TwelveLabsEMInvoker, VoyageEMInvoker as VoyageEMInvoker
|
|
3
2
|
from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
|
|
4
|
-
from gllm_inference.schema.model_id import ModelId as ModelId, ModelProvider as ModelProvider
|
|
3
|
+
from gllm_inference.schema.model_id import ModelId as ModelId, ModelProvider as ModelProvider
|
|
5
4
|
from typing import Any
|
|
6
5
|
|
|
7
6
|
PROVIDER_TO_EM_INVOKER_MAP: dict[str, type[BaseEMInvoker]]
|
|
8
|
-
logger: Incomplete
|
|
9
|
-
|
|
10
|
-
class Key:
|
|
11
|
-
"""Defines valid keys in the config."""
|
|
12
|
-
ACCESS_KEY_ID: str
|
|
13
|
-
API_KEY: str
|
|
14
|
-
AZURE_DEPLOYMENT: str
|
|
15
|
-
AZURE_ENDPOINT: str
|
|
16
|
-
BASE_URL: str
|
|
17
|
-
CREDENTIALS_PATH: str
|
|
18
|
-
MODEL_KWARGS: str
|
|
19
|
-
MODEL_NAME: str
|
|
20
|
-
MODEL_CLASS_PATH: str
|
|
21
|
-
SECRET_ACCESS_KEY: str
|
|
22
7
|
|
|
23
8
|
def build_em_invoker(model_id: str | ModelId, credentials: str | dict[str, Any] | None = None, config: dict[str, Any] | None = None) -> BaseEMInvoker:
|
|
24
9
|
'''Build an embedding model invoker based on the provided configurations.
|
|
@@ -1,29 +1,9 @@
|
|
|
1
|
-
from _typeshed import Incomplete
|
|
2
1
|
from gllm_inference.lm_invoker import AnthropicLMInvoker as AnthropicLMInvoker, AzureOpenAILMInvoker as AzureOpenAILMInvoker, BedrockLMInvoker as BedrockLMInvoker, DatasaurLMInvoker as DatasaurLMInvoker, GoogleLMInvoker as GoogleLMInvoker, LangChainLMInvoker as LangChainLMInvoker, LiteLLMLMInvoker as LiteLLMLMInvoker, OpenAIChatCompletionsLMInvoker as OpenAIChatCompletionsLMInvoker, OpenAICompatibleLMInvoker as OpenAICompatibleLMInvoker, OpenAILMInvoker as OpenAILMInvoker, PortkeyLMInvoker as PortkeyLMInvoker, XAILMInvoker as XAILMInvoker
|
|
3
2
|
from gllm_inference.lm_invoker.lm_invoker import BaseLMInvoker as BaseLMInvoker
|
|
4
3
|
from gllm_inference.schema.model_id import ModelId as ModelId, ModelProvider as ModelProvider
|
|
5
4
|
from typing import Any
|
|
6
5
|
|
|
7
6
|
PROVIDER_TO_LM_INVOKER_MAP: dict[str, type[BaseLMInvoker]]
|
|
8
|
-
logger: Incomplete
|
|
9
|
-
|
|
10
|
-
class Key:
|
|
11
|
-
"""Defines valid keys in the config."""
|
|
12
|
-
ACCESS_KEY_ID: str
|
|
13
|
-
API_KEY: str
|
|
14
|
-
AZURE_DEPLOYMENT: str
|
|
15
|
-
AZURE_ENDPOINT: str
|
|
16
|
-
BASE_URL: str
|
|
17
|
-
CONFIG: str
|
|
18
|
-
CUSTOM_HOST: str
|
|
19
|
-
CREDENTIALS_PATH: str
|
|
20
|
-
MODEL_ID: str
|
|
21
|
-
MODEL_KWARGS: str
|
|
22
|
-
MODEL_NAME: str
|
|
23
|
-
MODEL_CLASS_PATH: str
|
|
24
|
-
PORTKEY_API_KEY: str
|
|
25
|
-
PROVIDER: str
|
|
26
|
-
SECRET_ACCESS_KEY: str
|
|
27
7
|
|
|
28
8
|
def build_lm_invoker(model_id: str | ModelId, credentials: str | dict[str, Any] | None = None, config: dict[str, Any] | None = None) -> BaseLMInvoker:
|
|
29
9
|
'''Build a language model invoker based on the provided configurations.
|
|
@@ -10,6 +10,7 @@ from gllm_inference.schema import Attachment as Attachment, AttachmentType as At
|
|
|
10
10
|
from langchain_core.tools import Tool as LangChainTool
|
|
11
11
|
from typing import Any
|
|
12
12
|
|
|
13
|
+
FILENAME_SANITIZATION_REGEX: Incomplete
|
|
13
14
|
SUPPORTED_ATTACHMENTS: Incomplete
|
|
14
15
|
|
|
15
16
|
class BedrockLMInvoker(BaseLMInvoker):
|
|
@@ -26,8 +26,8 @@ class ModelProvider(StrEnum):
|
|
|
26
26
|
VOYAGE = 'voyage'
|
|
27
27
|
XAI = 'xai'
|
|
28
28
|
|
|
29
|
-
|
|
30
|
-
|
|
29
|
+
PROVIDERS_OPTIONAL_PATH: Incomplete
|
|
30
|
+
PROVIDERS_SUPPORT_PATH: Incomplete
|
|
31
31
|
|
|
32
32
|
class ModelId(BaseModel):
|
|
33
33
|
'''Defines a representation of a valid model id.
|
|
@@ -53,6 +53,11 @@ class ModelId(BaseModel):
|
|
|
53
53
|
model_id = ModelId.from_string("cohere/embed-english-v3.0")
|
|
54
54
|
```
|
|
55
55
|
|
|
56
|
+
# Using Cohere with custom endpoint
|
|
57
|
+
```python
|
|
58
|
+
model_id = ModelId.from_string("cohere/https://my-cohere-url:8000/v1:my-model-name")
|
|
59
|
+
```
|
|
60
|
+
|
|
56
61
|
# Using Datasaur
|
|
57
62
|
```python
|
|
58
63
|
model_id = ModelId.from_string("datasaur/https://deployment.datasaur.ai/api/deployment/teamId/deploymentId/")
|
|
@@ -67,8 +72,11 @@ class ModelId(BaseModel):
|
|
|
67
72
|
```python
|
|
68
73
|
model_id = ModelId.from_string("jina/jina-embeddings-v2-large")
|
|
69
74
|
```
|
|
70
|
-
|
|
71
|
-
|
|
75
|
+
|
|
76
|
+
# Using Jina with custom endpoint
|
|
77
|
+
```python
|
|
78
|
+
model_id = ModelId.from_string("jina/https://my-jina-url:8000/v1:my-model-name")
|
|
79
|
+
```
|
|
72
80
|
|
|
73
81
|
# Using OpenAI
|
|
74
82
|
```python
|
|
Binary file
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.2
|
|
2
2
|
Name: gllm-inference-binary
|
|
3
|
-
Version: 0.5.
|
|
3
|
+
Version: 0.5.50
|
|
4
4
|
Summary: A library containing components related to model inferences in Gen AI applications.
|
|
5
5
|
Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, Resti Febrina <resti.febrina@gdplabs.id>
|
|
6
6
|
Requires-Python: <3.14,>=3.11
|
|
@@ -1,10 +1,11 @@
|
|
|
1
|
-
gllm_inference.cp313-win_amd64.pyd,sha256=
|
|
1
|
+
gllm_inference.cp313-win_amd64.pyd,sha256=Le9KSLHrPw_25jEmrsuKOeZwauAI6KJGpK7Wm2ZkT_o,3828736
|
|
2
2
|
gllm_inference.pyi,sha256=1WeCtSLoqo97eCY-WiMP-LF9UUJG_pT5NTESuCoStRg,5211
|
|
3
3
|
gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
4
4
|
gllm_inference/constants.pyi,sha256=PncjVw-mkzcJ3ln1ohvVZGdJ-TD-VZy1Ygn4Va8Z7i0,350
|
|
5
5
|
gllm_inference/builder/__init__.pyi,sha256=-bw1uDx7CAM7pkvjvb1ZXku9zXlQ7aEAyC83KIn3bz8,506
|
|
6
|
-
gllm_inference/builder/
|
|
7
|
-
gllm_inference/builder/
|
|
6
|
+
gllm_inference/builder/_build_invoker.pyi,sha256=SZst1ocnRntcA7lWKp44lMJ3yRA5GP_jsUZkcsxwTjw,876
|
|
7
|
+
gllm_inference/builder/build_em_invoker.pyi,sha256=Z09hP6_ziumI2XB-7HCdyIE4AyoV1oL9DBdoU2G8izo,5879
|
|
8
|
+
gllm_inference/builder/build_lm_invoker.pyi,sha256=1Qgb0bD8tqobDh9kR_Kce7Vwu_Oljc2xv-HBIrC3HUY,9065
|
|
8
9
|
gllm_inference/builder/build_lm_request_processor.pyi,sha256=H7Rg88e7PTTCtuyY64r333moTmh4-ypOwgnG10gkEdY,4232
|
|
9
10
|
gllm_inference/builder/build_output_parser.pyi,sha256=sgSTrzUmSRxPzUUum0fDU7A3NXYoYhpi6bEx4Q2XMnA,965
|
|
10
11
|
gllm_inference/catalog/__init__.pyi,sha256=HWgPKWIzprpMHRKe_qN9BZSIQhVhrqiyjLjIXwvj1ho,291
|
|
@@ -42,7 +43,7 @@ gllm_inference/exceptions/provider_error_map.pyi,sha256=vWa4ZIHn7qIghECGvO-dS2Kz
|
|
|
42
43
|
gllm_inference/lm_invoker/__init__.pyi,sha256=L2nlkj13WwWbDYEBtM0mlAj0-UbSilMjVLpCJ_0Eock,1502
|
|
43
44
|
gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=JSgKUk9d1ZHlitv_ZjHlAk2hIW-J7u6yslVHflIeUro,16726
|
|
44
45
|
gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=FYfRNPG-oD4wIfitjTHnGib1uMZL7Pid0gbrRsymAHU,14601
|
|
45
|
-
gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=
|
|
46
|
+
gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=IMFD6Nr_4tkDDMy1WveOz91u_v1rjplVdtgep9UXsxw,11851
|
|
46
47
|
gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=LR0EM4vTfufq9OWk8JVIwLyFeJFTguPNmPgJBUooSq4,8342
|
|
47
48
|
gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=aSmEgoYj_V72Nb6erDResphw9RaHfbE5C6PhqpMfEeQ,17674
|
|
48
49
|
gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=tJIxkFUKjLF-yz0niaDjN3L0QNCbn4sT8hmPKtERpog,12742
|
|
@@ -119,7 +120,7 @@ gllm_inference/schema/lm_input.pyi,sha256=HxQiZgY7zcXh_Dw8nK8LSeBTZEHMPZVwmPmnfg
|
|
|
119
120
|
gllm_inference/schema/lm_output.pyi,sha256=DIV8BiIOPaSnMKxzKzH_Mp7j7-MScWCvmllegJDLqFg,2479
|
|
120
121
|
gllm_inference/schema/mcp.pyi,sha256=4SgQ83pEowfWm2p-w9lupV4NayqqVBOy7SuYxIFeWRs,1045
|
|
121
122
|
gllm_inference/schema/message.pyi,sha256=jJV6A0ihEcun2OhzyMtNkiHnf7d6v5R-GdpTBGfJ0AQ,2272
|
|
122
|
-
gllm_inference/schema/model_id.pyi,sha256=
|
|
123
|
+
gllm_inference/schema/model_id.pyi,sha256=Ozu46pp1uaPPvXJ6SPKpiL1vETPJqErEjNAYfOb5b6Y,6070
|
|
123
124
|
gllm_inference/schema/reasoning.pyi,sha256=jbPxkDRHt0Vt-zdcc8lTT1l2hIE1Jm3HIHeNd0hfXGo,577
|
|
124
125
|
gllm_inference/schema/token_usage.pyi,sha256=WJiGQyz5qatzBK2b-sABLCyTRLCBbAvxCRcqSJOzu-8,3025
|
|
125
126
|
gllm_inference/schema/tool_call.pyi,sha256=OWT9LUqs_xfUcOkPG0aokAAqzLYYDkfnjTa0zOWvugk,403
|
|
@@ -130,7 +131,7 @@ gllm_inference/utils/io_utils.pyi,sha256=Eg7dvHWdXslTKdjh1j3dG50i7r35XG2zTmJ9XXv
|
|
|
130
131
|
gllm_inference/utils/langchain.pyi,sha256=4AwFiVAO0ZpdgmqeC4Pb5NJwBt8vVr0MSUqLeCdTscc,1194
|
|
131
132
|
gllm_inference/utils/validation.pyi,sha256=-RdMmb8afH7F7q4Ao7x6FbwaDfxUHn3hA3WiOgzB-3s,397
|
|
132
133
|
gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
|
|
133
|
-
gllm_inference_binary-0.5.
|
|
134
|
-
gllm_inference_binary-0.5.
|
|
135
|
-
gllm_inference_binary-0.5.
|
|
136
|
-
gllm_inference_binary-0.5.
|
|
134
|
+
gllm_inference_binary-0.5.50.dist-info/METADATA,sha256=tbA509mUnhG0BXBzhX3qj3btekTzMYWDDD_PnnEyV8Q,5945
|
|
135
|
+
gllm_inference_binary-0.5.50.dist-info/WHEEL,sha256=O_u6PJIQ2pIcyIInxVQ9r-yArMuUZbBIaF1kpYVkYxA,96
|
|
136
|
+
gllm_inference_binary-0.5.50.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
|
|
137
|
+
gllm_inference_binary-0.5.50.dist-info/RECORD,,
|
|
File without changes
|
{gllm_inference_binary-0.5.48.dist-info → gllm_inference_binary-0.5.50.dist-info}/top_level.txt
RENAMED
|
File without changes
|