gllm-inference-binary 0.5.24__cp312-cp312-win_amd64.whl → 0.5.26.post1__cp312-cp312-win_amd64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of gllm-inference-binary might be problematic. Click here for more details.

@@ -86,7 +86,7 @@ def build_em_invoker(model_id: str | ModelId, credentials: str | dict[str, Any]
86
86
  # Using Azure OpenAI
87
87
  ```python
88
88
  em_invoker = build_em_invoker(
89
- model_id="azure-openai/https://my-resource.openai.azure.com:my-deployment",
89
+ model_id="azure-openai/https://my-resource.openai.azure.com/openai/v1:my-deployment",
90
90
  credentials="azure-api-key"
91
91
  )
92
92
  ```
@@ -1,6 +1,6 @@
1
1
  from _typeshed import Incomplete
2
2
 
3
- DEFAULT_AZURE_OPENAI_API_VERSION: str
3
+ AZURE_OPENAI_URL_SUFFIX: str
4
4
  DOCUMENT_MIME_TYPES: Incomplete
5
5
  GOOGLE_SCOPES: Incomplete
6
6
  GRPC_ENABLE_RETRIES_KEY: str
@@ -1,6 +1,6 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.utils.retry import RetryConfig as RetryConfig
3
- from gllm_inference.constants import DEFAULT_AZURE_OPENAI_API_VERSION as DEFAULT_AZURE_OPENAI_API_VERSION, INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
3
+ from gllm_inference.constants import AZURE_OPENAI_URL_SUFFIX as AZURE_OPENAI_URL_SUFFIX, INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
4
4
  from gllm_inference.em_invoker.openai_em_invoker import OpenAIEMInvoker as OpenAIEMInvoker
5
5
  from gllm_inference.em_invoker.schema.openai import Key as Key
6
6
  from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider, TruncationConfig as TruncationConfig
@@ -69,7 +69,7 @@ class AzureOpenAIEMInvoker(OpenAIEMInvoker):
69
69
  ```
70
70
  '''
71
71
  client: Incomplete
72
- def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
72
+ def __init__(self, azure_endpoint: str, azure_deployment: str, api_key: str | None = None, api_version: str | None = None, model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, retry_config: RetryConfig | None = None, truncation_config: TruncationConfig | None = None) -> None:
73
73
  """Initializes a new instance of the AzureOpenAIEMInvoker class.
74
74
 
75
75
  Args:
@@ -77,8 +77,7 @@ class AzureOpenAIEMInvoker(OpenAIEMInvoker):
77
77
  azure_deployment (str): The deployment name of the Azure OpenAI service.
78
78
  api_key (str | None, optional): The API key for authenticating with Azure OpenAI. Defaults to None, in
79
79
  which case the `AZURE_OPENAI_API_KEY` environment variable will be used.
80
- api_version (str, optional): The API version of the Azure OpenAI service. Defaults to
81
- `DEFAULT_AZURE_OPENAI_API_VERSION`.
80
+ api_version (str | None, optional): Deprecated parameter to be removed in v0.6. Defaults to None.
82
81
  model_kwargs (dict[str, Any] | None, optional): Additional model parameters. Defaults to None.
83
82
  default_hyperparameters (dict[str, Any] | None, optional): Default hyperparameters for invoking the model.
84
83
  Defaults to None.
@@ -1,15 +1,13 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.schema.tool import Tool as Tool
3
3
  from gllm_core.utils.retry import RetryConfig as RetryConfig
4
- from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
4
+ from gllm_inference.constants import AZURE_OPENAI_URL_SUFFIX as AZURE_OPENAI_URL_SUFFIX, INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
5
5
  from gllm_inference.lm_invoker.openai_lm_invoker import OpenAILMInvoker as OpenAILMInvoker, ReasoningEffort as ReasoningEffort, ReasoningSummary as ReasoningSummary
6
6
  from gllm_inference.lm_invoker.schema.openai import Key as Key
7
7
  from gllm_inference.schema import ModelId as ModelId, ModelProvider as ModelProvider, ResponseSchema as ResponseSchema
8
8
  from langchain_core.tools import Tool as LangChainTool
9
9
  from typing import Any
10
10
 
11
- URL_SUFFIX: str
12
-
13
11
  class AzureOpenAILMInvoker(OpenAILMInvoker):
14
12
  '''A language model invoker to interact with Azure OpenAI language models.
15
13
 
Binary file
@@ -1,110 +1,71 @@
1
- Metadata-Version: 2.1
2
- Name: gllm-inference-binary
3
- Version: 0.5.24
4
- Summary: A library containing components related to model inferences in Gen AI applications.
5
- Author: Henry Wicaksono
6
- Author-email: henry.wicaksono@gdplabs.id
7
- Requires-Python: >=3.11,<3.14
8
- Classifier: Programming Language :: Python :: 3
9
- Classifier: Programming Language :: Python :: 3.11
10
- Classifier: Programming Language :: Python :: 3.12
11
- Provides-Extra: anthropic
12
- Provides-Extra: bedrock
13
- Provides-Extra: datasaur
14
- Provides-Extra: google
15
- Provides-Extra: huggingface
16
- Provides-Extra: litellm
17
- Provides-Extra: openai
18
- Provides-Extra: twelvelabs
19
- Provides-Extra: voyage
20
- Provides-Extra: xai
21
- Requires-Dist: aioboto3 (>=15.0.0,<16.0.0) ; extra == "bedrock"
22
- Requires-Dist: aiohttp (>=3.12.14,<4.0.0)
23
- Requires-Dist: anthropic (>=0.60.0,<0.61.0) ; extra == "anthropic"
24
- Requires-Dist: filetype (>=1.2.0,<2.0.0)
25
- Requires-Dist: gllm-core-binary (>=0.3.0,<0.4.0)
26
- Requires-Dist: google-genai (>=1.26.0,<2.0.0) ; extra == "google"
27
- Requires-Dist: httpx (>=0.28.0,<0.29.0)
28
- Requires-Dist: huggingface-hub (>=0.30.0,<0.31.0) ; extra == "huggingface"
29
- Requires-Dist: jinja2 (>=3.1.4,<4.0.0)
30
- Requires-Dist: jsonschema (>=4.24.0,<5.0.0)
31
- Requires-Dist: langchain (>=0.3.0,<0.4.0)
32
- Requires-Dist: litellm (>=1.69.2,<2.0.0) ; extra == "litellm"
33
- Requires-Dist: openai (>=1.98.0,<2.0.0) ; extra == "datasaur" or extra == "openai"
34
- Requires-Dist: pandas (>=2.2.3,<3.0.0)
35
- Requires-Dist: poetry (>=2.1.3,<3.0.0)
36
- Requires-Dist: protobuf (>=6.0.0,<7.0.0)
37
- Requires-Dist: python-magic (>=0.4.27,<0.5.0) ; sys_platform != "win32"
38
- Requires-Dist: python-magic-bin (>=0.4.14,<0.5.0) ; sys_platform == "win32"
39
- Requires-Dist: sentencepiece (>=0.2.0,<0.3.0)
40
- Requires-Dist: transformers (==4.52.4) ; extra == "huggingface"
41
- Requires-Dist: twelvelabs (>=0.4.4,<0.5.0) ; extra == "twelvelabs"
42
- Requires-Dist: voyageai (>=0.3.0,<0.4.0) ; (python_version < "3.13") and (extra == "voyage")
43
- Requires-Dist: xai_sdk (>=1.0.0,<2.0.0) ; extra == "xai"
44
- Description-Content-Type: text/markdown
45
-
46
- # GLLM Inference
47
-
48
- ## Description
49
-
50
- A library containing components related to model inferences in Gen AI applications.
51
-
52
- ## Installation
53
-
54
- ### Prerequisites
55
- - Python 3.11+ - [Install here](https://www.python.org/downloads/)
56
- - Pip (if using Pip) - [Install here](https://pip.pypa.io/en/stable/installation/)
57
- - Poetry 1.8.1+ (if using Poetry) - [Install here](https://python-poetry.org/docs/#installation)
58
- - Git (if using Git) - [Install here](https://git-scm.com/downloads)
59
- - For git installation:
60
- - Access to the [GDP Labs SDK github repository](https://github.com/GDP-ADMIN/gen-ai-internal)
61
-
62
- ### 1. Installation from Artifact Registry
63
- Choose one of the following methods to install the package:
64
-
65
- #### Using pip
66
- ```bash
67
- pip install gllm-inference-binary
68
- ```
69
-
70
- #### Using Poetry
71
- ```bash
72
- poetry add gllm-inference-binary
73
- ```
74
-
75
- ### 2. Development Installation (Git)
76
- For development purposes, you can install directly from the Git repository:
77
- ```bash
78
- poetry add "git+ssh://git@github.com/GDP-ADMIN/gen-ai-internal.git#subdirectory=libs/gllm-inference"
79
- ```
80
-
81
- Available extras:
82
- - `anthropic`: Install Anthropic models dependencies
83
- - `google-genai`: Install Google Generative AI models dependencies
84
- - `google-vertexai`: Install Google Vertex AI models dependencies
85
- - `huggingface`: Install HuggingFace models dependencies
86
- - `openai`: Install OpenAI models dependencies
87
- - `twelvelabs`: Install TwelveLabs models dependencies
88
-
89
- ## Managing Dependencies
90
- 1. Go to root folder of `gllm-inference` module, e.g. `cd libs/gllm-inference`.
91
- 2. Run `poetry shell` to create a virtual environment.
92
- 3. Run `poetry lock` to create a lock file if you haven't done it yet.
93
- 4. Run `poetry install` to install the `gllm-inference` requirements for the first time.
94
- 5. Run `poetry update` if you update any dependency module version at `pyproject.toml`.
95
-
96
- ## Contributing
97
- Please refer to this [Python Style Guide](https://docs.google.com/document/d/1uRggCrHnVfDPBnG641FyQBwUwLoFw0kTzNqRm92vUwM/edit?usp=sharing)
98
- to get information about code style, documentation standard, and SCA that you need to use when contributing to this project
99
-
100
- 1. Activate `pre-commit` hooks using `pre-commit install`
101
- 2. Run `poetry shell` to create a virtual environment.
102
- 3. Run `poetry lock` to create a lock file if you haven't done it yet.
103
- 4. Run `poetry install` to install the `gllm-inference` requirements for the first time.
104
- 5. Run `which python` to get the path to be referenced at Visual Studio Code interpreter path (`Ctrl`+`Shift`+`P` or `Cmd`+`Shift`+`P`)
105
- 6. Try running the unit test to see if it's working:
106
- ```bash
107
- poetry run pytest -s tests/unit_tests/
108
- ```
109
-
110
-
1
+ Metadata-Version: 2.2
2
+ Name: gllm-inference-binary
3
+ Version: 0.5.26.post1
4
+ Summary: A library containing components related to model inferences in Gen AI applications.
5
+ Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, Resti Febrina <resti.febrina@gdplabs.id>
6
+ Description-Content-Type: text/markdown
7
+
8
+ # GLLM Inference
9
+
10
+ ## Description
11
+
12
+ A library containing components related to model inferences in Gen AI applications.
13
+
14
+ ## Installation
15
+
16
+ ### Prerequisites
17
+ - Python 3.11+ - [Install here](https://www.python.org/downloads/)
18
+ - Pip (if using Pip) - [Install here](https://pip.pypa.io/en/stable/installation/)
19
+ - Poetry 1.8.1+ (if using Poetry) - [Install here](https://python-poetry.org/docs/#installation)
20
+ - Git (if using Git) - [Install here](https://git-scm.com/downloads)
21
+ - For git installation:
22
+ - Access to the [GDP Labs SDK github repository](https://github.com/GDP-ADMIN/gen-ai-internal)
23
+
24
+ ### 1. Installation from Artifact Registry
25
+ Choose one of the following methods to install the package:
26
+
27
+ #### Using pip
28
+ ```bash
29
+ pip install gllm-inference-binary
30
+ ```
31
+
32
+ #### Using Poetry
33
+ ```bash
34
+ poetry add gllm-inference-binary
35
+ ```
36
+
37
+ ### 2. Development Installation (Git)
38
+ For development purposes, you can install directly from the Git repository:
39
+ ```bash
40
+ poetry add "git+ssh://git@github.com/GDP-ADMIN/gen-ai-internal.git#subdirectory=libs/gllm-inference"
41
+ ```
42
+
43
+ Available extras:
44
+ - `anthropic`: Install Anthropic models dependencies
45
+ - `google-genai`: Install Google Generative AI models dependencies
46
+ - `google-vertexai`: Install Google Vertex AI models dependencies
47
+ - `huggingface`: Install HuggingFace models dependencies
48
+ - `openai`: Install OpenAI models dependencies
49
+ - `twelvelabs`: Install TwelveLabs models dependencies
50
+
51
+ ## Managing Dependencies
52
+ 1. Go to root folder of `gllm-inference` module, e.g. `cd libs/gllm-inference`.
53
+ 2. Run `poetry shell` to create a virtual environment.
54
+ 3. Run `poetry lock` to create a lock file if you haven't done it yet.
55
+ 4. Run `poetry install` to install the `gllm-inference` requirements for the first time.
56
+ 5. Run `poetry update` if you update any dependency module version at `pyproject.toml`.
57
+
58
+ ## Contributing
59
+ Please refer to this [Python Style Guide](https://docs.google.com/document/d/1uRggCrHnVfDPBnG641FyQBwUwLoFw0kTzNqRm92vUwM/edit?usp=sharing)
60
+ to get information about code style, documentation standard, and SCA that you need to use when contributing to this project
61
+
62
+ 1. Activate `pre-commit` hooks using `pre-commit install`
63
+ 2. Run `poetry shell` to create a virtual environment.
64
+ 3. Run `poetry lock` to create a lock file if you haven't done it yet.
65
+ 4. Run `poetry install` to install the `gllm-inference` requirements for the first time.
66
+ 5. Run `which python` to get the path to be referenced at Visual Studio Code interpreter path (`Ctrl`+`Shift`+`P` or `Cmd`+`Shift`+`P`)
67
+ 6. Try running the unit test to see if it's working:
68
+ ```bash
69
+ poetry run pytest -s tests/unit_tests/
70
+ ```
71
+
@@ -1,6 +1,9 @@
1
+ gllm_inference.cp312-win_amd64.pyd,sha256=GKmSdulz-x2M3iUr471G6KOvRKhDEU-VrIsQ6f_q13s,3068928
2
+ gllm_inference.pyi,sha256=Pq6P04np3S3x7juGVCzC5sL2im4MsyligEvahVQNWzM,3820
1
3
  gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
4
+ gllm_inference/constants.pyi,sha256=KQmondDEkHK2P249ymmce3SdutVrx8kYm4v1eTCkW9U,277
2
5
  gllm_inference/builder/__init__.pyi,sha256=-bw1uDx7CAM7pkvjvb1ZXku9zXlQ7aEAyC83KIn3bz8,506
3
- gllm_inference/builder/build_em_invoker.pyi,sha256=cBqifw0frhYZinDndeCjqDVqv7oeW728-i5nH4JMyhk,6010
6
+ gllm_inference/builder/build_em_invoker.pyi,sha256=Mh1vRoJhsqc8hX4jUdopV14Fn44ql27NB7xbGjoHJtE,6020
4
7
  gllm_inference/builder/build_lm_invoker.pyi,sha256=p63iuVBOOpNizItGK6HDxYDrgXdovtfSe0VrvrEd-PA,7047
5
8
  gllm_inference/builder/build_lm_request_processor.pyi,sha256=0pJINCP4nnXVwuhIbhsaiwzjX8gohQt2oqXFZhTFSUs,4584
6
9
  gllm_inference/builder/build_output_parser.pyi,sha256=sgSTrzUmSRxPzUUum0fDU7A3NXYoYhpi6bEx4Q2XMnA,965
@@ -8,17 +11,18 @@ gllm_inference/catalog/__init__.pyi,sha256=HWgPKWIzprpMHRKe_qN9BZSIQhVhrqiyjLjIX
8
11
  gllm_inference/catalog/catalog.pyi,sha256=eWPqgQKi-SJGHabi_XOTEKpAj96OSRypKsb5ZEC1VWU,4911
9
12
  gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=GemCEjFRHNChtNOfbyXSVsJiA3klOCAe_X11fnymhYs,5540
10
13
  gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=iViWB4SaezzjQY4UY1YxeoXUNxqxa2cTJGaD9JSx4Q8,3279
11
- gllm_inference/constants.pyi,sha256=0WlEIaWQipEfW2yFmxUxHfZyAzw34SrWyGs_jHZMSts,286
12
14
  gllm_inference/em_invoker/__init__.pyi,sha256=pmbsjmsqXwfe4WPykMnrmasKrYuylJWnf2s0pbo0ioM,997
13
- gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=g1I3Aexg5VeDeU_zbZWCVgca2fhrUztVrpbzS5GBBYI,5072
15
+ gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=SfJPC_PJGiEfWS9JH5kRQPJztsR7jRhwVuETqdY-JsQ,5021
14
16
  gllm_inference/em_invoker/bedrock_em_invoker.pyi,sha256=UqodtpDmE7fEgpctXEETIlZGorX9i1lmmuTvGaJke6o,5829
15
17
  gllm_inference/em_invoker/em_invoker.pyi,sha256=YDYJ8TGScsz5Gg-OBnEENN1tI1RYvwoddypxUr6SAWw,5191
16
18
  gllm_inference/em_invoker/google_em_invoker.pyi,sha256=q69kdVuE44ZqziQ8BajFYZ1tYn-MPjKjzXS9cRh4oAo,6951
17
- gllm_inference/em_invoker/langchain/__init__.pyi,sha256=aOTlRvS9aG1tBErjsmhe75s4Sq-g2z9ArfGqNW7QyEs,151
18
- gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi,sha256=BBSDazMOckO9Aw17tC3LGUTPqLb01my1xUZLtKZlwJY,3388
19
19
  gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=nhX6LynrjhfySEt_44OlLoSBd15hoz3giWyNM9CYLKY,3544
20
20
  gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=zEYOBDXKQhvcMGer9DYDu50_3KRDjYyN8-JgpBIFPOI,5456
21
21
  gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=0TDIQa-5UwsPcVxgkze-QJJWrt-ToakAKbuAk9TW5SM,4746
22
+ gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=MMVgSnjMXksdhSDXIi3vOULIXnjbhtq19eR5LPnUmGo,5446
23
+ gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=vdB_qS8QKrCcb-HtXwKZS4WW1R1wGzpMBFmOKC39sjU,5619
24
+ gllm_inference/em_invoker/langchain/__init__.pyi,sha256=aOTlRvS9aG1tBErjsmhe75s4Sq-g2z9ArfGqNW7QyEs,151
25
+ gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi,sha256=BBSDazMOckO9Aw17tC3LGUTPqLb01my1xUZLtKZlwJY,3388
22
26
  gllm_inference/em_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
23
27
  gllm_inference/em_invoker/schema/bedrock.pyi,sha256=HoNgVi0T21aFd1JrCnSLu4yryv8k8RnYdR3-tIdHFgA,498
24
28
  gllm_inference/em_invoker/schema/google.pyi,sha256=bzdtu4DFH2kATLybIeNl_Lznj99H-6u2Fvx3Zx52oZg,190
@@ -27,15 +31,13 @@ gllm_inference/em_invoker/schema/openai.pyi,sha256=rNRqN62y5wHOKlr4T0n0m41ikAnSr
27
31
  gllm_inference/em_invoker/schema/openai_compatible.pyi,sha256=A9MOeBhI-IPuvewOk4YYOAGtgyKohERx6-9cEYtbwvs,157
28
32
  gllm_inference/em_invoker/schema/twelvelabs.pyi,sha256=D3F9_1F-UTzE6Ymxj6u0IFdL6OFVGlc7noZJr3iuA6I,389
29
33
  gllm_inference/em_invoker/schema/voyage.pyi,sha256=Aqvu6mhFkNb01aXAI5mChLKIgEnFnr-jNKq1lVWB54M,304
30
- gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=MMVgSnjMXksdhSDXIi3vOULIXnjbhtq19eR5LPnUmGo,5446
31
- gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=vdB_qS8QKrCcb-HtXwKZS4WW1R1wGzpMBFmOKC39sjU,5619
32
34
  gllm_inference/exceptions/__init__.pyi,sha256=nXOqwsuwUgsnBcJEANVuxbZ1nDfcJ6-pKUfKeZwltkk,1218
33
35
  gllm_inference/exceptions/error_parser.pyi,sha256=4aiJZhBzBOqlhdmpvaCvildGy7_XxlJzQpe3PzGt8eE,2040
34
36
  gllm_inference/exceptions/exceptions.pyi,sha256=6y3ECgHAStqMGgQv8Dv-Ui-5PDD07mSj6qaRZeSWea4,5857
35
37
  gllm_inference/exceptions/provider_error_map.pyi,sha256=4AsAgbXAh91mxEW2YiomEuhBoeSNeAIo9WbT9WK8gQk,1233
36
38
  gllm_inference/lm_invoker/__init__.pyi,sha256=eE_HDCl9A135mi6mtIV55q-T9J1O8OpbMcqWuny3w9A,1214
37
39
  gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=85uvShLv4-eiGOpTMgwWpQGZXPW6XaB6GrexBmxg_sQ,15200
38
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=Ek7pZYaSWWFdPKI6iPKiICvZxN7xCVglQleTSSwW5ok,14799
40
+ gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=Wzw6We1KwLcWW9-4tGGMZoPnnHSKofOmCuqoddTHr2Q,14832
39
41
  gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=HzpDRIhe4-XRj9n43bmsXQHxPwx5lcnetxIe5EMbHIE,12924
40
42
  gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=c4H3TOz0LIhWjokCCdQ4asiwQR4_LPyaimo4RAqU9es,9369
41
43
  gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=IPmVAFTtZGvBDb-veoeCq8u7R9chKU958vJoBlWbIvE,17369
@@ -44,6 +46,7 @@ gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=HHwW7i8ryXHI23JZQwscyva6
44
46
  gllm_inference/lm_invoker/lm_invoker.pyi,sha256=B00siZZ7F3i2GuU4nQk3xA8d-h_b37ADzyYBoXarbPA,8033
45
47
  gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=JemahodhaUsC2gsI7YSxnW4X3uX1cU4YCFdIvdWWY88,15203
46
48
  gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=VFMvYXuwMuUHarsu5Xz7tKF6Bx6Ket5HaXZ4-7AtBY0,20011
49
+ gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=6TwO3KU1DBWoe4UAsz97MY1yKBf-N38WjbrBqCmWCNU,15992
47
50
  gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
48
51
  gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=lGJ7xYLchdtv6003Is4GcaKiGdbmIOAzQsaldKG0Aww,1041
49
52
  gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=rB1AWfER2BBKZ5I219211YE2EUFPF25bhzysqjdPgiY,1080
@@ -53,7 +56,6 @@ gllm_inference/lm_invoker/schema/langchain.pyi,sha256=2OJOUQPlGdlUbIOTDOyiWDBOMm
53
56
  gllm_inference/lm_invoker/schema/openai.pyi,sha256=2KZkitU0jxFaR6x2AGe1FtawvxtUgTLDffY9T0Iq9yg,2017
54
57
  gllm_inference/lm_invoker/schema/openai_compatible.pyi,sha256=fVLRIrOvLJjhY7qPUgC3HRFoOFa7XimWLjr2EOo5qmQ,1226
55
58
  gllm_inference/lm_invoker/schema/xai.pyi,sha256=jpC6ZSBDUltzm9GjD6zvSFIPwqizn_ywLnjvwSa7KuU,663
56
- gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=6TwO3KU1DBWoe4UAsz97MY1yKBf-N38WjbrBqCmWCNU,15992
57
59
  gllm_inference/model/__init__.pyi,sha256=JKQB0wVSVYD-_tdRkG7N_oEVAKGCcoBw0BUOUMLieFo,602
58
60
  gllm_inference/model/em/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
59
61
  gllm_inference/model/em/google_em.pyi,sha256=c53H-KNdNOK9ppPLyOSkmCA890eF5FsMd05upkPIzF0,487
@@ -97,8 +99,7 @@ gllm_inference/utils/io_utils.pyi,sha256=Eg7dvHWdXslTKdjh1j3dG50i7r35XG2zTmJ9XXv
97
99
  gllm_inference/utils/langchain.pyi,sha256=4AwFiVAO0ZpdgmqeC4Pb5NJwBt8vVr0MSUqLeCdTscc,1194
98
100
  gllm_inference/utils/validation.pyi,sha256=-RdMmb8afH7F7q4Ao7x6FbwaDfxUHn3hA3WiOgzB-3s,397
99
101
  gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
100
- gllm_inference.cp312-win_amd64.pyd,sha256=QPT_t-K5qqQrk9GfD_HQhFyfNJbkSME_FYTNaWvUvCk,3065856
101
- gllm_inference.pyi,sha256=Pq6P04np3S3x7juGVCzC5sL2im4MsyligEvahVQNWzM,3820
102
- gllm_inference_binary-0.5.24.dist-info/METADATA,sha256=VRF96O0qP_WUnJN5O6eyphnYi4zYXo06tfLHbSlJMtI,4615
103
- gllm_inference_binary-0.5.24.dist-info/WHEEL,sha256=4N0hGcnWMI_Ty6ATf4qJqqSl-UNI-Ln828iTWGIywmU,98
104
- gllm_inference_binary-0.5.24.dist-info/RECORD,,
102
+ gllm_inference_binary-0.5.26.post1.dist-info/METADATA,sha256=FeFY1d9pZiaGA84b6DJ0puaOsqgkZJAwYT6BhEwMLzo,2996
103
+ gllm_inference_binary-0.5.26.post1.dist-info/WHEEL,sha256=x5rgv--I0NI0IT1Lh9tN1VG2cI637p3deednwYLKnxc,96
104
+ gllm_inference_binary-0.5.26.post1.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
105
+ gllm_inference_binary-0.5.26.post1.dist-info/RECORD,,
@@ -1,4 +1,5 @@
1
1
  Wheel-Version: 1.0
2
- Generator: poetry-core 1.9.0
2
+ Generator: Nuitka (2.6.9)
3
3
  Root-Is-Purelib: false
4
4
  Tag: cp312-cp312-win_amd64
5
+
@@ -0,0 +1 @@
1
+ gllm_inference