gllm-inference-binary 0.5.25__cp311-cp311-win_amd64.whl → 0.5.26b1__cp311-cp311-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of gllm-inference-binary might be problematic. Click here for more details.
- gllm_inference.cp311-win_amd64.pyd +0 -0
- {gllm_inference_binary-0.5.25.dist-info → gllm_inference_binary-0.5.26b1.dist-info}/METADATA +114 -110
- {gllm_inference_binary-0.5.25.dist-info → gllm_inference_binary-0.5.26b1.dist-info}/RECORD +12 -11
- {gllm_inference_binary-0.5.25.dist-info → gllm_inference_binary-0.5.26b1.dist-info}/WHEEL +2 -1
- gllm_inference_binary-0.5.26b1.dist-info/top_level.txt +1 -0
|
Binary file
|
{gllm_inference_binary-0.5.25.dist-info → gllm_inference_binary-0.5.26b1.dist-info}/METADATA
RENAMED
|
@@ -1,110 +1,114 @@
|
|
|
1
|
-
Metadata-Version: 2.
|
|
2
|
-
Name: gllm-inference-binary
|
|
3
|
-
Version: 0.5.
|
|
4
|
-
Summary: A library containing components related to model inferences in Gen AI applications.
|
|
5
|
-
Author: Henry Wicaksono
|
|
6
|
-
|
|
7
|
-
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
|
|
19
|
-
|
|
20
|
-
|
|
21
|
-
|
|
22
|
-
Requires-Dist:
|
|
23
|
-
Requires-Dist:
|
|
24
|
-
Requires-Dist:
|
|
25
|
-
Requires-Dist:
|
|
26
|
-
Requires-Dist:
|
|
27
|
-
Requires-Dist:
|
|
28
|
-
Requires-Dist:
|
|
29
|
-
|
|
30
|
-
Requires-Dist:
|
|
31
|
-
|
|
32
|
-
Requires-Dist:
|
|
33
|
-
|
|
34
|
-
Requires-Dist:
|
|
35
|
-
|
|
36
|
-
Requires-Dist:
|
|
37
|
-
|
|
38
|
-
Requires-Dist:
|
|
39
|
-
Requires-Dist:
|
|
40
|
-
|
|
41
|
-
Requires-Dist:
|
|
42
|
-
|
|
43
|
-
Requires-Dist:
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
|
|
47
|
-
|
|
48
|
-
|
|
49
|
-
|
|
50
|
-
|
|
51
|
-
|
|
52
|
-
|
|
53
|
-
|
|
54
|
-
|
|
55
|
-
|
|
56
|
-
|
|
57
|
-
|
|
58
|
-
|
|
59
|
-
|
|
60
|
-
|
|
61
|
-
|
|
62
|
-
|
|
63
|
-
|
|
64
|
-
|
|
65
|
-
|
|
66
|
-
|
|
67
|
-
|
|
68
|
-
|
|
69
|
-
|
|
70
|
-
#### Using
|
|
71
|
-
```bash
|
|
72
|
-
|
|
73
|
-
```
|
|
74
|
-
|
|
75
|
-
|
|
76
|
-
|
|
77
|
-
|
|
78
|
-
|
|
79
|
-
|
|
80
|
-
|
|
81
|
-
|
|
82
|
-
|
|
83
|
-
|
|
84
|
-
|
|
85
|
-
|
|
86
|
-
|
|
87
|
-
- `
|
|
88
|
-
|
|
89
|
-
|
|
90
|
-
|
|
91
|
-
|
|
92
|
-
|
|
93
|
-
|
|
94
|
-
|
|
95
|
-
|
|
96
|
-
|
|
97
|
-
|
|
98
|
-
|
|
99
|
-
|
|
100
|
-
|
|
101
|
-
|
|
102
|
-
|
|
103
|
-
|
|
104
|
-
|
|
105
|
-
|
|
106
|
-
|
|
107
|
-
poetry
|
|
108
|
-
|
|
109
|
-
|
|
110
|
-
|
|
1
|
+
Metadata-Version: 2.2
|
|
2
|
+
Name: gllm-inference-binary
|
|
3
|
+
Version: 0.5.26b1
|
|
4
|
+
Summary: A library containing components related to model inferences in Gen AI applications.
|
|
5
|
+
Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, Resti Febrina <resti.febrina@gdplabs.id>
|
|
6
|
+
Requires-Python: <3.14,>=3.11
|
|
7
|
+
Description-Content-Type: text/markdown
|
|
8
|
+
Requires-Dist: poetry<3.0.0,>=2.1.3
|
|
9
|
+
Requires-Dist: gllm-core-binary<0.4.0,>=0.3.0
|
|
10
|
+
Requires-Dist: aiohttp<4.0.0,>=3.12.14
|
|
11
|
+
Requires-Dist: filetype<2.0.0,>=1.2.0
|
|
12
|
+
Requires-Dist: httpx<0.29.0,>=0.28.0
|
|
13
|
+
Requires-Dist: jinja2<4.0.0,>=3.1.4
|
|
14
|
+
Requires-Dist: jsonschema<5.0.0,>=4.24.0
|
|
15
|
+
Requires-Dist: langchain<0.4.0,>=0.3.0
|
|
16
|
+
Requires-Dist: pandas<3.0.0,>=2.2.3
|
|
17
|
+
Requires-Dist: protobuf<7.0.0,>=6.0.0
|
|
18
|
+
Requires-Dist: python-magic<0.5.0,>=0.4.27; sys_platform != "win32"
|
|
19
|
+
Requires-Dist: python-magic-bin<0.5.0,>=0.4.14; sys_platform == "win32"
|
|
20
|
+
Requires-Dist: sentencepiece<0.3.0,>=0.2.0
|
|
21
|
+
Provides-Extra: dev
|
|
22
|
+
Requires-Dist: coverage<8.0.0,>=7.4.4; extra == "dev"
|
|
23
|
+
Requires-Dist: mypy<2.0.0,>=1.15.0; extra == "dev"
|
|
24
|
+
Requires-Dist: pre-commit<4.0.0,>=3.7.0; extra == "dev"
|
|
25
|
+
Requires-Dist: pytest<9.0.0,>=8.1.1; extra == "dev"
|
|
26
|
+
Requires-Dist: pytest-asyncio<1.0.0,>=0.23.6; extra == "dev"
|
|
27
|
+
Requires-Dist: pytest-cov<6.0.0,>=5.0.0; extra == "dev"
|
|
28
|
+
Requires-Dist: ruff<1.0.0,>=0.6.7; extra == "dev"
|
|
29
|
+
Provides-Extra: anthropic
|
|
30
|
+
Requires-Dist: anthropic<0.61.0,>=0.60.0; extra == "anthropic"
|
|
31
|
+
Provides-Extra: bedrock
|
|
32
|
+
Requires-Dist: aioboto3<16.0.0,>=15.0.0; extra == "bedrock"
|
|
33
|
+
Provides-Extra: datasaur
|
|
34
|
+
Requires-Dist: openai<2.0.0,>=1.98.0; extra == "datasaur"
|
|
35
|
+
Provides-Extra: google
|
|
36
|
+
Requires-Dist: google-genai<=1.36,>=1.23; extra == "google"
|
|
37
|
+
Provides-Extra: huggingface
|
|
38
|
+
Requires-Dist: huggingface-hub<0.31.0,>=0.30.0; extra == "huggingface"
|
|
39
|
+
Requires-Dist: transformers==4.52.4; extra == "huggingface"
|
|
40
|
+
Provides-Extra: openai
|
|
41
|
+
Requires-Dist: openai<2.0.0,>=1.98.0; extra == "openai"
|
|
42
|
+
Provides-Extra: litellm
|
|
43
|
+
Requires-Dist: litellm<2.0.0,>=1.69.2; extra == "litellm"
|
|
44
|
+
Provides-Extra: twelvelabs
|
|
45
|
+
Requires-Dist: twelvelabs<0.5.0,>=0.4.4; extra == "twelvelabs"
|
|
46
|
+
Provides-Extra: voyage
|
|
47
|
+
Requires-Dist: voyageai<0.4.0,>=0.3.0; python_version < "3.13" and extra == "voyage"
|
|
48
|
+
Provides-Extra: xai
|
|
49
|
+
Requires-Dist: xai_sdk<2.0.0,>=1.0.0; extra == "xai"
|
|
50
|
+
|
|
51
|
+
# GLLM Inference
|
|
52
|
+
|
|
53
|
+
## Description
|
|
54
|
+
|
|
55
|
+
A library containing components related to model inferences in Gen AI applications.
|
|
56
|
+
|
|
57
|
+
## Installation
|
|
58
|
+
|
|
59
|
+
### Prerequisites
|
|
60
|
+
- Python 3.11+ - [Install here](https://www.python.org/downloads/)
|
|
61
|
+
- Pip (if using Pip) - [Install here](https://pip.pypa.io/en/stable/installation/)
|
|
62
|
+
- Poetry 1.8.1+ (if using Poetry) - [Install here](https://python-poetry.org/docs/#installation)
|
|
63
|
+
- Git (if using Git) - [Install here](https://git-scm.com/downloads)
|
|
64
|
+
- For git installation:
|
|
65
|
+
- Access to the [GDP Labs SDK github repository](https://github.com/GDP-ADMIN/gen-ai-internal)
|
|
66
|
+
|
|
67
|
+
### 1. Installation from Artifact Registry
|
|
68
|
+
Choose one of the following methods to install the package:
|
|
69
|
+
|
|
70
|
+
#### Using pip
|
|
71
|
+
```bash
|
|
72
|
+
pip install gllm-inference-binary
|
|
73
|
+
```
|
|
74
|
+
|
|
75
|
+
#### Using Poetry
|
|
76
|
+
```bash
|
|
77
|
+
poetry add gllm-inference-binary
|
|
78
|
+
```
|
|
79
|
+
|
|
80
|
+
### 2. Development Installation (Git)
|
|
81
|
+
For development purposes, you can install directly from the Git repository:
|
|
82
|
+
```bash
|
|
83
|
+
poetry add "git+ssh://git@github.com/GDP-ADMIN/gen-ai-internal.git#subdirectory=libs/gllm-inference"
|
|
84
|
+
```
|
|
85
|
+
|
|
86
|
+
Available extras:
|
|
87
|
+
- `anthropic`: Install Anthropic models dependencies
|
|
88
|
+
- `google-genai`: Install Google Generative AI models dependencies
|
|
89
|
+
- `google-vertexai`: Install Google Vertex AI models dependencies
|
|
90
|
+
- `huggingface`: Install HuggingFace models dependencies
|
|
91
|
+
- `openai`: Install OpenAI models dependencies
|
|
92
|
+
- `twelvelabs`: Install TwelveLabs models dependencies
|
|
93
|
+
|
|
94
|
+
## Managing Dependencies
|
|
95
|
+
1. Go to root folder of `gllm-inference` module, e.g. `cd libs/gllm-inference`.
|
|
96
|
+
2. Run `poetry shell` to create a virtual environment.
|
|
97
|
+
3. Run `poetry lock` to create a lock file if you haven't done it yet.
|
|
98
|
+
4. Run `poetry install` to install the `gllm-inference` requirements for the first time.
|
|
99
|
+
5. Run `poetry update` if you update any dependency module version at `pyproject.toml`.
|
|
100
|
+
|
|
101
|
+
## Contributing
|
|
102
|
+
Please refer to this [Python Style Guide](https://docs.google.com/document/d/1uRggCrHnVfDPBnG641FyQBwUwLoFw0kTzNqRm92vUwM/edit?usp=sharing)
|
|
103
|
+
to get information about code style, documentation standard, and SCA that you need to use when contributing to this project
|
|
104
|
+
|
|
105
|
+
1. Activate `pre-commit` hooks using `pre-commit install`
|
|
106
|
+
2. Run `poetry shell` to create a virtual environment.
|
|
107
|
+
3. Run `poetry lock` to create a lock file if you haven't done it yet.
|
|
108
|
+
4. Run `poetry install` to install the `gllm-inference` requirements for the first time.
|
|
109
|
+
5. Run `which python` to get the path to be referenced at Visual Studio Code interpreter path (`Ctrl`+`Shift`+`P` or `Cmd`+`Shift`+`P`)
|
|
110
|
+
6. Try running the unit test to see if it's working:
|
|
111
|
+
```bash
|
|
112
|
+
poetry run pytest -s tests/unit_tests/
|
|
113
|
+
```
|
|
114
|
+
|
|
@@ -1,4 +1,7 @@
|
|
|
1
|
+
gllm_inference.cp311-win_amd64.pyd,sha256=PUthY8eDvMpBT2O45s5iga9V3aj5YwdE5tqR82i5vG0,3077632
|
|
2
|
+
gllm_inference.pyi,sha256=Pq6P04np3S3x7juGVCzC5sL2im4MsyligEvahVQNWzM,3820
|
|
1
3
|
gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
4
|
+
gllm_inference/constants.pyi,sha256=KQmondDEkHK2P249ymmce3SdutVrx8kYm4v1eTCkW9U,277
|
|
2
5
|
gllm_inference/builder/__init__.pyi,sha256=-bw1uDx7CAM7pkvjvb1ZXku9zXlQ7aEAyC83KIn3bz8,506
|
|
3
6
|
gllm_inference/builder/build_em_invoker.pyi,sha256=Mh1vRoJhsqc8hX4jUdopV14Fn44ql27NB7xbGjoHJtE,6020
|
|
4
7
|
gllm_inference/builder/build_lm_invoker.pyi,sha256=p63iuVBOOpNizItGK6HDxYDrgXdovtfSe0VrvrEd-PA,7047
|
|
@@ -8,17 +11,18 @@ gllm_inference/catalog/__init__.pyi,sha256=HWgPKWIzprpMHRKe_qN9BZSIQhVhrqiyjLjIX
|
|
|
8
11
|
gllm_inference/catalog/catalog.pyi,sha256=eWPqgQKi-SJGHabi_XOTEKpAj96OSRypKsb5ZEC1VWU,4911
|
|
9
12
|
gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=GemCEjFRHNChtNOfbyXSVsJiA3klOCAe_X11fnymhYs,5540
|
|
10
13
|
gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=iViWB4SaezzjQY4UY1YxeoXUNxqxa2cTJGaD9JSx4Q8,3279
|
|
11
|
-
gllm_inference/constants.pyi,sha256=KQmondDEkHK2P249ymmce3SdutVrx8kYm4v1eTCkW9U,277
|
|
12
14
|
gllm_inference/em_invoker/__init__.pyi,sha256=pmbsjmsqXwfe4WPykMnrmasKrYuylJWnf2s0pbo0ioM,997
|
|
13
15
|
gllm_inference/em_invoker/azure_openai_em_invoker.pyi,sha256=SfJPC_PJGiEfWS9JH5kRQPJztsR7jRhwVuETqdY-JsQ,5021
|
|
14
16
|
gllm_inference/em_invoker/bedrock_em_invoker.pyi,sha256=UqodtpDmE7fEgpctXEETIlZGorX9i1lmmuTvGaJke6o,5829
|
|
15
17
|
gllm_inference/em_invoker/em_invoker.pyi,sha256=YDYJ8TGScsz5Gg-OBnEENN1tI1RYvwoddypxUr6SAWw,5191
|
|
16
18
|
gllm_inference/em_invoker/google_em_invoker.pyi,sha256=q69kdVuE44ZqziQ8BajFYZ1tYn-MPjKjzXS9cRh4oAo,6951
|
|
17
|
-
gllm_inference/em_invoker/langchain/__init__.pyi,sha256=aOTlRvS9aG1tBErjsmhe75s4Sq-g2z9ArfGqNW7QyEs,151
|
|
18
|
-
gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi,sha256=BBSDazMOckO9Aw17tC3LGUTPqLb01my1xUZLtKZlwJY,3388
|
|
19
19
|
gllm_inference/em_invoker/langchain_em_invoker.pyi,sha256=nhX6LynrjhfySEt_44OlLoSBd15hoz3giWyNM9CYLKY,3544
|
|
20
20
|
gllm_inference/em_invoker/openai_compatible_em_invoker.pyi,sha256=zEYOBDXKQhvcMGer9DYDu50_3KRDjYyN8-JgpBIFPOI,5456
|
|
21
21
|
gllm_inference/em_invoker/openai_em_invoker.pyi,sha256=0TDIQa-5UwsPcVxgkze-QJJWrt-ToakAKbuAk9TW5SM,4746
|
|
22
|
+
gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=MMVgSnjMXksdhSDXIi3vOULIXnjbhtq19eR5LPnUmGo,5446
|
|
23
|
+
gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=vdB_qS8QKrCcb-HtXwKZS4WW1R1wGzpMBFmOKC39sjU,5619
|
|
24
|
+
gllm_inference/em_invoker/langchain/__init__.pyi,sha256=aOTlRvS9aG1tBErjsmhe75s4Sq-g2z9ArfGqNW7QyEs,151
|
|
25
|
+
gllm_inference/em_invoker/langchain/em_invoker_embeddings.pyi,sha256=BBSDazMOckO9Aw17tC3LGUTPqLb01my1xUZLtKZlwJY,3388
|
|
22
26
|
gllm_inference/em_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
23
27
|
gllm_inference/em_invoker/schema/bedrock.pyi,sha256=HoNgVi0T21aFd1JrCnSLu4yryv8k8RnYdR3-tIdHFgA,498
|
|
24
28
|
gllm_inference/em_invoker/schema/google.pyi,sha256=bzdtu4DFH2kATLybIeNl_Lznj99H-6u2Fvx3Zx52oZg,190
|
|
@@ -27,8 +31,6 @@ gllm_inference/em_invoker/schema/openai.pyi,sha256=rNRqN62y5wHOKlr4T0n0m41ikAnSr
|
|
|
27
31
|
gllm_inference/em_invoker/schema/openai_compatible.pyi,sha256=A9MOeBhI-IPuvewOk4YYOAGtgyKohERx6-9cEYtbwvs,157
|
|
28
32
|
gllm_inference/em_invoker/schema/twelvelabs.pyi,sha256=D3F9_1F-UTzE6Ymxj6u0IFdL6OFVGlc7noZJr3iuA6I,389
|
|
29
33
|
gllm_inference/em_invoker/schema/voyage.pyi,sha256=Aqvu6mhFkNb01aXAI5mChLKIgEnFnr-jNKq1lVWB54M,304
|
|
30
|
-
gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=MMVgSnjMXksdhSDXIi3vOULIXnjbhtq19eR5LPnUmGo,5446
|
|
31
|
-
gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=vdB_qS8QKrCcb-HtXwKZS4WW1R1wGzpMBFmOKC39sjU,5619
|
|
32
34
|
gllm_inference/exceptions/__init__.pyi,sha256=nXOqwsuwUgsnBcJEANVuxbZ1nDfcJ6-pKUfKeZwltkk,1218
|
|
33
35
|
gllm_inference/exceptions/error_parser.pyi,sha256=4aiJZhBzBOqlhdmpvaCvildGy7_XxlJzQpe3PzGt8eE,2040
|
|
34
36
|
gllm_inference/exceptions/exceptions.pyi,sha256=6y3ECgHAStqMGgQv8Dv-Ui-5PDD07mSj6qaRZeSWea4,5857
|
|
@@ -44,6 +46,7 @@ gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=HHwW7i8ryXHI23JZQwscyva6
|
|
|
44
46
|
gllm_inference/lm_invoker/lm_invoker.pyi,sha256=B00siZZ7F3i2GuU4nQk3xA8d-h_b37ADzyYBoXarbPA,8033
|
|
45
47
|
gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=JemahodhaUsC2gsI7YSxnW4X3uX1cU4YCFdIvdWWY88,15203
|
|
46
48
|
gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=VFMvYXuwMuUHarsu5Xz7tKF6Bx6Ket5HaXZ4-7AtBY0,20011
|
|
49
|
+
gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=6TwO3KU1DBWoe4UAsz97MY1yKBf-N38WjbrBqCmWCNU,15992
|
|
47
50
|
gllm_inference/lm_invoker/schema/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
48
51
|
gllm_inference/lm_invoker/schema/anthropic.pyi,sha256=lGJ7xYLchdtv6003Is4GcaKiGdbmIOAzQsaldKG0Aww,1041
|
|
49
52
|
gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=rB1AWfER2BBKZ5I219211YE2EUFPF25bhzysqjdPgiY,1080
|
|
@@ -53,7 +56,6 @@ gllm_inference/lm_invoker/schema/langchain.pyi,sha256=2OJOUQPlGdlUbIOTDOyiWDBOMm
|
|
|
53
56
|
gllm_inference/lm_invoker/schema/openai.pyi,sha256=2KZkitU0jxFaR6x2AGe1FtawvxtUgTLDffY9T0Iq9yg,2017
|
|
54
57
|
gllm_inference/lm_invoker/schema/openai_compatible.pyi,sha256=fVLRIrOvLJjhY7qPUgC3HRFoOFa7XimWLjr2EOo5qmQ,1226
|
|
55
58
|
gllm_inference/lm_invoker/schema/xai.pyi,sha256=jpC6ZSBDUltzm9GjD6zvSFIPwqizn_ywLnjvwSa7KuU,663
|
|
56
|
-
gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=6TwO3KU1DBWoe4UAsz97MY1yKBf-N38WjbrBqCmWCNU,15992
|
|
57
59
|
gllm_inference/model/__init__.pyi,sha256=JKQB0wVSVYD-_tdRkG7N_oEVAKGCcoBw0BUOUMLieFo,602
|
|
58
60
|
gllm_inference/model/em/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
59
61
|
gllm_inference/model/em/google_em.pyi,sha256=c53H-KNdNOK9ppPLyOSkmCA890eF5FsMd05upkPIzF0,487
|
|
@@ -97,8 +99,7 @@ gllm_inference/utils/io_utils.pyi,sha256=Eg7dvHWdXslTKdjh1j3dG50i7r35XG2zTmJ9XXv
|
|
|
97
99
|
gllm_inference/utils/langchain.pyi,sha256=4AwFiVAO0ZpdgmqeC4Pb5NJwBt8vVr0MSUqLeCdTscc,1194
|
|
98
100
|
gllm_inference/utils/validation.pyi,sha256=-RdMmb8afH7F7q4Ao7x6FbwaDfxUHn3hA3WiOgzB-3s,397
|
|
99
101
|
gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
|
|
100
|
-
|
|
101
|
-
|
|
102
|
-
gllm_inference_binary-0.5.
|
|
103
|
-
gllm_inference_binary-0.5.
|
|
104
|
-
gllm_inference_binary-0.5.25.dist-info/RECORD,,
|
|
102
|
+
gllm_inference_binary-0.5.26b1.dist-info/METADATA,sha256=eMDdk3EK3o3Ag2h49phs0uKaDhkT338rJU9ZlwLvFEk,4964
|
|
103
|
+
gllm_inference_binary-0.5.26b1.dist-info/WHEEL,sha256=l2aKBREYfqJ7T2ljmr6hUiXPoNvvXF47bG4IHjuSyS4,96
|
|
104
|
+
gllm_inference_binary-0.5.26b1.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
|
|
105
|
+
gllm_inference_binary-0.5.26b1.dist-info/RECORD,,
|
|
@@ -0,0 +1 @@
|
|
|
1
|
+
gllm_inference
|