gllm-inference-binary 0.5.56__cp312-cp312-win_amd64.whl → 0.5.58__cp312-cp312-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- gllm_inference/lm_invoker/google_lm_invoker.pyi +2 -2
- gllm_inference/lm_invoker/openai_lm_invoker.pyi +32 -1
- gllm_inference/lm_invoker/schema/openai.pyi +2 -0
- gllm_inference/schema/attachment.pyi +19 -5
- gllm_inference.cp312-win_amd64.pyd +0 -0
- gllm_inference.pyi +1 -0
- {gllm_inference_binary-0.5.56.dist-info → gllm_inference_binary-0.5.58.dist-info}/METADATA +1 -1
- {gllm_inference_binary-0.5.56.dist-info → gllm_inference_binary-0.5.58.dist-info}/RECORD +10 -10
- {gllm_inference_binary-0.5.56.dist-info → gllm_inference_binary-0.5.58.dist-info}/WHEEL +0 -0
- {gllm_inference_binary-0.5.56.dist-info → gllm_inference_binary-0.5.58.dist-info}/top_level.txt +0 -0
|
@@ -31,10 +31,10 @@ class GoogleLMInvoker(BaseLMInvoker):
|
|
|
31
31
|
structured response as defined by the schema. Supports both Pydantic BaseModel and JSON schema dictionary.
|
|
32
32
|
output_analytics (bool): Whether to output the invocation analytics.
|
|
33
33
|
retry_config (RetryConfig | None): The retry configuration for the language model.
|
|
34
|
-
generate_image (bool): Whether to generate image. Only allowed for image generation models.
|
|
35
34
|
thinking (bool): Whether to enable thinking. Only allowed for thinking models.
|
|
36
35
|
thinking_budget (int): The tokens allowed for thinking process. Only allowed for thinking models.
|
|
37
36
|
If set to -1, the model will control the budget automatically.
|
|
37
|
+
image_generation (bool): Whether to generate image. Only allowed for image generation models.
|
|
38
38
|
|
|
39
39
|
Basic usage:
|
|
40
40
|
The `GoogleLMInvoker` can be used as follows:
|
|
@@ -262,7 +262,7 @@ class GoogleLMInvoker(BaseLMInvoker):
|
|
|
262
262
|
```
|
|
263
263
|
'''
|
|
264
264
|
client_params: Incomplete
|
|
265
|
-
|
|
265
|
+
image_generation: Incomplete
|
|
266
266
|
thinking: Incomplete
|
|
267
267
|
thinking_budget: Incomplete
|
|
268
268
|
def __init__(self, model_name: str, api_key: str | None = None, credentials_path: str | None = None, project_id: str | None = None, location: str = 'us-central1', model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, thinking: bool | None = None, thinking_budget: int = ..., simplify_events: bool = False) -> None:
|
|
@@ -36,6 +36,7 @@ class OpenAILMInvoker(BaseLMInvoker):
|
|
|
36
36
|
for non-reasoning models. If None, the model will perform medium reasoning effort.
|
|
37
37
|
reasoning_summary (ReasoningSummary | None): The reasoning summary level for reasoning models. Not allowed
|
|
38
38
|
for non-reasoning models. If None, no summary will be generated.
|
|
39
|
+
image_generation (bool): Whether to enable image generation.
|
|
39
40
|
mcp_servers (list[MCPServer]): The list of MCP servers to enable MCP tool calling.
|
|
40
41
|
code_interpreter (bool): Whether to enable the code interpreter.
|
|
41
42
|
web_search (bool): Whether to enable the web search.
|
|
@@ -123,6 +124,35 @@ class OpenAILMInvoker(BaseLMInvoker):
|
|
|
123
124
|
|
|
124
125
|
When structured output is enabled, streaming is disabled.
|
|
125
126
|
|
|
127
|
+
Image generation:
|
|
128
|
+
The `OpenAILMInvoker` can be configured to generate images.
|
|
129
|
+
This feature can be enabled by setting the `image_generation` parameter to `True`.
|
|
130
|
+
|
|
131
|
+
Image outputs are stored in the `outputs` attribute of the `LMOutput` object and can be accessed
|
|
132
|
+
via the `attachments` property.
|
|
133
|
+
|
|
134
|
+
Usage example:
|
|
135
|
+
```python
|
|
136
|
+
lm_invoker = OpenAILMInvoker(..., image_generation=True)
|
|
137
|
+
result = await lm_invoker.invoke("Create a picture...")
|
|
138
|
+
result.attachments[0].write_to_file("path/to/local/image.png")
|
|
139
|
+
```
|
|
140
|
+
|
|
141
|
+
Output example:
|
|
142
|
+
```python
|
|
143
|
+
LMOutput(
|
|
144
|
+
outputs=[
|
|
145
|
+
LMOutputItem(
|
|
146
|
+
type="attachment",
|
|
147
|
+
output=Attachment(filename="image.png", mime_type="image/png", data=b"..."),
|
|
148
|
+
),
|
|
149
|
+
],
|
|
150
|
+
)
|
|
151
|
+
```
|
|
152
|
+
|
|
153
|
+
When image generation is enabled, streaming is disabled.
|
|
154
|
+
Image generation is only available for certain models.
|
|
155
|
+
|
|
126
156
|
Tool calling:
|
|
127
157
|
The `OpenAILMInvoker` can be configured to call tools to perform certain tasks.
|
|
128
158
|
This feature can be enabled by providing a list of `Tool` objects to the `tools` parameter.
|
|
@@ -354,7 +384,7 @@ class OpenAILMInvoker(BaseLMInvoker):
|
|
|
354
384
|
```
|
|
355
385
|
'''
|
|
356
386
|
client_kwargs: Incomplete
|
|
357
|
-
def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None, mcp_servers: list[MCPServer] | None = None, code_interpreter: bool = False, web_search: bool = False, simplify_events: bool = False) -> None:
|
|
387
|
+
def __init__(self, model_name: str, api_key: str | None = None, base_url: str = ..., model_kwargs: dict[str, Any] | None = None, default_hyperparameters: dict[str, Any] | None = None, tools: list[Tool | LangChainTool] | None = None, response_schema: ResponseSchema | None = None, output_analytics: bool = False, retry_config: RetryConfig | None = None, reasoning_effort: ReasoningEffort | None = None, reasoning_summary: ReasoningSummary | None = None, image_generation: bool = False, mcp_servers: list[MCPServer] | None = None, code_interpreter: bool = False, web_search: bool = False, simplify_events: bool = False) -> None:
|
|
358
388
|
'''Initializes a new instance of the OpenAILMInvoker class.
|
|
359
389
|
|
|
360
390
|
Args:
|
|
@@ -379,6 +409,7 @@ class OpenAILMInvoker(BaseLMInvoker):
|
|
|
379
409
|
for non-reasoning models. If None, the model will perform medium reasoning effort. Defaults to None.
|
|
380
410
|
reasoning_summary (ReasoningSummary | None, optional): The reasoning summary level for reasoning models.
|
|
381
411
|
Not allowed for non-reasoning models. If None, no summary will be generated. Defaults to None.
|
|
412
|
+
image_generation (bool, optional): Whether to enable image generation. Defaults to False.
|
|
382
413
|
mcp_servers (list[MCPServer] | None, optional): The MCP servers containing tools to be accessed by the
|
|
383
414
|
language model. Defaults to None.
|
|
384
415
|
code_interpreter (bool, optional): Whether to enable the code interpreter. Defaults to False.
|
|
@@ -18,6 +18,7 @@ class Key:
|
|
|
18
18
|
FILENAME: str
|
|
19
19
|
FORMAT: str
|
|
20
20
|
ID: str
|
|
21
|
+
IMAGE_GENERATION_CALL: str
|
|
21
22
|
IMAGE_URL: str
|
|
22
23
|
INCLUDE: str
|
|
23
24
|
INCOMPLETE_DETAILS: str
|
|
@@ -54,6 +55,7 @@ class InputType:
|
|
|
54
55
|
FUNCTION: str
|
|
55
56
|
FUNCTION_CALL: str
|
|
56
57
|
FUNCTION_CALL_OUTPUT: str
|
|
58
|
+
IMAGE_GENERATION: str
|
|
57
59
|
INPUT_FILE: str
|
|
58
60
|
INPUT_IMAGE: str
|
|
59
61
|
INPUT_TEXT: str
|
|
@@ -1,7 +1,9 @@
|
|
|
1
1
|
from _typeshed import Incomplete
|
|
2
2
|
from pydantic import BaseModel
|
|
3
|
+
from typing import Any
|
|
3
4
|
|
|
4
5
|
HEX_REPR_LENGTH: int
|
|
6
|
+
METADATA_ITEM_REPR_LENGTH: int
|
|
5
7
|
logger: Incomplete
|
|
6
8
|
|
|
7
9
|
class Attachment(BaseModel):
|
|
@@ -13,68 +15,80 @@ class Attachment(BaseModel):
|
|
|
13
15
|
mime_type (str): The mime type of the file attachment.
|
|
14
16
|
extension (str): The extension of the file attachment.
|
|
15
17
|
url (str | None): The URL of the file attachment. Defaults to None.
|
|
18
|
+
metadata (dict[str, Any]): The metadata of the file attachment. Defaults to an empty dictionary.
|
|
16
19
|
"""
|
|
17
20
|
data: bytes
|
|
18
21
|
filename: str
|
|
19
22
|
mime_type: str
|
|
20
23
|
extension: str
|
|
21
24
|
url: str | None
|
|
25
|
+
metadata: dict[str, Any]
|
|
22
26
|
@classmethod
|
|
23
|
-
def from_bytes(cls, bytes: bytes, filename: str | None = None) -> Attachment:
|
|
27
|
+
def from_bytes(cls, bytes: bytes, filename: str | None = None, metadata: dict[str, Any] | None = None) -> Attachment:
|
|
24
28
|
"""Creates an Attachment from bytes.
|
|
25
29
|
|
|
26
30
|
Args:
|
|
27
31
|
bytes (bytes): The bytes of the file.
|
|
28
32
|
filename (str | None, optional): The filename of the file. Defaults to None,
|
|
29
33
|
in which case the filename will be derived from the extension.
|
|
34
|
+
metadata (dict[str, Any] | None, optional): The metadata of the file attachment. Defaults to None,
|
|
35
|
+
in which case an empty dictionary will be used.
|
|
30
36
|
|
|
31
37
|
Returns:
|
|
32
38
|
Attachment: The instantiated Attachment.
|
|
33
39
|
"""
|
|
34
40
|
@classmethod
|
|
35
|
-
def from_base64(cls, base64_data: str, filename: str | None = None) -> Attachment:
|
|
41
|
+
def from_base64(cls, base64_data: str, filename: str | None = None, metadata: dict[str, Any] | None = None) -> Attachment:
|
|
36
42
|
"""Creates an Attachment from a base64 string.
|
|
37
43
|
|
|
38
44
|
Args:
|
|
39
45
|
base64_data (str): The base64 string of the file.
|
|
40
46
|
filename (str | None, optional): The filename of the file. Defaults to None,
|
|
41
47
|
in which case the filename will be derived from the mime type.
|
|
48
|
+
metadata (dict[str, Any] | None, optional): The metadata of the file attachment. Defaults to None,
|
|
49
|
+
in which case an empty dictionary will be used.
|
|
42
50
|
|
|
43
51
|
Returns:
|
|
44
52
|
Attachment: The instantiated Attachment.
|
|
45
53
|
"""
|
|
46
54
|
@classmethod
|
|
47
|
-
def from_data_url(cls, data_url: str, filename: str | None = None) -> Attachment:
|
|
55
|
+
def from_data_url(cls, data_url: str, filename: str | None = None, metadata: dict[str, Any] | None = None) -> Attachment:
|
|
48
56
|
"""Creates an Attachment from a data URL (data:[mime/type];base64,[bytes]).
|
|
49
57
|
|
|
50
58
|
Args:
|
|
51
59
|
data_url (str): The data URL of the file.
|
|
52
60
|
filename (str | None, optional): The filename of the file. Defaults to None,
|
|
53
61
|
in which case the filename will be derived from the mime type.
|
|
62
|
+
metadata (dict[str, Any] | None, optional): The metadata of the file attachment. Defaults to None,
|
|
63
|
+
in which case an empty dictionary will be used.
|
|
54
64
|
|
|
55
65
|
Returns:
|
|
56
66
|
Attachment: The instantiated Attachment.
|
|
57
67
|
"""
|
|
58
68
|
@classmethod
|
|
59
|
-
def from_url(cls, url: str, filename: str | None = None) -> Attachment:
|
|
69
|
+
def from_url(cls, url: str, filename: str | None = None, metadata: dict[str, Any] | None = None) -> Attachment:
|
|
60
70
|
"""Creates an Attachment from a URL.
|
|
61
71
|
|
|
62
72
|
Args:
|
|
63
73
|
url (str): The URL of the file.
|
|
64
74
|
filename (str | None, optional): The filename of the file. Defaults to None,
|
|
65
75
|
in which case the filename will be derived from the URL.
|
|
76
|
+
metadata (dict[str, Any] | None, optional): The metadata of the file attachment. Defaults to None,
|
|
77
|
+
in which case an empty dictionary will be used.
|
|
66
78
|
|
|
67
79
|
Returns:
|
|
68
80
|
Attachment: The instantiated Attachment.
|
|
69
81
|
"""
|
|
70
82
|
@classmethod
|
|
71
|
-
def from_path(cls, path: str, filename: str | None = None) -> Attachment:
|
|
83
|
+
def from_path(cls, path: str, filename: str | None = None, metadata: dict[str, Any] | None = None) -> Attachment:
|
|
72
84
|
"""Creates an Attachment from a path.
|
|
73
85
|
|
|
74
86
|
Args:
|
|
75
87
|
path (str): The path to the file.
|
|
76
88
|
filename (str | None, optional): The filename of the file. Defaults to None,
|
|
77
89
|
in which case the filename will be derived from the path.
|
|
90
|
+
metadata (dict[str, Any] | None, optional): The metadata of the file attachment. Defaults to None,
|
|
91
|
+
in which case an empty dictionary will be used.
|
|
78
92
|
|
|
79
93
|
Returns:
|
|
80
94
|
Attachment: The instantiated Attachment.
|
|
Binary file
|
gllm_inference.pyi
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.2
|
|
2
2
|
Name: gllm-inference-binary
|
|
3
|
-
Version: 0.5.
|
|
3
|
+
Version: 0.5.58
|
|
4
4
|
Summary: A library containing components related to model inferences in Gen AI applications.
|
|
5
5
|
Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, Resti Febrina <resti.febrina@gdplabs.id>
|
|
6
6
|
Requires-Python: <3.14,>=3.11
|
|
@@ -1,5 +1,5 @@
|
|
|
1
|
-
gllm_inference.cp312-win_amd64.pyd,sha256=
|
|
2
|
-
gllm_inference.pyi,sha256=
|
|
1
|
+
gllm_inference.cp312-win_amd64.pyd,sha256=gXu-_J8DNAWsilArVoVdmEbfp-xR-FoOlVTddWx7xDI,3974144
|
|
2
|
+
gllm_inference.pyi,sha256=t6lCcDjutByvoSK439LdRhQ9i7dq9tmp0Vwr5aAuv4Y,5077
|
|
3
3
|
gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
4
4
|
gllm_inference/constants.pyi,sha256=8jIYOyxJYVWUYXSXF3vag9HhHwjq1iU9tzPiosRHkWk,328
|
|
5
5
|
gllm_inference/builder/__init__.pyi,sha256=-bw1uDx7CAM7pkvjvb1ZXku9zXlQ7aEAyC83KIn3bz8,506
|
|
@@ -45,13 +45,13 @@ gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=_Dst_88LOpC-FN01hApihx
|
|
|
45
45
|
gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=uV98H2nJsElCTsxAuInZ9KSk1jOTq6SROAGQRPR-_r0,13173
|
|
46
46
|
gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=qXmFK6zsOM3nPfueEhY5pAfG24bZytA1jqemPa63vLY,10951
|
|
47
47
|
gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=FnpayOW_Zi0pWFSawLX8XahEnknbnpsRWrkhKZe8Y3U,8035
|
|
48
|
-
gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=
|
|
48
|
+
gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=8U68jWVxQKy6DOn-wJAoqyBEPhTn-Hx7133MQ9F6b28,16690
|
|
49
49
|
gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=ull3cX-iUT4hYMbixcxqfrNUxR8ZoR4Vt9ACVILQWSM,12126
|
|
50
50
|
gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=qG8pPTiDJZR2e7wr5Q2VyceC227tz3QybX3UPihT5ng,11400
|
|
51
51
|
gllm_inference/lm_invoker/lm_invoker.pyi,sha256=L_PHRCeHo0dNs6BjnB8H29irGib-qhxKYf7F7pZlU0E,8652
|
|
52
52
|
gllm_inference/lm_invoker/openai_chat_completions_lm_invoker.pyi,sha256=qt9DAdJM7YBB4op-6SOJB0kCouPYVxtIamGUXLGLUeA,13888
|
|
53
53
|
gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=T9sShA_9fgEuaaAuT2gJZq_EYNbEhf3IkWwMCwfszY8,4244
|
|
54
|
-
gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=
|
|
54
|
+
gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=ReU37hrmYZFbLfCD_c14ryRgnfpPC2YyDx2S5Ft_tXQ,22747
|
|
55
55
|
gllm_inference/lm_invoker/portkey_lm_invoker.pyi,sha256=FYOp4BaDfOtompWIRhDqzMVVSK-TiFyw7JA4TznANQE,15236
|
|
56
56
|
gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=6beZsQjGUTo7TdzWBWksRzVGT58XyipErpGfiRq6NH0,13017
|
|
57
57
|
gllm_inference/lm_invoker/batch/__init__.pyi,sha256=vJOTHRJ83oq8Bq0UsMdID9_HW5JAxr06gUs4aPRZfEE,130
|
|
@@ -62,7 +62,7 @@ gllm_inference/lm_invoker/schema/bedrock.pyi,sha256=rB1AWfER2BBKZ5I219211YE2EUFP
|
|
|
62
62
|
gllm_inference/lm_invoker/schema/datasaur.pyi,sha256=8lmb1PRbkqBsF_l7iOffxW0K5Xxpi69GW9Z7KxyxHTE,279
|
|
63
63
|
gllm_inference/lm_invoker/schema/google.pyi,sha256=elXHrUMS46pbTsulk7hBXVVFcT022iD-_U_I590xeV8,529
|
|
64
64
|
gllm_inference/lm_invoker/schema/langchain.pyi,sha256=2OJOUQPlGdlUbIOTDOyiWDBOMm3MoVX-kU2nK0zQsF0,452
|
|
65
|
-
gllm_inference/lm_invoker/schema/openai.pyi,sha256=
|
|
65
|
+
gllm_inference/lm_invoker/schema/openai.pyi,sha256=GIrqEtUPinn8VD-w-38gOw0qiIYuVzM9cj5dRYuGIoQ,2387
|
|
66
66
|
gllm_inference/lm_invoker/schema/openai_chat_completions.pyi,sha256=nNPb7ETC9IrJwkV5wfbGf6Co3-qdq4lhcXz0l_qYCE4,1261
|
|
67
67
|
gllm_inference/lm_invoker/schema/portkey.pyi,sha256=V2q4JIwDAR7BidqfmO01u1_1mLOMtm5OCon6sN2zNt0,662
|
|
68
68
|
gllm_inference/lm_invoker/schema/xai.pyi,sha256=jpC6ZSBDUltzm9GjD6zvSFIPwqizn_ywLnjvwSa7KuU,663
|
|
@@ -111,7 +111,7 @@ gllm_inference/request_processor/lm_request_processor.pyi,sha256=0fy1HyILCVDw6y4
|
|
|
111
111
|
gllm_inference/request_processor/uses_lm_mixin.pyi,sha256=LYHq-zLoXEMel1LfVdYv7W3BZ8WtBLo_WWFjRf10Yto,6512
|
|
112
112
|
gllm_inference/schema/__init__.pyi,sha256=4qSVD761xZz1r_GML6g7DDAJhg4nrmG17eYdfocRk-I,2422
|
|
113
113
|
gllm_inference/schema/activity.pyi,sha256=atrU4OwLesA9FEt1H7K3gsUWYNdOqpI5i2VdWkmo6cs,2367
|
|
114
|
-
gllm_inference/schema/attachment.pyi,sha256=
|
|
114
|
+
gllm_inference/schema/attachment.pyi,sha256=myJ_cI_h5mwUdvmMrWpSQIwj3nIxe8SD7HxO37o_3D4,4611
|
|
115
115
|
gllm_inference/schema/code_exec_result.pyi,sha256=WQ-ARoGM9r6nyRX-A0Ro1XKiqrc9R3jRYXZpu_xo5S4,573
|
|
116
116
|
gllm_inference/schema/config.pyi,sha256=NVmjQK6HipIE0dKSfx12hgIC0O-S1HEcAc-TWlXAF5A,689
|
|
117
117
|
gllm_inference/schema/enums.pyi,sha256=aJjmCCUV4ASDM8VTiSJlxPxdapKkxXjS4_NqRplyUnE,2145
|
|
@@ -131,7 +131,7 @@ gllm_inference/utils/io_utils.pyi,sha256=Eg7dvHWdXslTKdjh1j3dG50i7r35XG2zTmJ9XXv
|
|
|
131
131
|
gllm_inference/utils/langchain.pyi,sha256=4AwFiVAO0ZpdgmqeC4Pb5NJwBt8vVr0MSUqLeCdTscc,1194
|
|
132
132
|
gllm_inference/utils/validation.pyi,sha256=OWRZxeVGIuuvNU0LqLGB-9gNmypvbH-LcSJx91rnH1k,453
|
|
133
133
|
gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
|
|
134
|
-
gllm_inference_binary-0.5.
|
|
135
|
-
gllm_inference_binary-0.5.
|
|
136
|
-
gllm_inference_binary-0.5.
|
|
137
|
-
gllm_inference_binary-0.5.
|
|
134
|
+
gllm_inference_binary-0.5.58.dist-info/METADATA,sha256=vkXLbZSNRDU_sSSB8_3vsYupuJJGtsEiauFswv46_No,5945
|
|
135
|
+
gllm_inference_binary-0.5.58.dist-info/WHEEL,sha256=x5rgv--I0NI0IT1Lh9tN1VG2cI637p3deednwYLKnxc,96
|
|
136
|
+
gllm_inference_binary-0.5.58.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
|
|
137
|
+
gllm_inference_binary-0.5.58.dist-info/RECORD,,
|
|
File without changes
|
{gllm_inference_binary-0.5.56.dist-info → gllm_inference_binary-0.5.58.dist-info}/top_level.txt
RENAMED
|
File without changes
|