gllm-inference-binary 0.5.18__cp313-cp313-macosx_13_0_x86_64.whl → 0.5.19__cp313-cp313-macosx_13_0_x86_64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of gllm-inference-binary might be problematic. Click here for more details.

@@ -1,6 +1,5 @@
1
1
  from _typeshed import Incomplete
2
2
  from gllm_core.utils.retry import RetryConfig as RetryConfig
3
- from gllm_inference.constants import INVOKER_PROPAGATED_MAX_RETRIES as INVOKER_PROPAGATED_MAX_RETRIES
4
3
  from gllm_inference.em_invoker.em_invoker import BaseEMInvoker as BaseEMInvoker
5
4
  from gllm_inference.em_invoker.schema.twelvelabs import InputType as InputType, Key as Key, OutputType as OutputType
6
5
  from gllm_inference.schema import Attachment as Attachment, AttachmentType as AttachmentType, EMContent as EMContent, ModelId as ModelId, ModelProvider as ModelProvider, TruncationConfig as TruncationConfig, Vector as Vector
Binary file
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: gllm-inference-binary
3
- Version: 0.5.18
3
+ Version: 0.5.19
4
4
  Summary: A library containing components related to model inferences in Gen AI applications.
5
5
  Author: Henry Wicaksono
6
6
  Author-email: henry.wicaksono@gdplabs.id
@@ -27,7 +27,7 @@ gllm_inference/em_invoker/schema/openai.pyi,sha256=Q_dsEcodkOXYXPdrkOkW0LnuLhfeq
27
27
  gllm_inference/em_invoker/schema/openai_compatible.pyi,sha256=gmvGtsWoOMBelke_tZjC6dKimFBW9f4Vrgv0Ig0OM9Q,150
28
28
  gllm_inference/em_invoker/schema/twelvelabs.pyi,sha256=F6wKHgG01bYskJpKoheBSpRpHUfFpteKn9sj9n5YfcU,372
29
29
  gllm_inference/em_invoker/schema/voyage.pyi,sha256=HVpor0fqNy-IwapCICfsgFmqf1FJXCOMIxS2vOXhHd8,289
30
- gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=RYY4td4N3cPSy-yXTEUKwYLz0AH8mLxi2j5Bfr9PS0g,5447
30
+ gllm_inference/em_invoker/twelevelabs_em_invoker.pyi,sha256=9F37VuLYTH5xsPSS_uzhrsPJug0QifX_qh2GwG0jSTU,5345
31
31
  gllm_inference/em_invoker/voyage_em_invoker.pyi,sha256=zJZqMvvFKu3sHdrNM773UYjfHVlnwE2w2BmvdFcHzV0,5515
32
32
  gllm_inference/exceptions/__init__.pyi,sha256=v9uxjW5DssIn7n_bKqT7L83CeqFET2Z45GFOvi78UuE,977
33
33
  gllm_inference/exceptions/error_parser.pyi,sha256=4RkVfS2Fl9kjz_h2bK9eoAeI-Y-VkHcUqXWj68BsYig,2393
@@ -96,8 +96,8 @@ gllm_inference/utils/io_utils.pyi,sha256=7kUTacHAVRYoemFUOjCH7-Qmw-YsQGd6rGYxjf_
96
96
  gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
97
97
  gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
98
98
  gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
99
- gllm_inference.cpython-313-darwin.so,sha256=KSsTBbpELj6uU0DXWlgq-KpEzG-i5rpnYNjjuNmQI-k,4190792
99
+ gllm_inference.cpython-313-darwin.so,sha256=RO6__6IZ4Qan03-pTFhkeFz9sv0GWDs6kEfDTBFkcbw,4190352
100
100
  gllm_inference.pyi,sha256=uxl1voKdn19LurAHKEZLWbq9ryPO4UkJ1Nk1MM8IL34,3636
101
- gllm_inference_binary-0.5.18.dist-info/METADATA,sha256=Heo4b0XS1gQgOtbWPzIYPTic4-esBmsLMJ2LO68SP3U,4608
102
- gllm_inference_binary-0.5.18.dist-info/WHEEL,sha256=PCOZcL_jcbAVhuFR5ylE4Mr-7HPGHAcfJk9OpuMh7RQ,107
103
- gllm_inference_binary-0.5.18.dist-info/RECORD,,
101
+ gllm_inference_binary-0.5.19.dist-info/METADATA,sha256=RYPC2mk8-uXt5cFOZh7VdB6ccfFPjZ7AjvSsCoCfZQI,4608
102
+ gllm_inference_binary-0.5.19.dist-info/WHEEL,sha256=PCOZcL_jcbAVhuFR5ylE4Mr-7HPGHAcfJk9OpuMh7RQ,107
103
+ gllm_inference_binary-0.5.19.dist-info/RECORD,,