gllm-inference-binary 0.5.32b1__cp311-cp311-manylinux_2_31_x86_64.whl → 0.5.33__cp311-cp311-manylinux_2_31_x86_64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -134,7 +134,12 @@ class AzureOpenAILMInvoker(OpenAILMInvoker):
134
134
  ```python
135
135
  LMOutput(
136
136
  response="Golden retriever is a good dog breed.",
137
- token_usage=TokenUsage(input_tokens=100, output_tokens=50),
137
+ token_usage=TokenUsage(
138
+ input_tokens=1500,
139
+ output_tokens=200,
140
+ input_token_details=InputTokenDetails(cached_tokens=1200, uncached_tokens=300),
141
+ output_token_details=OutputTokenDetails(reasoning_tokens=180, response_tokens=20),
142
+ ),
138
143
  duration=0.729,
139
144
  finish_details={"status": "completed", "incomplete_details": {"reason": None}},
140
145
  )
@@ -137,7 +137,12 @@ class OpenAILMInvoker(BaseLMInvoker):
137
137
  ```python
138
138
  LMOutput(
139
139
  response="Golden retriever is a good dog breed.",
140
- token_usage=TokenUsage(input_tokens=100, output_tokens=50),
140
+ token_usage=TokenUsage(
141
+ input_tokens=1500,
142
+ output_tokens=200,
143
+ input_token_details=InputTokenDetails(cached_tokens=1200, uncached_tokens=300),
144
+ output_token_details=OutputTokenDetails(reasoning_tokens=180, response_tokens=20),
145
+ ),
141
146
  duration=0.729,
142
147
  finish_details={"status": "completed", "incomplete_details": {"reason": None}},
143
148
  )
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.2
2
2
  Name: gllm-inference-binary
3
- Version: 0.5.32b1
3
+ Version: 0.5.33
4
4
  Summary: A library containing components related to model inferences in Gen AI applications.
5
5
  Author-email: Henry Wicaksono <henry.wicaksono@gdplabs.id>, Resti Febrina <resti.febrina@gdplabs.id>
6
6
  Requires-Python: <3.14,>=3.11
@@ -1,4 +1,4 @@
1
- gllm_inference.cpython-311-x86_64-linux-gnu.so,sha256=JEGvv1Psr4swuUxGdfdvOb9LbWDYxfUVzAXAICsjebA,4425392
1
+ gllm_inference.cpython-311-x86_64-linux-gnu.so,sha256=g5WSHBiOraH_OsostDXYAuh5qU3izIcUJj0a1r9GUXQ,4425392
2
2
  gllm_inference.pyi,sha256=bvMQNMzysfZtXgjW4ZX0KwSOV4uroNakpE0NUHKCMmk,4199
3
3
  gllm_inference/__init__.pyi,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
4
4
  gllm_inference/constants.pyi,sha256=EFVMtK3xDK2yjGoHp8EL3LeRZWhIefVKClI9jvbfQQ0,267
@@ -37,7 +37,7 @@ gllm_inference/exceptions/exceptions.pyi,sha256=Bv996qLa_vju0Qjf4GewMxdkq8CV9LRZ
37
37
  gllm_inference/exceptions/provider_error_map.pyi,sha256=P1WnhWkM103FW6hqMfNZBOmYSWOmsJtll3VQV8DGb8E,1210
38
38
  gllm_inference/lm_invoker/__init__.pyi,sha256=NmQSqObPjevEP1KbbrNnaz4GMh175EVPERZ19vK5Emc,1202
39
39
  gllm_inference/lm_invoker/anthropic_lm_invoker.pyi,sha256=rJeQ9jpUIvcf5z1BB9Lksqf37ZgUzcnFqDMstOl3-kk,17235
40
- gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=EXPFTsPwUk42B12MVDhh98maCFSkdPWfqJeht5Wjpq4,14783
40
+ gllm_inference/lm_invoker/azure_openai_lm_invoker.pyi,sha256=7zF8pj3LbuOwu1QArPX0ra6IrqUq5AkkGbC_wXiGhlA,15027
41
41
  gllm_inference/lm_invoker/bedrock_lm_invoker.pyi,sha256=uZ9wpzOKSOvgu1ICMLqEXcrOE3RIbUmqHmgtuwBekPg,12802
42
42
  gllm_inference/lm_invoker/datasaur_lm_invoker.pyi,sha256=J_tfnIgVDr-zQ7YE5_TKMyZyA336ly04g1l-ZKnr1As,9315
43
43
  gllm_inference/lm_invoker/google_lm_invoker.pyi,sha256=4-3CwfBcDh6thxkidRcYbGVp9bCDkQTemat6VBHsUC8,17164
@@ -45,7 +45,7 @@ gllm_inference/lm_invoker/langchain_lm_invoker.pyi,sha256=hnQcScOHs31xx4GB6YI-Rn
45
45
  gllm_inference/lm_invoker/litellm_lm_invoker.pyi,sha256=eEPvDOCj55f9wJ0neNl4O9XQWvSI6YWJgHZMHOaYGRk,13240
46
46
  gllm_inference/lm_invoker/lm_invoker.pyi,sha256=hjolpN8BzUrhgy8MSpnYxhrlWPJO1LXeCFGlBhQ-eBw,8152
47
47
  gllm_inference/lm_invoker/openai_compatible_lm_invoker.pyi,sha256=XV-KjulVYAhU0e2giqOdHUGCSCrybXRWsrtzZByqOXI,15050
48
- gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=xrNdXmsy8i4XZRUTddaRdNbgtZtgal7l6zOogNAEa4M,21854
48
+ gllm_inference/lm_invoker/openai_lm_invoker.pyi,sha256=8U-b2dUcHGHm0uoPcZr3UBNhJsLaNCOHftfkebZP_DU,22098
49
49
  gllm_inference/lm_invoker/xai_lm_invoker.pyi,sha256=rV8D3E730OUmwK7jELKSziMUl7MnpbfxMAvMuq8-Aew,15687
50
50
  gllm_inference/lm_invoker/batch/__init__.pyi,sha256=W4W-_yfk7lL20alREJai6GnwuQvdlKRfwQCX4mQK4XI,127
51
51
  gllm_inference/lm_invoker/batch/batch_operations.pyi,sha256=Oo7hoyPSfPZdy1mXvSdvtRndvq-XTIbPIjEoGvJj5C0,5372
@@ -103,7 +103,7 @@ gllm_inference/utils/io_utils.pyi,sha256=7kUTacHAVRYoemFUOjCH7-Qmw-YsQGd6rGYxjf_
103
103
  gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
104
104
  gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
105
105
  gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
106
- gllm_inference_binary-0.5.32b1.dist-info/METADATA,sha256=_Vq9GKxNtOxODXNNsnQgpaKANbJuC3DQqWUllr0ZS7k,4859
107
- gllm_inference_binary-0.5.32b1.dist-info/WHEEL,sha256=WMelAR6z66VnlU3tu68fV-jM5qbG8iPyeTqaBcpU3pI,108
108
- gllm_inference_binary-0.5.32b1.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
109
- gllm_inference_binary-0.5.32b1.dist-info/RECORD,,
106
+ gllm_inference_binary-0.5.33.dist-info/METADATA,sha256=g86H1HI56Bl8-7LIVHMWyexW9ldz5hbbLC0qtU_xBAg,4857
107
+ gllm_inference_binary-0.5.33.dist-info/WHEEL,sha256=WMelAR6z66VnlU3tu68fV-jM5qbG8iPyeTqaBcpU3pI,108
108
+ gllm_inference_binary-0.5.33.dist-info/top_level.txt,sha256=FpOjtN80F-qVNgbScXSEyqa0w09FYn6301iq6qt69IQ,15
109
+ gllm_inference_binary-0.5.33.dist-info/RECORD,,