gllm-inference-binary 0.5.13__cp313-cp313-manylinux_2_31_x86_64.whl → 0.5.14__cp313-cp313-manylinux_2_31_x86_64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -60,7 +60,7 @@ class LMRequestProcessor:
60
60
 
61
61
  This method clears the response schema for the LM invoker.
62
62
  """
63
- async def process(self, prompt_kwargs: dict[str, Any] | None = None, history: list[Message] | None = None, extra_contents: list[MessageContent] | None = None, hyperparameters: dict[str, Any] | None = None, event_emitter: EventEmitter | None = None, auto_execute_tools: bool = True, max_lm_calls: int = 5) -> Any:
63
+ async def process(self, prompt_kwargs: dict[str, Any] | None = None, history: list[Message] | None = None, extra_contents: list[MessageContent] | None = None, hyperparameters: dict[str, Any] | None = None, event_emitter: EventEmitter | None = None, auto_execute_tools: bool = True, max_lm_calls: int = 5, **kwargs: Any) -> Any:
64
64
  """Processes a language model inference request.
65
65
 
66
66
  This method processes the language model inference request as follows:
@@ -72,8 +72,8 @@ class LMRequestProcessor:
72
72
  LMOutput object, the output parser will process the `response` attribute of the LMOutput object.
73
73
 
74
74
  Args:
75
- prompt_kwargs (dict[str, Any], optional): A dictionary of arguments used to format the prompt.
76
- Defaults to None, in which case no arguments will be passed to the prompt builder.
75
+ prompt_kwargs (dict[str, Any], optional): Deprecated parameter for passing prompt kwargs.
76
+ Replaced by **kwargs. Defaults to None
77
77
  history (list[Message] | None, optional): A list of conversation history to be included in the prompt.
78
78
  Defaults to None.
79
79
  extra_contents (list[MessageContent] | None, optional): A list of extra contents to be included in the
@@ -86,6 +86,15 @@ class LMRequestProcessor:
86
86
  tool calls. Defaults to True.
87
87
  max_lm_calls (int, optional): The maximum number of times the language model can be invoked
88
88
  when `auto_execute_tools` is True. Defaults to 5.
89
+ **kwargs (Any): Keyword arguments that will be passed to format the prompt builder.
90
+ Values must be either a string or an object that can be serialized to a string.
91
+ Reserved keyword arguments that cannot be passed to the prompt builder include:
92
+ 1. `history`
93
+ 2. `extra_contents`
94
+ 3. `hyperparameters`
95
+ 4. `event_emitter`
96
+ 5. `auto_execute_tools`
97
+ 6. `max_lm_calls`
89
98
 
90
99
  Returns:
91
100
  Any: The result of the language model invocation, optionally parsed by the output parser.
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: gllm-inference-binary
3
- Version: 0.5.13
3
+ Version: 0.5.14
4
4
  Summary: A library containing components related to model inferences in Gen AI applications.
5
5
  Author: Henry Wicaksono
6
6
  Author-email: henry.wicaksono@gdplabs.id
@@ -74,7 +74,7 @@ gllm_inference/prompt_formatter/mistral_prompt_formatter.pyi,sha256=DgFitxfvCSJf
74
74
  gllm_inference/prompt_formatter/openai_prompt_formatter.pyi,sha256=QB7gHddipQuAolUKO01ApZeaRxBtRZzglS5B78mzsV8,1286
75
75
  gllm_inference/prompt_formatter/prompt_formatter.pyi,sha256=UkcPi5ao98OGJyNRsqfhYTlMW-ZLNITaGZUTwzvBzqk,1146
76
76
  gllm_inference/request_processor/__init__.pyi,sha256=hVnfdNZnkTBJHnmLtN3Na4ANP0yK6AstWdIizVr2Apo,227
77
- gllm_inference/request_processor/lm_request_processor.pyi,sha256=7pVNb2GwITb1jTflZP498qZ321G15b16jayZuuhuO1o,5424
77
+ gllm_inference/request_processor/lm_request_processor.pyi,sha256=VnYc8E3Iayyhw-rPnGPfTKuO3ohgFsS8HPrZJeyES5I,5889
78
78
  gllm_inference/request_processor/uses_lm_mixin.pyi,sha256=He-ytjwv2H5Hn312WFBAlBK96ALKTtDO3AT_80hCGTg,2321
79
79
  gllm_inference/schema/__init__.pyi,sha256=aQx7eKVsFNgJkYtbbKppa9Ww2WfTe0yzl_ewajDCo6w,1363
80
80
  gllm_inference/schema/attachment.pyi,sha256=jApuzjOHJDCz4lr4MlHzBgIndh559nbWu2Xp1fk3hso,3297
@@ -92,8 +92,8 @@ gllm_inference/utils/__init__.pyi,sha256=npmBmmlBv7cPHMg1hdL3S2_RelD6vk_LhCsGELh
92
92
  gllm_inference/utils/langchain.pyi,sha256=VluQiHkGigDdqLUbhB6vnXiISCP5hHqV0qokYY6dC1A,1164
93
93
  gllm_inference/utils/validation.pyi,sha256=toxBtRp-VItC_X7sNi-GDd7sjibBdWMrR0q01OI2D7k,385
94
94
  gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
95
- gllm_inference.cpython-313-x86_64-linux-gnu.so,sha256=zZvCVUM1RJgrOYb7k2HqytO95jTy-G1TNbI0Dzl45Os,4195240
95
+ gllm_inference.cpython-313-x86_64-linux-gnu.so,sha256=eMDbRV61RFuo8CAsrCmJMcVLPf8IbkVFraO5STsVnwk,4199336
96
96
  gllm_inference.pyi,sha256=896j4U7emFAFPObQk6dVwj92K147TZaBmr7ywLj8kEk,3540
97
- gllm_inference_binary-0.5.13.dist-info/METADATA,sha256=_xhbXMSCIbgUComS2gsY5OxhVqzhF1LYxahGn9E7hoE,4608
98
- gllm_inference_binary-0.5.13.dist-info/WHEEL,sha256=qGYSeeDMRvGsNMRKS15OK05VQRV6Z0DMQkqDjYiypg0,110
99
- gllm_inference_binary-0.5.13.dist-info/RECORD,,
97
+ gllm_inference_binary-0.5.14.dist-info/METADATA,sha256=_PX1mmZ2KxIQu7Tp9zlYLjTwSBl0UzG9ewo8UPe5y6w,4608
98
+ gllm_inference_binary-0.5.14.dist-info/WHEEL,sha256=qGYSeeDMRvGsNMRKS15OK05VQRV6Z0DMQkqDjYiypg0,110
99
+ gllm_inference_binary-0.5.14.dist-info/RECORD,,