gllm-inference-binary 0.4.61__cp313-cp313-manylinux_2_31_x86_64.whl → 0.4.62__cp313-cp313-manylinux_2_31_x86_64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of gllm-inference-binary might be problematic. Click here for more details.
- gllm_inference/catalog/lm_request_processor_catalog.pyi +7 -4
- gllm_inference.cpython-313-x86_64-linux-gnu.so +0 -0
- {gllm_inference_binary-0.4.61.dist-info → gllm_inference_binary-0.4.62.dist-info}/METADATA +1 -1
- {gllm_inference_binary-0.4.61.dist-info → gllm_inference_binary-0.4.62.dist-info}/RECORD +5 -5
- {gllm_inference_binary-0.4.61.dist-info → gllm_inference_binary-0.4.62.dist-info}/WHEEL +0 -0
|
@@ -93,10 +93,13 @@ class LMRequestProcessorCatalog(BaseCatalog[LMRequestProcessor]):
|
|
|
93
93
|
2.3. For the available model ID formats, see: https://gdplabs.gitbook.io/sdk/resources/supported-models
|
|
94
94
|
3. `credentials` is optional. If it is filled, it can either be:
|
|
95
95
|
3.1. An environment variable name containing the API key (e.g. OPENAI_API_KEY).
|
|
96
|
-
3.2.
|
|
97
|
-
|
|
98
|
-
|
|
99
|
-
|
|
96
|
+
3.2. An environment variable name containing the path to a credentials JSON file
|
|
97
|
+
(e.g. GOOGLE_CREDENTIALS_FILE_PATH). Currently only supported for Google Vertex AI.
|
|
98
|
+
3.3. A dictionary of credentials, with each value being an environment variable name corresponding to the
|
|
99
|
+
credential (e.g. {"api_key": "OPENAI_API_KEY"}). Currently supported for Bedrock and LangChain.
|
|
100
|
+
If it is empty, the LM invoker will use the default credentials loaded from the environment variables.
|
|
101
|
+
4. `config` is optional. If filled, must be a dictionary containing the configuration for the LM invoker.
|
|
102
|
+
If it is empty, the LM invoker will use the default configuration.
|
|
100
103
|
5. `output_parser_type` can either be:
|
|
101
104
|
5.1. none: No output parser will be used.
|
|
102
105
|
5.2. json: The JSONOutputParser will be used.
|
|
Binary file
|
|
@@ -8,7 +8,7 @@ gllm_inference/builder/model_id.pyi,sha256=julTH1t5XVhMynbCB_JIXcEAexRrs3TqGvap3
|
|
|
8
8
|
gllm_inference/catalog/__init__.pyi,sha256=JBkPGTyiiZ30GECzJBW-mW8LekWyY2qyzal3eW7ynaM,287
|
|
9
9
|
gllm_inference/catalog/catalog.pyi,sha256=MH1CSaGPa1EgnxwIsUPKwL0zMEbWeUAvGJQemRpOzs4,4635
|
|
10
10
|
gllm_inference/catalog/component_map.pyi,sha256=X0F4M94Jm8SjnDtRG9zkfUknGvlnnNAbSJR43h6JwVI,1346
|
|
11
|
-
gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=
|
|
11
|
+
gllm_inference/catalog/lm_request_processor_catalog.pyi,sha256=HgorZ_sIWZRPURDNN_jaOC0yVUnEVw5yVmaZBneQC5w,5357
|
|
12
12
|
gllm_inference/catalog/prompt_builder_catalog.pyi,sha256=QHSrSZp8z0tbeDw-5Eksr-VCcXuUtmdAikc26wVc1X0,3805
|
|
13
13
|
gllm_inference/constants.pyi,sha256=J3fxih6ArNPsRNhvugodRgKAFjewSzP5wrj6knjbmw0,273
|
|
14
14
|
gllm_inference/em_invoker/__init__.pyi,sha256=ciD_057T702P4gAJgJMPkYhl8lU4Dp7HtSzJf6n83z0,1273
|
|
@@ -108,8 +108,8 @@ gllm_inference/utils/openai_multimodal_lm_helper.pyi,sha256=eF3MVWpQOyu_oYdHRWpR
|
|
|
108
108
|
gllm_inference/utils/retry.pyi,sha256=PVDHBDWfWj9Frvl0yf5X6nqI6oromc35LvOs-jDk3So,76
|
|
109
109
|
gllm_inference/utils/utils.pyi,sha256=px3RqfrgMTR_IvC2byKjSkNSrvE9_80nIe5UUw-d09s,6017
|
|
110
110
|
gllm_inference.build/.gitignore,sha256=aEiIwOuxfzdCmLZe4oB1JsBmCUxwG8x-u-HBCV9JT8E,1
|
|
111
|
-
gllm_inference.cpython-313-x86_64-linux-gnu.so,sha256=
|
|
111
|
+
gllm_inference.cpython-313-x86_64-linux-gnu.so,sha256=c8GaoYOIg-FJ5kq4MIY_qFyC9M26laJgu2I42p2i_To,4999176
|
|
112
112
|
gllm_inference.pyi,sha256=U8ZHrNyDfYAAEMXVFrCFSuf6o6NSuxcJBnMnnMTf330,5050
|
|
113
|
-
gllm_inference_binary-0.4.
|
|
114
|
-
gllm_inference_binary-0.4.
|
|
115
|
-
gllm_inference_binary-0.4.
|
|
113
|
+
gllm_inference_binary-0.4.62.dist-info/METADATA,sha256=_2NkG0nqiOli_nrerx5mdrbsgFlQkewk9pS6zIB9OeY,4958
|
|
114
|
+
gllm_inference_binary-0.4.62.dist-info/WHEEL,sha256=qGYSeeDMRvGsNMRKS15OK05VQRV6Z0DMQkqDjYiypg0,110
|
|
115
|
+
gllm_inference_binary-0.4.62.dist-info/RECORD,,
|
|
File without changes
|