lm-deluge 0.0.26__tar.gz → 0.0.27__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of lm-deluge might be problematic. Click here for more details.
- {lm_deluge-0.0.26/src/lm_deluge.egg-info → lm_deluge-0.0.27}/PKG-INFO +2 -2
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/README.md +1 -1
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/pyproject.toml +1 -1
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/gemini.py +11 -8
- {lm_deluge-0.0.26 → lm_deluge-0.0.27/src/lm_deluge.egg-info}/PKG-INFO +2 -2
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/LICENSE +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/setup.cfg +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/__init__.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/agent.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/__init__.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/anthropic.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/base.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/bedrock.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/common.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/deprecated/bedrock.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/deprecated/cohere.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/deprecated/deepseek.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/deprecated/mistral.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/deprecated/vertex.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/mistral.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/openai.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/api_requests/response.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/batches.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/built_in_tools/anthropic/__init__.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/built_in_tools/anthropic/bash.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/built_in_tools/anthropic/computer_use.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/built_in_tools/anthropic/editor.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/built_in_tools/base.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/built_in_tools/openai.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/cache.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/client.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/config.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/embed.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/errors.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/file.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/gemini_limits.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/image.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/llm_tools/__init__.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/llm_tools/classify.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/llm_tools/extract.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/llm_tools/locate.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/llm_tools/ocr.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/llm_tools/score.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/llm_tools/translate.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/models.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/prompt.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/request_context.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/rerank.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/tool.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/tracker.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/usage.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/util/json.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/util/logprobs.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/util/spatial.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/util/validation.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/util/xml.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge.egg-info/SOURCES.txt +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge.egg-info/dependency_links.txt +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge.egg-info/requires.txt +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge.egg-info/top_level.txt +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/tests/test_builtin_tools.py +0 -0
- {lm_deluge-0.0.26 → lm_deluge-0.0.27}/tests/test_native_mcp_server.py +0 -0
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: lm_deluge
|
|
3
|
-
Version: 0.0.
|
|
3
|
+
Version: 0.0.27
|
|
4
4
|
Summary: Python utility for using LLM API models.
|
|
5
5
|
Author-email: Benjamin Anderson <ben@trytaylor.ai>
|
|
6
6
|
Requires-Python: >=3.10
|
|
@@ -274,7 +274,7 @@ We support all models in `src/lm_deluge/models.py`. Vertex support is not planne
|
|
|
274
274
|
|
|
275
275
|
## Feature Support
|
|
276
276
|
|
|
277
|
-
We support structured outputs via `json_mode` parameter provided to `SamplingParams`. Structured outputs with a schema are planned. Reasoning models are supported via the `reasoning_effort` parameter, which is translated to a thinking budget for Claude/Gemini. Image models are supported. We support tool use as documented above. We support logprobs for OpenAI models that return them.
|
|
277
|
+
We support structured outputs via `json_mode` parameter provided to `SamplingParams`. Structured outputs with a schema are planned. Reasoning models are supported via the `reasoning_effort` parameter, which is translated to a thinking budget for Claude/Gemini. Passing `None` (or the string `"none"`) disables Gemini thoughts entirely. Image models are supported. We support tool use as documented above. We support logprobs for OpenAI models that return them.
|
|
278
278
|
|
|
279
279
|
## Built‑in tools
|
|
280
280
|
|
|
@@ -247,7 +247,7 @@ We support all models in `src/lm_deluge/models.py`. Vertex support is not planne
|
|
|
247
247
|
|
|
248
248
|
## Feature Support
|
|
249
249
|
|
|
250
|
-
We support structured outputs via `json_mode` parameter provided to `SamplingParams`. Structured outputs with a schema are planned. Reasoning models are supported via the `reasoning_effort` parameter, which is translated to a thinking budget for Claude/Gemini. Image models are supported. We support tool use as documented above. We support logprobs for OpenAI models that return them.
|
|
250
|
+
We support structured outputs via `json_mode` parameter provided to `SamplingParams`. Structured outputs with a schema are planned. Reasoning models are supported via the `reasoning_effort` parameter, which is translated to a thinking budget for Claude/Gemini. Passing `None` (or the string `"none"`) disables Gemini thoughts entirely. Image models are supported. We support tool use as documented above. We support logprobs for OpenAI models that return them.
|
|
251
251
|
|
|
252
252
|
## Built‑in tools
|
|
253
253
|
|
|
@@ -37,14 +37,17 @@ async def _build_gemini_request(
|
|
|
37
37
|
|
|
38
38
|
# Handle reasoning models (thinking)
|
|
39
39
|
if model.reasoning_model:
|
|
40
|
-
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
|
|
47
|
-
|
|
40
|
+
thinking_config = None
|
|
41
|
+
effort = sampling_params.reasoning_effort
|
|
42
|
+
if effort is None or effort == "none":
|
|
43
|
+
# Explicitly disable thoughts when no effort is requested
|
|
44
|
+
thinking_config = {"includeThoughts": False, "thinkingBudget": 0}
|
|
45
|
+
else:
|
|
46
|
+
thinking_config = {"includeThoughts": True}
|
|
47
|
+
if effort in {"low", "medium", "high"} and "flash" in model.id:
|
|
48
|
+
budget = {"low": 1024, "medium": 4096, "high": 16384}[effort]
|
|
49
|
+
thinking_config["thinkingBudget"] = budget
|
|
50
|
+
request_json["generationConfig"]["thinkingConfig"] = thinking_config
|
|
48
51
|
|
|
49
52
|
else:
|
|
50
53
|
if sampling_params.reasoning_effort:
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: lm_deluge
|
|
3
|
-
Version: 0.0.
|
|
3
|
+
Version: 0.0.27
|
|
4
4
|
Summary: Python utility for using LLM API models.
|
|
5
5
|
Author-email: Benjamin Anderson <ben@trytaylor.ai>
|
|
6
6
|
Requires-Python: >=3.10
|
|
@@ -274,7 +274,7 @@ We support all models in `src/lm_deluge/models.py`. Vertex support is not planne
|
|
|
274
274
|
|
|
275
275
|
## Feature Support
|
|
276
276
|
|
|
277
|
-
We support structured outputs via `json_mode` parameter provided to `SamplingParams`. Structured outputs with a schema are planned. Reasoning models are supported via the `reasoning_effort` parameter, which is translated to a thinking budget for Claude/Gemini. Image models are supported. We support tool use as documented above. We support logprobs for OpenAI models that return them.
|
|
277
|
+
We support structured outputs via `json_mode` parameter provided to `SamplingParams`. Structured outputs with a schema are planned. Reasoning models are supported via the `reasoning_effort` parameter, which is translated to a thinking budget for Claude/Gemini. Passing `None` (or the string `"none"`) disables Gemini thoughts entirely. Image models are supported. We support tool use as documented above. We support logprobs for OpenAI models that return them.
|
|
278
278
|
|
|
279
279
|
## Built‑in tools
|
|
280
280
|
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{lm_deluge-0.0.26 → lm_deluge-0.0.27}/src/lm_deluge/built_in_tools/anthropic/computer_use.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|