lm-deluge 0.0.25__py3-none-any.whl → 0.0.27__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of lm-deluge might be problematic. Click here for more details.

@@ -37,14 +37,17 @@ async def _build_gemini_request(
37
37
 
38
38
  # Handle reasoning models (thinking)
39
39
  if model.reasoning_model:
40
- request_json["generationConfig"]["thinkingConfig"] = {"includeThoughts": True}
41
- if sampling_params.reasoning_effort and "flash" in model.id:
42
- budget = {"low": 1024, "medium": 4096, "high": 16384}.get(
43
- sampling_params.reasoning_effort
44
- )
45
- request_json["generationConfig"]["thinkingConfig"]["thinkingBudget"] = (
46
- budget
47
- )
40
+ thinking_config = None
41
+ effort = sampling_params.reasoning_effort
42
+ if effort is None or effort == "none":
43
+ # Explicitly disable thoughts when no effort is requested
44
+ thinking_config = {"includeThoughts": False, "thinkingBudget": 0}
45
+ else:
46
+ thinking_config = {"includeThoughts": True}
47
+ if effort in {"low", "medium", "high"} and "flash" in model.id:
48
+ budget = {"low": 1024, "medium": 4096, "high": 16384}[effort]
49
+ thinking_config["thinkingBudget"] = budget
50
+ request_json["generationConfig"]["thinkingConfig"] = thinking_config
48
51
 
49
52
  else:
50
53
  if sampling_params.reasoning_effort:
lm_deluge/file.py CHANGED
@@ -1,3 +1,4 @@
1
+ from functools import cached_property
1
2
  import os
2
3
  import io
3
4
  import requests
@@ -68,13 +69,13 @@ class File:
68
69
  return encoded
69
70
  return f"data:{self._mime()};base64,{encoded}"
70
71
 
71
- @property
72
+ @cached_property
72
73
  def fingerprint(self) -> str:
73
74
  # Hash the file contents for fingerprinting
74
75
  file_bytes = self._bytes()
75
76
  return xxhash.xxh64(file_bytes).hexdigest()
76
77
 
77
- @property
78
+ @cached_property
78
79
  def size(self) -> int:
79
80
  """Return file size in bytes."""
80
81
  return len(self._bytes())
@@ -1,4 +1,5 @@
1
1
  from dataclasses import dataclass, field
2
+ from functools import cached_property
2
3
  from typing import Any, Callable
3
4
 
4
5
  from .config import SamplingParams
@@ -39,14 +40,18 @@ class RequestContext:
39
40
 
40
41
  # Computed properties
41
42
  cache_key: str = field(init=False)
42
- num_tokens: int = field(init=False)
43
+ # num_tokens: int = field(init=False)
43
44
 
44
- def __post_init__(self):
45
- # Compute cache key from prompt fingerprint
46
- self.cache_key = self.prompt.fingerprint
45
+ # def __post_init__(self):
46
+ # # Compute cache key from prompt fingerprint
47
+ # # self.cache_key = self.prompt.fingerprint
47
48
 
48
- # Compute token count
49
- self.num_tokens = self.prompt.count_tokens(self.sampling_params.max_new_tokens)
49
+ # # Compute token count
50
+ # self.num_tokens =
51
+
52
+ @cached_property
53
+ def num_tokens(self):
54
+ return self.prompt.count_tokens(self.sampling_params.max_new_tokens)
50
55
 
51
56
  def maybe_callback(self, response, tracker):
52
57
  if not self.callback:
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: lm_deluge
3
- Version: 0.0.25
3
+ Version: 0.0.27
4
4
  Summary: Python utility for using LLM API models.
5
5
  Author-email: Benjamin Anderson <ben@trytaylor.ai>
6
6
  Requires-Python: >=3.10
@@ -274,7 +274,7 @@ We support all models in `src/lm_deluge/models.py`. Vertex support is not planne
274
274
 
275
275
  ## Feature Support
276
276
 
277
- We support structured outputs via `json_mode` parameter provided to `SamplingParams`. Structured outputs with a schema are planned. Reasoning models are supported via the `reasoning_effort` parameter, which is translated to a thinking budget for Claude/Gemini. Image models are supported. We support tool use as documented above. We support logprobs for OpenAI models that return them.
277
+ We support structured outputs via `json_mode` parameter provided to `SamplingParams`. Structured outputs with a schema are planned. Reasoning models are supported via the `reasoning_effort` parameter, which is translated to a thinking budget for Claude/Gemini. Passing `None` (or the string `"none"`) disables Gemini thoughts entirely. Image models are supported. We support tool use as documented above. We support logprobs for OpenAI models that return them.
278
278
 
279
279
  ## Built‑in tools
280
280
 
@@ -6,12 +6,12 @@ lm_deluge/client.py,sha256=nAGMwdUPDVx-x23hZF6U5Yhug6Zf5FT27RHj_kj8nZk,26369
6
6
  lm_deluge/config.py,sha256=H1tQyJDNHGFuwxqQNL5Z-CjWAC0luHSBA3iY_pxmACM,932
7
7
  lm_deluge/embed.py,sha256=CO-TOlC5kOTAM8lcnicoG4u4K664vCBwHF1vHa-nAGg,13382
8
8
  lm_deluge/errors.py,sha256=oHjt7YnxWbh-eXMScIzov4NvpJMo0-2r5J6Wh5DQ1tk,209
9
- lm_deluge/file.py,sha256=zQH1STMjCG9pczO7Fk9Jw0_0Pj_8CogcdIxTe4J4AJw,5414
9
+ lm_deluge/file.py,sha256=ggOXcAOI-LAykXLb3N3WfLlHdBjQIJtW9ltmfKCIsNQ,5466
10
10
  lm_deluge/gemini_limits.py,sha256=V9mpS9JtXYz7AY6OuKyQp5TuIMRH1BVv9YrSNmGmHNA,1569
11
11
  lm_deluge/image.py,sha256=Qpa0k5yXfrpSaHzVUwW_TEn7yEgmwzYGL17Sa7-KhSA,7729
12
12
  lm_deluge/models.py,sha256=3vgI1BlfT4_Higev25QhhXJufQvsI6pd0yjF9YL0crA,49812
13
13
  lm_deluge/prompt.py,sha256=cfwzCAmT-1K0v7SfEMUrxpBkJGgf7IFlWfNLJrCcoBM,37025
14
- lm_deluge/request_context.py,sha256=0X-5m8BKn51rnnjzGDDXqbuSUEFGjdayirQjbvPcjMI,2425
14
+ lm_deluge/request_context.py,sha256=o33LSEwnK6YPhZeulUoSE_VrdKCXiCQa0tjjixK2K6M,2540
15
15
  lm_deluge/rerank.py,sha256=-NBAJdHz9OB-SWWJnHzkFmeVO4wR6lFV7Vw-SxG7aVo,11457
16
16
  lm_deluge/tool.py,sha256=_coOKB9nPNVZoseMRumRyQ8BMR7_d0IlstzMHNT69JY,15732
17
17
  lm_deluge/tracker.py,sha256=-EkFDAklh5mclIFR-5SthAwNL4p1yKS8LUN7rhpOVPQ,9266
@@ -21,7 +21,7 @@ lm_deluge/api_requests/anthropic.py,sha256=Vi2iGQw6LGPW8d4X489Jg-dlEye529jf8i5Ys
21
21
  lm_deluge/api_requests/base.py,sha256=EVHNFtlttKbN7Tt1MnLaO-NjvKHPSV5CqlRv-OnpVAE,5593
22
22
  lm_deluge/api_requests/bedrock.py,sha256=p6leW5stnvb406lwbmVFmfTGxdEI-t0GfaTExpol3qk,10900
23
23
  lm_deluge/api_requests/common.py,sha256=BZ3vRO5TB669_UsNKugkkuFSzoLHOYJIKt4nV4sf4vc,422
24
- lm_deluge/api_requests/gemini.py,sha256=UeA4gsIePsIu6IQQ7izR359LLQQTi0Qky_ykjH1XQf4,7461
24
+ lm_deluge/api_requests/gemini.py,sha256=W4NjQ0buBsdS7RYpzDahrXNQWMzDHRMLNRSphCOmIqg,7685
25
25
  lm_deluge/api_requests/mistral.py,sha256=S_LpOfCGbCVEROH_od3P-tYeNYTKFMamMTL-c_wFCBI,4597
26
26
  lm_deluge/api_requests/openai.py,sha256=jb_pBGqSbqs7SvbA45Odu3JkwS3jsKn-p3hG8-qkYbc,21509
27
27
  lm_deluge/api_requests/response.py,sha256=FtkVYk_rDH93Kj9pqbB-l7a4dQHzVr6ivKL9khYKLbs,5966
@@ -48,8 +48,8 @@ lm_deluge/util/logprobs.py,sha256=UkBZakOxWluaLqHrjARu7xnJ0uCHVfLGHJdnYlEcutk,11
48
48
  lm_deluge/util/spatial.py,sha256=BsF_UKhE-x0xBirc-bV1xSKZRTUhsOBdGqsMKme20C8,4099
49
49
  lm_deluge/util/validation.py,sha256=hz5dDb3ebvZrZhnaWxOxbNSVMI6nmaOODBkk0htAUhs,1575
50
50
  lm_deluge/util/xml.py,sha256=Ft4zajoYBJR3HHCt2oHwGfymGLdvp_gegVmJ-Wqk4Ck,10547
51
- lm_deluge-0.0.25.dist-info/licenses/LICENSE,sha256=uNNXGXPCw2TC7CUs7SEBkA-Mz6QBQFWUUEWDMgEs1dU,1058
52
- lm_deluge-0.0.25.dist-info/METADATA,sha256=I-AUU79sc9sxjRV8UAPVWtlmzg10voZ2irH-bLkh_9E,12978
53
- lm_deluge-0.0.25.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
54
- lm_deluge-0.0.25.dist-info/top_level.txt,sha256=hqU-TJX93yBwpgkDtYcXyLr3t7TLSCCZ_reytJjwBaE,10
55
- lm_deluge-0.0.25.dist-info/RECORD,,
51
+ lm_deluge-0.0.27.dist-info/licenses/LICENSE,sha256=uNNXGXPCw2TC7CUs7SEBkA-Mz6QBQFWUUEWDMgEs1dU,1058
52
+ lm_deluge-0.0.27.dist-info/METADATA,sha256=vKjAfWLWBC9_fgSvYXlQQMxgUEf1E3n57Y4kZBy9_7I,13053
53
+ lm_deluge-0.0.27.dist-info/WHEEL,sha256=_zCd3N1l69ArxyTb8rzEoP9TpbYXkqRFSNOD5OuxnTs,91
54
+ lm_deluge-0.0.27.dist-info/top_level.txt,sha256=hqU-TJX93yBwpgkDtYcXyLr3t7TLSCCZ_reytJjwBaE,10
55
+ lm_deluge-0.0.27.dist-info/RECORD,,