lm-deluge 0.0.67__tar.gz → 0.0.68__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (81) hide show
  1. {lm_deluge-0.0.67/src/lm_deluge.egg-info → lm_deluge-0.0.68}/PKG-INFO +1 -1
  2. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/pyproject.toml +1 -1
  3. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/client.py +9 -8
  4. {lm_deluge-0.0.67 → lm_deluge-0.0.68/src/lm_deluge.egg-info}/PKG-INFO +1 -1
  5. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/LICENSE +0 -0
  6. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/README.md +0 -0
  7. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/setup.cfg +0 -0
  8. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/__init__.py +0 -0
  9. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/__init__.py +0 -0
  10. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/anthropic.py +0 -0
  11. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/base.py +0 -0
  12. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/bedrock.py +0 -0
  13. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/common.py +0 -0
  14. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/deprecated/bedrock.py +0 -0
  15. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/deprecated/cohere.py +0 -0
  16. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/deprecated/deepseek.py +0 -0
  17. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/deprecated/mistral.py +0 -0
  18. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/deprecated/vertex.py +0 -0
  19. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/gemini.py +0 -0
  20. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/mistral.py +0 -0
  21. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/openai.py +0 -0
  22. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/api_requests/response.py +0 -0
  23. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/batches.py +0 -0
  24. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/built_in_tools/anthropic/__init__.py +0 -0
  25. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/built_in_tools/anthropic/bash.py +0 -0
  26. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/built_in_tools/anthropic/computer_use.py +0 -0
  27. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/built_in_tools/anthropic/editor.py +0 -0
  28. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/built_in_tools/base.py +0 -0
  29. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/built_in_tools/openai.py +0 -0
  30. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/cache.py +0 -0
  31. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/cli.py +0 -0
  32. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/config.py +0 -0
  33. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/embed.py +0 -0
  34. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/errors.py +0 -0
  35. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/file.py +0 -0
  36. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/image.py +0 -0
  37. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/llm_tools/__init__.py +0 -0
  38. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/llm_tools/classify.py +0 -0
  39. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/llm_tools/extract.py +0 -0
  40. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/llm_tools/locate.py +0 -0
  41. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/llm_tools/ocr.py +0 -0
  42. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/llm_tools/score.py +0 -0
  43. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/llm_tools/translate.py +0 -0
  44. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/__init__.py +0 -0
  45. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/anthropic.py +0 -0
  46. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/bedrock.py +0 -0
  47. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/cerebras.py +0 -0
  48. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/cohere.py +0 -0
  49. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/deepseek.py +0 -0
  50. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/fireworks.py +0 -0
  51. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/google.py +0 -0
  52. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/grok.py +0 -0
  53. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/groq.py +0 -0
  54. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/meta.py +0 -0
  55. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/mistral.py +0 -0
  56. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/openai.py +0 -0
  57. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/openrouter.py +0 -0
  58. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/models/together.py +0 -0
  59. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/presets/cerebras.py +0 -0
  60. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/presets/meta.py +0 -0
  61. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/prompt.py +0 -0
  62. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/request_context.py +0 -0
  63. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/rerank.py +0 -0
  64. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/tool.py +0 -0
  65. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/tracker.py +0 -0
  66. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/usage.py +0 -0
  67. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/util/harmony.py +0 -0
  68. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/util/json.py +0 -0
  69. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/util/logprobs.py +0 -0
  70. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/util/spatial.py +0 -0
  71. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/util/validation.py +0 -0
  72. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/util/xml.py +0 -0
  73. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge/warnings.py +0 -0
  74. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge.egg-info/SOURCES.txt +0 -0
  75. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge.egg-info/dependency_links.txt +0 -0
  76. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge.egg-info/requires.txt +0 -0
  77. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/src/lm_deluge.egg-info/top_level.txt +0 -0
  78. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/tests/test_builtin_tools.py +0 -0
  79. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/tests/test_file_upload.py +0 -0
  80. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/tests/test_native_mcp_server.py +0 -0
  81. {lm_deluge-0.0.67 → lm_deluge-0.0.68}/tests/test_openrouter_generic.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: lm_deluge
3
- Version: 0.0.67
3
+ Version: 0.0.68
4
4
  Summary: Python utility for using LLM API models.
5
5
  Author-email: Benjamin Anderson <ben@trytaylor.ai>
6
6
  Requires-Python: >=3.10
@@ -3,7 +3,7 @@ requires = ["setuptools", "wheel"]
3
3
 
4
4
  [project]
5
5
  name = "lm_deluge"
6
- version = "0.0.67"
6
+ version = "0.0.68"
7
7
  authors = [{ name = "Benjamin Anderson", email = "ben@trytaylor.ai" }]
8
8
  description = "Python utility for using LLM API models."
9
9
  readme = "README.md"
@@ -702,7 +702,7 @@ class _LLMClient(BaseModel):
702
702
 
703
703
  async def start(
704
704
  self,
705
- prompt: str | Conversation,
705
+ prompt: Prompt,
706
706
  *,
707
707
  tools: list[Tool | dict | MCPServer] | None = None,
708
708
  cache: CachePattern | None = None,
@@ -780,12 +780,12 @@ class _LLMClient(BaseModel):
780
780
 
781
781
  async def stream(
782
782
  self,
783
- prompt: str | Conversation,
783
+ prompt: Prompt,
784
784
  tools: list[Tool | dict | MCPServer] | None = None,
785
785
  ):
786
786
  model, sampling_params = self._select_model()
787
- if isinstance(prompt, str):
788
- prompt = Conversation.user(prompt)
787
+ prompt = prompts_to_conversations([prompt])[0]
788
+ assert isinstance(prompt, Conversation)
789
789
  async for item in stream_chat(
790
790
  model, prompt, sampling_params, tools, None, self.extra_headers
791
791
  ):
@@ -799,7 +799,7 @@ class _LLMClient(BaseModel):
799
799
 
800
800
  async def run_agent_loop(
801
801
  self,
802
- conversation: str | Conversation,
802
+ conversation: Prompt,
803
803
  *,
804
804
  tools: list[Tool | dict | MCPServer] | None = None,
805
805
  max_rounds: int = 5,
@@ -812,8 +812,9 @@ class _LLMClient(BaseModel):
812
812
  instances or built‑in tool dictionaries.
813
813
  """
814
814
 
815
- if isinstance(conversation, str):
816
- conversation = Conversation.user(conversation)
815
+ if not isinstance(conversation, Conversation):
816
+ conversation = prompts_to_conversations([conversation])[0]
817
+ assert isinstance(conversation, Conversation)
817
818
 
818
819
  # Expand MCPServer objects to their constituent tools for tool execution
819
820
  expanded_tools: list[Tool] = []
@@ -870,7 +871,7 @@ class _LLMClient(BaseModel):
870
871
 
871
872
  def run_agent_loop_sync(
872
873
  self,
873
- conversation: str | Conversation,
874
+ conversation: Prompt,
874
875
  *,
875
876
  tools: list[Tool | dict | MCPServer] | None = None,
876
877
  max_rounds: int = 5,
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: lm_deluge
3
- Version: 0.0.67
3
+ Version: 0.0.68
4
4
  Summary: Python utility for using LLM API models.
5
5
  Author-email: Benjamin Anderson <ben@trytaylor.ai>
6
6
  Requires-Python: >=3.10
File without changes
File without changes
File without changes