ccs-llmconnector 1.0.5__tar.gz → 1.0.6__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (20) hide show
  1. {ccs_llmconnector-1.0.5/src/ccs_llmconnector.egg-info → ccs_llmconnector-1.0.6}/PKG-INFO +3 -1
  2. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/README.md +2 -0
  3. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/pyproject.toml +1 -1
  4. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6/src/ccs_llmconnector.egg-info}/PKG-INFO +3 -1
  5. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/src/llmconnector/client.py +5 -2
  6. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/src/llmconnector/gemini_client.py +5 -2
  7. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/src/llmconnector/openai_client.py +2 -1
  8. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/LICENSE +0 -0
  9. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/MANIFEST.in +0 -0
  10. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/setup.cfg +0 -0
  11. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/src/ccs_llmconnector.egg-info/SOURCES.txt +0 -0
  12. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/src/ccs_llmconnector.egg-info/dependency_links.txt +0 -0
  13. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/src/ccs_llmconnector.egg-info/entry_points.txt +0 -0
  14. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/src/ccs_llmconnector.egg-info/requires.txt +0 -0
  15. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/src/ccs_llmconnector.egg-info/top_level.txt +0 -0
  16. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/src/llmconnector/__init__.py +0 -0
  17. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/src/llmconnector/anthropic_client.py +0 -0
  18. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/src/llmconnector/client_cli.py +0 -0
  19. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/src/llmconnector/grok_client.py +0 -0
  20. {ccs_llmconnector-1.0.5 → ccs_llmconnector-1.0.6}/src/llmconnector/py.typed +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: ccs-llmconnector
3
- Version: 1.0.5
3
+ Version: 1.0.6
4
4
  Summary: Lightweight wrapper around different LLM provider Python SDK Responses APIs.
5
5
  Author: CCS
6
6
  License: MIT
@@ -104,6 +104,7 @@ image_bytes = client.generate_image(
104
104
  prompt="Generate an infographic of the current weather in Tokyo.",
105
105
  model="gemini-3-pro-image-preview",
106
106
  image_size="2K", # Optional, defaults to "2K"
107
+ aspect_ratio="16:9", # Optional, e.g. "16:9", "4:3"
107
108
  )
108
109
 
109
110
  with open("weather_tokyo.png", "wb") as f:
@@ -351,6 +352,7 @@ image_bytes = llm_client.generate_image(
351
352
  api_key="your-gemini-api-key",
352
353
  prompt="A futuristic city",
353
354
  model="gemini-3-pro-image-preview",
355
+ aspect_ratio="16:9",
354
356
  )
355
357
  ```
356
358
 
@@ -88,6 +88,7 @@ image_bytes = client.generate_image(
88
88
  prompt="Generate an infographic of the current weather in Tokyo.",
89
89
  model="gemini-3-pro-image-preview",
90
90
  image_size="2K", # Optional, defaults to "2K"
91
+ aspect_ratio="16:9", # Optional, e.g. "16:9", "4:3"
91
92
  )
92
93
 
93
94
  with open("weather_tokyo.png", "wb") as f:
@@ -335,6 +336,7 @@ image_bytes = llm_client.generate_image(
335
336
  api_key="your-gemini-api-key",
336
337
  prompt="A futuristic city",
337
338
  model="gemini-3-pro-image-preview",
339
+ aspect_ratio="16:9",
338
340
  )
339
341
  ```
340
342
 
@@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
4
4
 
5
5
  [project]
6
6
  name = "ccs-llmconnector"
7
- version = "1.0.5"
7
+ version = "1.0.6"
8
8
  description = "Lightweight wrapper around different LLM provider Python SDK Responses APIs."
9
9
  readme = "README.md"
10
10
  requires-python = ">=3.8"
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: ccs-llmconnector
3
- Version: 1.0.5
3
+ Version: 1.0.6
4
4
  Summary: Lightweight wrapper around different LLM provider Python SDK Responses APIs.
5
5
  Author: CCS
6
6
  License: MIT
@@ -104,6 +104,7 @@ image_bytes = client.generate_image(
104
104
  prompt="Generate an infographic of the current weather in Tokyo.",
105
105
  model="gemini-3-pro-image-preview",
106
106
  image_size="2K", # Optional, defaults to "2K"
107
+ aspect_ratio="16:9", # Optional, e.g. "16:9", "4:3"
107
108
  )
108
109
 
109
110
  with open("weather_tokyo.png", "wb") as f:
@@ -351,6 +352,7 @@ image_bytes = llm_client.generate_image(
351
352
  api_key="your-gemini-api-key",
352
353
  prompt="A futuristic city",
353
354
  model="gemini-3-pro-image-preview",
355
+ aspect_ratio="16:9",
354
356
  )
355
357
  ```
356
358
 
@@ -32,7 +32,8 @@ class SupportsGenerateResponse(Protocol):
32
32
  api_key: str,
33
33
  prompt: str,
34
34
  model: str,
35
- image_size: str = "2K",
35
+ image_size: Optional[str] = None,
36
+ aspect_ratio: Optional[str] = None,
36
37
  image: Optional[ImageInput] = None,
37
38
  ) -> bytes:
38
39
  ...
@@ -106,7 +107,8 @@ class LLMClient:
106
107
  api_key: str,
107
108
  prompt: str,
108
109
  model: str,
109
- image_size: str = "2K",
110
+ image_size: Optional[str] = None,
111
+ aspect_ratio: Optional[str] = None,
110
112
  image: Optional[ImageInput] = None,
111
113
  ) -> bytes:
112
114
  """Generate an image using the selected provider."""
@@ -125,6 +127,7 @@ class LLMClient:
125
127
  prompt=prompt,
126
128
  model=model,
127
129
  image_size=image_size,
130
+ aspect_ratio=aspect_ratio,
128
131
  image=image,
129
132
  )
130
133
 
@@ -136,7 +136,8 @@ class GeminiClient:
136
136
  api_key: str,
137
137
  prompt: str,
138
138
  model: str,
139
- image_size: str = "2K",
139
+ image_size: Optional[str] = None,
140
+ aspect_ratio: Optional[str] = None,
140
141
  image: Optional[ImageInput] = None,
141
142
  ) -> bytes:
142
143
  """Generate an image using Gemini 3 Pro Image.
@@ -146,6 +147,7 @@ class GeminiClient:
146
147
  prompt: Text prompt for image generation.
147
148
  model: Identifier of the Gemini model to target (e.g., "gemini-3-pro-image-preview").
148
149
  image_size: Size of the generated image (e.g., "2K", "4K"). Defaults to "2K".
150
+ aspect_ratio: Aspect ratio of the generated image (e.g., "16:9", "4:3").
149
151
  image: Optional input image for editing tasks.
150
152
 
151
153
  Returns:
@@ -167,7 +169,8 @@ class GeminiClient:
167
169
  config = types.GenerateContentConfig(
168
170
  tools=[{"google_search": {}}],
169
171
  image_config=types.ImageConfig(
170
- image_size=image_size
172
+ image_size=image_size or "2K",
173
+ aspect_ratio=aspect_ratio,
171
174
  )
172
175
  )
173
176
 
@@ -124,7 +124,8 @@ class OpenAIResponsesClient:
124
124
  api_key: str,
125
125
  prompt: str,
126
126
  model: str,
127
- image_size: str = "2K",
127
+ image_size: Optional[str] = None,
128
+ aspect_ratio: Optional[str] = None,
128
129
  image: Optional[ImageInput] = None,
129
130
  ) -> bytes:
130
131
  """Generate an image using the OpenAI API.