vlm-dataset-captioner 0.0.1__tar.gz → 0.0.2__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {vlm_dataset_captioner-0.0.1 → vlm_dataset_captioner-0.0.2}/PKG-INFO +1 -1
- vlm_dataset_captioner-0.0.2/vlm_dataset_captioner/__init__.py +5 -0
- {vlm_dataset_captioner-0.0.1 → vlm_dataset_captioner-0.0.2}/vlm_dataset_captioner/vlm_caption.py +13 -9
- {vlm_dataset_captioner-0.0.1 → vlm_dataset_captioner-0.0.2}/.gitignore +0 -0
- {vlm_dataset_captioner-0.0.1 → vlm_dataset_captioner-0.0.2}/README.md +0 -0
- {vlm_dataset_captioner-0.0.1 → vlm_dataset_captioner-0.0.2}/pyproject.toml +0 -0
- {vlm_dataset_captioner-0.0.1 → vlm_dataset_captioner-0.0.2}/vlm_dataset_captioner/vlm_caption_cli.py +0 -0
{vlm_dataset_captioner-0.0.1 → vlm_dataset_captioner-0.0.2}/vlm_dataset_captioner/vlm_caption.py
RENAMED
|
@@ -103,7 +103,7 @@ def caption_image(prompt, image, model, processor, max_new_tokens=None):
|
|
|
103
103
|
# Generate caption
|
|
104
104
|
generated_ids = model.generate(
|
|
105
105
|
**inputs,
|
|
106
|
-
max_new_tokens=128,
|
|
106
|
+
max_new_tokens=max_new_tokens or 128,
|
|
107
107
|
do_sample=True,
|
|
108
108
|
top_p=1.0,
|
|
109
109
|
temperature=0.7,
|
|
@@ -159,13 +159,17 @@ def requires_caption(image_file, output_directory, overwrite):
|
|
|
159
159
|
def caption_entire_directory(
|
|
160
160
|
directory_path,
|
|
161
161
|
output_directory,
|
|
162
|
-
|
|
163
|
-
|
|
162
|
+
model_name="Qwen/Qwen2.5-VL-32B-Instruct",
|
|
163
|
+
model=None,
|
|
164
|
+
processor=None,
|
|
164
165
|
max_new_tokens=None,
|
|
165
166
|
ignore_substring=None,
|
|
166
167
|
num_captions=None,
|
|
167
168
|
overwrite=False,
|
|
168
169
|
):
|
|
170
|
+
if model is None or processor is None:
|
|
171
|
+
model, processor = init_model(model_name=model_name)
|
|
172
|
+
|
|
169
173
|
print(
|
|
170
174
|
f"INFO: Processing directory {directory_path} for image captions.", flush=True
|
|
171
175
|
)
|
|
@@ -178,12 +182,12 @@ def caption_entire_directory(
|
|
|
178
182
|
caption_entire_directory(
|
|
179
183
|
subdir_path,
|
|
180
184
|
os.path.join(output_directory, subdir),
|
|
181
|
-
model,
|
|
182
|
-
processor,
|
|
183
|
-
max_new_tokens,
|
|
184
|
-
ignore_substring,
|
|
185
|
-
num_captions,
|
|
186
|
-
overwrite,
|
|
185
|
+
model=model,
|
|
186
|
+
processor=processor,
|
|
187
|
+
max_new_tokens=max_new_tokens,
|
|
188
|
+
ignore_substring=ignore_substring,
|
|
189
|
+
num_captions=num_captions,
|
|
190
|
+
overwrite=overwrite,
|
|
187
191
|
)
|
|
188
192
|
else:
|
|
189
193
|
prompt = get_prompt_for_directory(directory_path)
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{vlm_dataset_captioner-0.0.1 → vlm_dataset_captioner-0.0.2}/vlm_dataset_captioner/vlm_caption_cli.py
RENAMED
|
File without changes
|