chunkr-ai 0.1.0a10__tar.gz → 0.1.0a11__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- chunkr_ai-0.1.0a11/.release-please-manifest.json +3 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/CHANGELOG.md +8 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/PKG-INFO +1 -1
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/pyproject.toml +1 -1
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_version.py +1 -1
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/resources/tasks/parse.py +9 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/__init__.py +2 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/file_info.py +0 -3
- chunkr_ai-0.1.0a11/src/chunkr_ai/types/llm_processing.py +36 -0
- chunkr_ai-0.1.0a11/src/chunkr_ai/types/llm_processing_param.py +36 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/ocr_result.py +6 -6
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/parse_configuration.py +4 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/parse_configuration_param.py +4 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/segment.py +5 -8
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/segment_processing.py +2 -92
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/segment_processing_param.py +2 -92
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/task_response.py +2 -8
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/tasks/extract_create_response.py +1 -7
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/tasks/extract_get_response.py +1 -7
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/tasks/parse_create_params.py +4 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/tasks/parse_create_response.py +0 -6
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/tasks/parse_get_response.py +0 -6
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/version_info.py +1 -1
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/api_resources/tasks/test_extract.py +16 -74
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/api_resources/tasks/test_parse.py +16 -74
- chunkr_ai-0.1.0a10/.release-please-manifest.json +0 -3
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/.gitignore +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/CONTRIBUTING.md +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/LICENSE +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/README.md +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/SECURITY.md +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/api.md +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/bin/check-release-environment +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/bin/publish-pypi +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/examples/.keep +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/noxfile.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/release-please-config.json +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/requirements-dev.lock +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/requirements.lock +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr/lib/.keep +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/__init__.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_base_client.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_client.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_compat.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_constants.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_exceptions.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_files.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_models.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_qs.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_resource.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_response.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_streaming.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_types.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_utils/__init__.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_utils/_compat.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_utils/_datetime_parse.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_utils/_logs.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_utils/_proxy.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_utils/_reflection.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_utils/_resources_proxy.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_utils/_streams.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_utils/_sync.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_utils/_transform.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_utils/_typing.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/_utils/_utils.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/lib/.keep +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/pagination.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/py.typed +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/resources/__init__.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/resources/files.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/resources/health.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/resources/tasks/__init__.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/resources/tasks/extract.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/resources/tasks/tasks.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/resources/webhooks.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/bounding_box.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/cell.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/cell_style.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/chunk.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/chunk_processing.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/chunk_processing_param.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/delete.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/extract_configuration.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/extract_output_response.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/file.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/file_create_params.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/file_list_params.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/file_url.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/file_url_params.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/files_list_response.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/generation_config.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/generation_config_param.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/health_check_response.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/page.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/parse_output_response.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/task_extract_updated_webhook_event.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/task_get_params.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/task_list_params.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/task_parse_updated_webhook_event.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/tasks/__init__.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/tasks/extract_create_params.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/tasks/extract_get_params.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/tasks/parse_get_params.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/unwrap_webhook_event.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/webhook_url_response.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/__init__.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/api_resources/__init__.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/api_resources/tasks/__init__.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/api_resources/test_files.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/api_resources/test_health.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/api_resources/test_tasks.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/api_resources/test_webhooks.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/conftest.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/sample_file.txt +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/test_client.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/test_deepcopy.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/test_extract_files.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/test_files.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/test_models.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/test_qs.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/test_required_args.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/test_response.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/test_streaming.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/test_transform.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/test_utils/test_datetime_parse.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/test_utils/test_proxy.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/test_utils/test_typing.py +0 -0
- {chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/tests/utils.py +0 -0
@@ -1,5 +1,13 @@
|
|
1
1
|
# Changelog
|
2
2
|
|
3
|
+
## 0.1.0-alpha.11 (2025-10-02)
|
4
|
+
|
5
|
+
Full Changelog: [v0.1.0-alpha.10...v0.1.0-alpha.11](https://github.com/lumina-ai-inc/chunkr-python/compare/v0.1.0-alpha.10...v0.1.0-alpha.11)
|
6
|
+
|
7
|
+
### Features
|
8
|
+
|
9
|
+
* **api:** api update ([335fcdd](https://github.com/lumina-ai-inc/chunkr-python/commit/335fcdd892d4d009ea966d12c5a3357fa38b595f))
|
10
|
+
|
3
11
|
## 0.1.0-alpha.10 (2025-10-02)
|
4
12
|
|
5
13
|
Full Changelog: [v0.1.0-alpha.9...v0.1.0-alpha.10](https://github.com/lumina-ai-inc/chunkr-python/compare/v0.1.0-alpha.9...v0.1.0-alpha.10)
|
@@ -1,6 +1,6 @@
|
|
1
1
|
Metadata-Version: 2.3
|
2
2
|
Name: chunkr-ai
|
3
|
-
Version: 0.1.
|
3
|
+
Version: 0.1.0a11
|
4
4
|
Summary: The official Python library for the chunkr API
|
5
5
|
Project-URL: Homepage, https://github.com/lumina-ai-inc/chunkr-python
|
6
6
|
Project-URL: Repository, https://github.com/lumina-ai-inc/chunkr-python
|
@@ -19,6 +19,7 @@ from ..._response import (
|
|
19
19
|
)
|
20
20
|
from ...types.tasks import parse_get_params, parse_create_params
|
21
21
|
from ..._base_client import make_request_options
|
22
|
+
from ...types.llm_processing_param import LlmProcessingParam
|
22
23
|
from ...types.chunk_processing_param import ChunkProcessingParam
|
23
24
|
from ...types.segment_processing_param import SegmentProcessingParam
|
24
25
|
from ...types.tasks.parse_get_response import ParseGetResponse
|
@@ -55,6 +56,7 @@ class ParseResource(SyncAPIResource):
|
|
55
56
|
error_handling: Literal["Fail", "Continue"] | Omit = omit,
|
56
57
|
expires_in: Optional[int] | Omit = omit,
|
57
58
|
file_name: Optional[str] | Omit = omit,
|
59
|
+
llm_processing: LlmProcessingParam | Omit = omit,
|
58
60
|
ocr_strategy: Literal["All", "Auto"] | Omit = omit,
|
59
61
|
pipeline: Literal["Azure", "Chunkr"] | Omit = omit,
|
60
62
|
segment_processing: Optional[SegmentProcessingParam] | Omit = omit,
|
@@ -97,6 +99,8 @@ class ParseResource(SyncAPIResource):
|
|
97
99
|
|
98
100
|
file_name: The name of the file to be parsed. If not set a name will be generated.
|
99
101
|
|
102
|
+
llm_processing: Controls the LLM used for the task.
|
103
|
+
|
100
104
|
ocr_strategy: Controls the Optical Character Recognition (OCR) strategy.
|
101
105
|
|
102
106
|
- `All`: Processes all pages with OCR. (Latency penalty: ~0.5 seconds per page)
|
@@ -146,6 +150,7 @@ class ParseResource(SyncAPIResource):
|
|
146
150
|
"error_handling": error_handling,
|
147
151
|
"expires_in": expires_in,
|
148
152
|
"file_name": file_name,
|
153
|
+
"llm_processing": llm_processing,
|
149
154
|
"ocr_strategy": ocr_strategy,
|
150
155
|
"pipeline": pipeline,
|
151
156
|
"segment_processing": segment_processing,
|
@@ -251,6 +256,7 @@ class AsyncParseResource(AsyncAPIResource):
|
|
251
256
|
error_handling: Literal["Fail", "Continue"] | Omit = omit,
|
252
257
|
expires_in: Optional[int] | Omit = omit,
|
253
258
|
file_name: Optional[str] | Omit = omit,
|
259
|
+
llm_processing: LlmProcessingParam | Omit = omit,
|
254
260
|
ocr_strategy: Literal["All", "Auto"] | Omit = omit,
|
255
261
|
pipeline: Literal["Azure", "Chunkr"] | Omit = omit,
|
256
262
|
segment_processing: Optional[SegmentProcessingParam] | Omit = omit,
|
@@ -293,6 +299,8 @@ class AsyncParseResource(AsyncAPIResource):
|
|
293
299
|
|
294
300
|
file_name: The name of the file to be parsed. If not set a name will be generated.
|
295
301
|
|
302
|
+
llm_processing: Controls the LLM used for the task.
|
303
|
+
|
296
304
|
ocr_strategy: Controls the Optical Character Recognition (OCR) strategy.
|
297
305
|
|
298
306
|
- `All`: Processes all pages with OCR. (Latency penalty: ~0.5 seconds per page)
|
@@ -342,6 +350,7 @@ class AsyncParseResource(AsyncAPIResource):
|
|
342
350
|
"error_handling": error_handling,
|
343
351
|
"expires_in": expires_in,
|
344
352
|
"file_name": file_name,
|
353
|
+
"llm_processing": llm_processing,
|
345
354
|
"ocr_strategy": ocr_strategy,
|
346
355
|
"pipeline": pipeline,
|
347
356
|
"segment_processing": segment_processing,
|
@@ -15,6 +15,7 @@ from .ocr_result import OcrResult as OcrResult
|
|
15
15
|
from .bounding_box import BoundingBox as BoundingBox
|
16
16
|
from .version_info import VersionInfo as VersionInfo
|
17
17
|
from .task_response import TaskResponse as TaskResponse
|
18
|
+
from .llm_processing import LlmProcessing as LlmProcessing
|
18
19
|
from .file_url_params import FileURLParams as FileURLParams
|
19
20
|
from .task_get_params import TaskGetParams as TaskGetParams
|
20
21
|
from .chunk_processing import ChunkProcessing as ChunkProcessing
|
@@ -25,6 +26,7 @@ from .file_create_params import FileCreateParams as FileCreateParams
|
|
25
26
|
from .segment_processing import SegmentProcessing as SegmentProcessing
|
26
27
|
from .files_list_response import FilesListResponse as FilesListResponse
|
27
28
|
from .parse_configuration import ParseConfiguration as ParseConfiguration
|
29
|
+
from .llm_processing_param import LlmProcessingParam as LlmProcessingParam
|
28
30
|
from .unwrap_webhook_event import UnwrapWebhookEvent as UnwrapWebhookEvent
|
29
31
|
from .webhook_url_response import WebhookURLResponse as WebhookURLResponse
|
30
32
|
from .extract_configuration import ExtractConfiguration as ExtractConfiguration
|
@@ -0,0 +1,36 @@
|
|
1
|
+
# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
|
2
|
+
|
3
|
+
from typing import Union, Optional
|
4
|
+
from typing_extensions import Literal, TypeAlias
|
5
|
+
|
6
|
+
from pydantic import Field as FieldInfo
|
7
|
+
|
8
|
+
from .._models import BaseModel
|
9
|
+
|
10
|
+
__all__ = ["LlmProcessing", "FallbackStrategy", "FallbackStrategyModel"]
|
11
|
+
|
12
|
+
|
13
|
+
class FallbackStrategyModel(BaseModel):
|
14
|
+
model: str = FieldInfo(alias="Model")
|
15
|
+
"""Use a specific model as fallback"""
|
16
|
+
|
17
|
+
|
18
|
+
FallbackStrategy: TypeAlias = Union[Literal["None", "Default"], FallbackStrategyModel]
|
19
|
+
|
20
|
+
|
21
|
+
class LlmProcessing(BaseModel):
|
22
|
+
fallback_strategy: Optional[FallbackStrategy] = None
|
23
|
+
"""The fallback strategy to use for the LLMs in the task."""
|
24
|
+
|
25
|
+
llm_model_id: Optional[str] = None
|
26
|
+
"""The ID of the model to use for the task.
|
27
|
+
|
28
|
+
If not provided, the default model will be used. Please check the documentation
|
29
|
+
for the model you want to use.
|
30
|
+
"""
|
31
|
+
|
32
|
+
max_completion_tokens: Optional[int] = None
|
33
|
+
"""The maximum number of tokens to generate."""
|
34
|
+
|
35
|
+
temperature: Optional[float] = None
|
36
|
+
"""The temperature to use for the LLM."""
|
@@ -0,0 +1,36 @@
|
|
1
|
+
# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
|
2
|
+
|
3
|
+
from __future__ import annotations
|
4
|
+
|
5
|
+
from typing import Union, Optional
|
6
|
+
from typing_extensions import Literal, Required, Annotated, TypeAlias, TypedDict
|
7
|
+
|
8
|
+
from .._utils import PropertyInfo
|
9
|
+
|
10
|
+
__all__ = ["LlmProcessingParam", "FallbackStrategy", "FallbackStrategyModel"]
|
11
|
+
|
12
|
+
|
13
|
+
class FallbackStrategyModel(TypedDict, total=False):
|
14
|
+
model: Required[Annotated[str, PropertyInfo(alias="Model")]]
|
15
|
+
"""Use a specific model as fallback"""
|
16
|
+
|
17
|
+
|
18
|
+
FallbackStrategy: TypeAlias = Union[Literal["None", "Default"], FallbackStrategyModel]
|
19
|
+
|
20
|
+
|
21
|
+
class LlmProcessingParam(TypedDict, total=False):
|
22
|
+
fallback_strategy: FallbackStrategy
|
23
|
+
"""The fallback strategy to use for the LLMs in the task."""
|
24
|
+
|
25
|
+
llm_model_id: Optional[str]
|
26
|
+
"""The ID of the model to use for the task.
|
27
|
+
|
28
|
+
If not provided, the default model will be used. Please check the documentation
|
29
|
+
for the model you want to use.
|
30
|
+
"""
|
31
|
+
|
32
|
+
max_completion_tokens: Optional[int]
|
33
|
+
"""The maximum number of tokens to generate."""
|
34
|
+
|
35
|
+
temperature: float
|
36
|
+
"""The temperature to use for the LLM."""
|
@@ -15,14 +15,14 @@ class OcrResult(BaseModel):
|
|
15
15
|
text: str
|
16
16
|
"""The recognized text of the OCR result."""
|
17
17
|
|
18
|
+
cell_ref: Optional[str] = None
|
19
|
+
"""
|
20
|
+
Excel-style cell reference (e.g., "A1" or "A1:B2") when OCR originates from a
|
21
|
+
spreadsheet cell
|
22
|
+
"""
|
23
|
+
|
18
24
|
confidence: Optional[float] = None
|
19
25
|
"""The confidence score of the recognized text."""
|
20
26
|
|
21
27
|
ocr_id: Optional[str] = None
|
22
28
|
"""The unique identifier for the OCR result."""
|
23
|
-
|
24
|
-
ss_cell_ref: Optional[str] = None
|
25
|
-
"""
|
26
|
-
Excel-style cell reference (e.g., "A1" or "A1:B2") when OCR originates from a
|
27
|
-
spreadsheet cell
|
28
|
-
"""
|
@@ -4,6 +4,7 @@ from typing import Optional
|
|
4
4
|
from typing_extensions import Literal
|
5
5
|
|
6
6
|
from .._models import BaseModel
|
7
|
+
from .llm_processing import LlmProcessing
|
7
8
|
from .chunk_processing import ChunkProcessing
|
8
9
|
from .segment_processing import SegmentProcessing
|
9
10
|
|
@@ -22,6 +23,9 @@ class ParseConfiguration(BaseModel):
|
|
22
23
|
LLM refusals etc.)
|
23
24
|
"""
|
24
25
|
|
26
|
+
llm_processing: Optional[LlmProcessing] = None
|
27
|
+
"""Controls the LLM used for the task."""
|
28
|
+
|
25
29
|
ocr_strategy: Optional[Literal["All", "Auto"]] = None
|
26
30
|
"""Controls the Optical Character Recognition (OCR) strategy.
|
27
31
|
|
@@ -5,6 +5,7 @@ from __future__ import annotations
|
|
5
5
|
from typing import Optional
|
6
6
|
from typing_extensions import Literal, TypedDict
|
7
7
|
|
8
|
+
from .llm_processing_param import LlmProcessingParam
|
8
9
|
from .chunk_processing_param import ChunkProcessingParam
|
9
10
|
from .segment_processing_param import SegmentProcessingParam
|
10
11
|
|
@@ -23,6 +24,9 @@ class ParseConfigurationParam(TypedDict, total=False):
|
|
23
24
|
LLM refusals etc.)
|
24
25
|
"""
|
25
26
|
|
27
|
+
llm_processing: LlmProcessingParam
|
28
|
+
"""Controls the LLM used for the task."""
|
29
|
+
|
26
30
|
ocr_strategy: Literal["All", "Auto"]
|
27
31
|
"""Controls the Optical Character Recognition (OCR) strategy.
|
28
32
|
|
@@ -31,23 +31,20 @@ class Segment(BaseModel):
|
|
31
31
|
"Caption",
|
32
32
|
"Footnote",
|
33
33
|
"Formula",
|
34
|
-
"FormRegion",
|
35
|
-
"GraphicalItem",
|
36
|
-
"Legend",
|
37
|
-
"LineNumber",
|
38
34
|
"ListItem",
|
39
35
|
"Page",
|
40
36
|
"PageFooter",
|
41
37
|
"PageHeader",
|
42
|
-
"PageNumber",
|
43
38
|
"Picture",
|
39
|
+
"SectionHeader",
|
44
40
|
"Table",
|
45
41
|
"Text",
|
46
42
|
"Title",
|
47
|
-
"Unknown",
|
48
|
-
"SectionHeader",
|
49
43
|
]
|
50
|
-
"""
|
44
|
+
"""
|
45
|
+
All the possible types for a segment. Note: Different configurations will
|
46
|
+
produce different types. Please refer to the documentation for more information.
|
47
|
+
"""
|
51
48
|
|
52
49
|
confidence: Optional[float] = None
|
53
50
|
"""Confidence score of the layout analysis model"""
|
@@ -47,24 +47,6 @@ class SegmentProcessing(BaseModel):
|
|
47
47
|
- `extended_context` uses the full page image as context for LLM generation.
|
48
48
|
"""
|
49
49
|
|
50
|
-
form_region: Optional[GenerationConfig] = FieldInfo(alias="FormRegion", default=None)
|
51
|
-
"""Controls the processing and generation for the segment.
|
52
|
-
|
53
|
-
- `crop_image` controls whether to crop the file's images to the segment's
|
54
|
-
bounding box. The cropped image will be stored in the segment's `image` field.
|
55
|
-
Use `All` to always crop, or `Auto` to only crop when needed for
|
56
|
-
post-processing.
|
57
|
-
- `format` specifies the output format: `Html` or `Markdown`
|
58
|
-
- `strategy` determines how the content is generated: `Auto`, `LLM`, or `Ignore`
|
59
|
-
- `Auto`: Process content automatically
|
60
|
-
- `LLM`: Use large language models for processing
|
61
|
-
- `Ignore`: Exclude segments from final output
|
62
|
-
- `description` enables LLM-generated descriptions for segments. **Note:** This
|
63
|
-
uses chunkr's own VLM models and is not configurable via LLM processing
|
64
|
-
configuration.
|
65
|
-
- `extended_context` uses the full page image as context for LLM generation.
|
66
|
-
"""
|
67
|
-
|
68
50
|
formula: Optional[GenerationConfig] = FieldInfo(alias="Formula", default=None)
|
69
51
|
"""Controls the processing and generation for the segment.
|
70
52
|
|
@@ -83,60 +65,6 @@ class SegmentProcessing(BaseModel):
|
|
83
65
|
- `extended_context` uses the full page image as context for LLM generation.
|
84
66
|
"""
|
85
67
|
|
86
|
-
graphical_item: Optional[GenerationConfig] = FieldInfo(alias="GraphicalItem", default=None)
|
87
|
-
"""Controls the processing and generation for the segment.
|
88
|
-
|
89
|
-
- `crop_image` controls whether to crop the file's images to the segment's
|
90
|
-
bounding box. The cropped image will be stored in the segment's `image` field.
|
91
|
-
Use `All` to always crop, or `Auto` to only crop when needed for
|
92
|
-
post-processing.
|
93
|
-
- `format` specifies the output format: `Html` or `Markdown`
|
94
|
-
- `strategy` determines how the content is generated: `Auto`, `LLM`, or `Ignore`
|
95
|
-
- `Auto`: Process content automatically
|
96
|
-
- `LLM`: Use large language models for processing
|
97
|
-
- `Ignore`: Exclude segments from final output
|
98
|
-
- `description` enables LLM-generated descriptions for segments. **Note:** This
|
99
|
-
uses chunkr's own VLM models and is not configurable via LLM processing
|
100
|
-
configuration.
|
101
|
-
- `extended_context` uses the full page image as context for LLM generation.
|
102
|
-
"""
|
103
|
-
|
104
|
-
legend: Optional[GenerationConfig] = FieldInfo(alias="Legend", default=None)
|
105
|
-
"""Controls the processing and generation for the segment.
|
106
|
-
|
107
|
-
- `crop_image` controls whether to crop the file's images to the segment's
|
108
|
-
bounding box. The cropped image will be stored in the segment's `image` field.
|
109
|
-
Use `All` to always crop, or `Auto` to only crop when needed for
|
110
|
-
post-processing.
|
111
|
-
- `format` specifies the output format: `Html` or `Markdown`
|
112
|
-
- `strategy` determines how the content is generated: `Auto`, `LLM`, or `Ignore`
|
113
|
-
- `Auto`: Process content automatically
|
114
|
-
- `LLM`: Use large language models for processing
|
115
|
-
- `Ignore`: Exclude segments from final output
|
116
|
-
- `description` enables LLM-generated descriptions for segments. **Note:** This
|
117
|
-
uses chunkr's own VLM models and is not configurable via LLM processing
|
118
|
-
configuration.
|
119
|
-
- `extended_context` uses the full page image as context for LLM generation.
|
120
|
-
"""
|
121
|
-
|
122
|
-
line_number: Optional[GenerationConfig] = FieldInfo(alias="LineNumber", default=None)
|
123
|
-
"""Controls the processing and generation for the segment.
|
124
|
-
|
125
|
-
- `crop_image` controls whether to crop the file's images to the segment's
|
126
|
-
bounding box. The cropped image will be stored in the segment's `image` field.
|
127
|
-
Use `All` to always crop, or `Auto` to only crop when needed for
|
128
|
-
post-processing.
|
129
|
-
- `format` specifies the output format: `Html` or `Markdown`
|
130
|
-
- `strategy` determines how the content is generated: `Auto`, `LLM`, or `Ignore`
|
131
|
-
- `Auto`: Process content automatically
|
132
|
-
- `LLM`: Use large language models for processing
|
133
|
-
- `Ignore`: Exclude segments from final output
|
134
|
-
- `description` enables LLM-generated descriptions for segments. **Note:** This
|
135
|
-
uses chunkr's own VLM models and is not configurable via LLM processing
|
136
|
-
configuration.
|
137
|
-
- `extended_context` uses the full page image as context for LLM generation.
|
138
|
-
"""
|
139
|
-
|
140
68
|
list_item: Optional[GenerationConfig] = FieldInfo(alias="ListItem", default=None)
|
141
69
|
"""Controls the processing and generation for the segment.
|
142
70
|
|
@@ -209,7 +137,7 @@ class SegmentProcessing(BaseModel):
|
|
209
137
|
- `extended_context` uses the full page image as context for LLM generation.
|
210
138
|
"""
|
211
139
|
|
212
|
-
|
140
|
+
picture: Optional[GenerationConfig] = FieldInfo(alias="Picture", default=None)
|
213
141
|
"""Controls the processing and generation for the segment.
|
214
142
|
|
215
143
|
- `crop_image` controls whether to crop the file's images to the segment's
|
@@ -227,7 +155,7 @@ class SegmentProcessing(BaseModel):
|
|
227
155
|
- `extended_context` uses the full page image as context for LLM generation.
|
228
156
|
"""
|
229
157
|
|
230
|
-
|
158
|
+
section_header: Optional[GenerationConfig] = FieldInfo(alias="SectionHeader", default=None)
|
231
159
|
"""Controls the processing and generation for the segment.
|
232
160
|
|
233
161
|
- `crop_image` controls whether to crop the file's images to the segment's
|
@@ -298,21 +226,3 @@ class SegmentProcessing(BaseModel):
|
|
298
226
|
configuration.
|
299
227
|
- `extended_context` uses the full page image as context for LLM generation.
|
300
228
|
"""
|
301
|
-
|
302
|
-
unknown: Optional[GenerationConfig] = FieldInfo(alias="Unknown", default=None)
|
303
|
-
"""Controls the processing and generation for the segment.
|
304
|
-
|
305
|
-
- `crop_image` controls whether to crop the file's images to the segment's
|
306
|
-
bounding box. The cropped image will be stored in the segment's `image` field.
|
307
|
-
Use `All` to always crop, or `Auto` to only crop when needed for
|
308
|
-
post-processing.
|
309
|
-
- `format` specifies the output format: `Html` or `Markdown`
|
310
|
-
- `strategy` determines how the content is generated: `Auto`, `LLM`, or `Ignore`
|
311
|
-
- `Auto`: Process content automatically
|
312
|
-
- `LLM`: Use large language models for processing
|
313
|
-
- `Ignore`: Exclude segments from final output
|
314
|
-
- `description` enables LLM-generated descriptions for segments. **Note:** This
|
315
|
-
uses chunkr's own VLM models and is not configurable via LLM processing
|
316
|
-
configuration.
|
317
|
-
- `extended_context` uses the full page image as context for LLM generation.
|
318
|
-
"""
|
@@ -48,24 +48,6 @@ class SegmentProcessingParam(TypedDict, total=False):
|
|
48
48
|
- `extended_context` uses the full page image as context for LLM generation.
|
49
49
|
"""
|
50
50
|
|
51
|
-
form_region: Annotated[Optional[GenerationConfigParam], PropertyInfo(alias="FormRegion")]
|
52
|
-
"""Controls the processing and generation for the segment.
|
53
|
-
|
54
|
-
- `crop_image` controls whether to crop the file's images to the segment's
|
55
|
-
bounding box. The cropped image will be stored in the segment's `image` field.
|
56
|
-
Use `All` to always crop, or `Auto` to only crop when needed for
|
57
|
-
post-processing.
|
58
|
-
- `format` specifies the output format: `Html` or `Markdown`
|
59
|
-
- `strategy` determines how the content is generated: `Auto`, `LLM`, or `Ignore`
|
60
|
-
- `Auto`: Process content automatically
|
61
|
-
- `LLM`: Use large language models for processing
|
62
|
-
- `Ignore`: Exclude segments from final output
|
63
|
-
- `description` enables LLM-generated descriptions for segments. **Note:** This
|
64
|
-
uses chunkr's own VLM models and is not configurable via LLM processing
|
65
|
-
configuration.
|
66
|
-
- `extended_context` uses the full page image as context for LLM generation.
|
67
|
-
"""
|
68
|
-
|
69
51
|
formula: Annotated[Optional[GenerationConfigParam], PropertyInfo(alias="Formula")]
|
70
52
|
"""Controls the processing and generation for the segment.
|
71
53
|
|
@@ -84,60 +66,6 @@ class SegmentProcessingParam(TypedDict, total=False):
|
|
84
66
|
- `extended_context` uses the full page image as context for LLM generation.
|
85
67
|
"""
|
86
68
|
|
87
|
-
graphical_item: Annotated[Optional[GenerationConfigParam], PropertyInfo(alias="GraphicalItem")]
|
88
|
-
"""Controls the processing and generation for the segment.
|
89
|
-
|
90
|
-
- `crop_image` controls whether to crop the file's images to the segment's
|
91
|
-
bounding box. The cropped image will be stored in the segment's `image` field.
|
92
|
-
Use `All` to always crop, or `Auto` to only crop when needed for
|
93
|
-
post-processing.
|
94
|
-
- `format` specifies the output format: `Html` or `Markdown`
|
95
|
-
- `strategy` determines how the content is generated: `Auto`, `LLM`, or `Ignore`
|
96
|
-
- `Auto`: Process content automatically
|
97
|
-
- `LLM`: Use large language models for processing
|
98
|
-
- `Ignore`: Exclude segments from final output
|
99
|
-
- `description` enables LLM-generated descriptions for segments. **Note:** This
|
100
|
-
uses chunkr's own VLM models and is not configurable via LLM processing
|
101
|
-
configuration.
|
102
|
-
- `extended_context` uses the full page image as context for LLM generation.
|
103
|
-
"""
|
104
|
-
|
105
|
-
legend: Annotated[Optional[GenerationConfigParam], PropertyInfo(alias="Legend")]
|
106
|
-
"""Controls the processing and generation for the segment.
|
107
|
-
|
108
|
-
- `crop_image` controls whether to crop the file's images to the segment's
|
109
|
-
bounding box. The cropped image will be stored in the segment's `image` field.
|
110
|
-
Use `All` to always crop, or `Auto` to only crop when needed for
|
111
|
-
post-processing.
|
112
|
-
- `format` specifies the output format: `Html` or `Markdown`
|
113
|
-
- `strategy` determines how the content is generated: `Auto`, `LLM`, or `Ignore`
|
114
|
-
- `Auto`: Process content automatically
|
115
|
-
- `LLM`: Use large language models for processing
|
116
|
-
- `Ignore`: Exclude segments from final output
|
117
|
-
- `description` enables LLM-generated descriptions for segments. **Note:** This
|
118
|
-
uses chunkr's own VLM models and is not configurable via LLM processing
|
119
|
-
configuration.
|
120
|
-
- `extended_context` uses the full page image as context for LLM generation.
|
121
|
-
"""
|
122
|
-
|
123
|
-
line_number: Annotated[Optional[GenerationConfigParam], PropertyInfo(alias="LineNumber")]
|
124
|
-
"""Controls the processing and generation for the segment.
|
125
|
-
|
126
|
-
- `crop_image` controls whether to crop the file's images to the segment's
|
127
|
-
bounding box. The cropped image will be stored in the segment's `image` field.
|
128
|
-
Use `All` to always crop, or `Auto` to only crop when needed for
|
129
|
-
post-processing.
|
130
|
-
- `format` specifies the output format: `Html` or `Markdown`
|
131
|
-
- `strategy` determines how the content is generated: `Auto`, `LLM`, or `Ignore`
|
132
|
-
- `Auto`: Process content automatically
|
133
|
-
- `LLM`: Use large language models for processing
|
134
|
-
- `Ignore`: Exclude segments from final output
|
135
|
-
- `description` enables LLM-generated descriptions for segments. **Note:** This
|
136
|
-
uses chunkr's own VLM models and is not configurable via LLM processing
|
137
|
-
configuration.
|
138
|
-
- `extended_context` uses the full page image as context for LLM generation.
|
139
|
-
"""
|
140
|
-
|
141
69
|
list_item: Annotated[Optional[GenerationConfigParam], PropertyInfo(alias="ListItem")]
|
142
70
|
"""Controls the processing and generation for the segment.
|
143
71
|
|
@@ -210,7 +138,7 @@ class SegmentProcessingParam(TypedDict, total=False):
|
|
210
138
|
- `extended_context` uses the full page image as context for LLM generation.
|
211
139
|
"""
|
212
140
|
|
213
|
-
|
141
|
+
picture: Annotated[Optional[GenerationConfigParam], PropertyInfo(alias="Picture")]
|
214
142
|
"""Controls the processing and generation for the segment.
|
215
143
|
|
216
144
|
- `crop_image` controls whether to crop the file's images to the segment's
|
@@ -228,7 +156,7 @@ class SegmentProcessingParam(TypedDict, total=False):
|
|
228
156
|
- `extended_context` uses the full page image as context for LLM generation.
|
229
157
|
"""
|
230
158
|
|
231
|
-
|
159
|
+
section_header: Annotated[Optional[GenerationConfigParam], PropertyInfo(alias="SectionHeader")]
|
232
160
|
"""Controls the processing and generation for the segment.
|
233
161
|
|
234
162
|
- `crop_image` controls whether to crop the file's images to the segment's
|
@@ -299,21 +227,3 @@ class SegmentProcessingParam(TypedDict, total=False):
|
|
299
227
|
configuration.
|
300
228
|
- `extended_context` uses the full page image as context for LLM generation.
|
301
229
|
"""
|
302
|
-
|
303
|
-
unknown: Annotated[Optional[GenerationConfigParam], PropertyInfo(alias="Unknown")]
|
304
|
-
"""Controls the processing and generation for the segment.
|
305
|
-
|
306
|
-
- `crop_image` controls whether to crop the file's images to the segment's
|
307
|
-
bounding box. The cropped image will be stored in the segment's `image` field.
|
308
|
-
Use `All` to always crop, or `Auto` to only crop when needed for
|
309
|
-
post-processing.
|
310
|
-
- `format` specifies the output format: `Html` or `Markdown`
|
311
|
-
- `strategy` determines how the content is generated: `Auto`, `LLM`, or `Ignore`
|
312
|
-
- `Auto`: Process content automatically
|
313
|
-
- `LLM`: Use large language models for processing
|
314
|
-
- `Ignore`: Exclude segments from final output
|
315
|
-
- `description` enables LLM-generated descriptions for segments. **Note:** This
|
316
|
-
uses chunkr's own VLM models and is not configurable via LLM processing
|
317
|
-
configuration.
|
318
|
-
- `extended_context` uses the full page image as context for LLM generation.
|
319
|
-
"""
|
@@ -20,12 +20,6 @@ Output: TypeAlias = Union[ParseOutputResponse, ExtractOutputResponse, None]
|
|
20
20
|
|
21
21
|
|
22
22
|
class TaskResponse(BaseModel):
|
23
|
-
completed: bool
|
24
|
-
"""True when the task reaches a terminal state i.e.
|
25
|
-
|
26
|
-
`status` is `Succeeded` or `Failed` or `Cancelled`
|
27
|
-
"""
|
28
|
-
|
29
23
|
configuration: Configuration
|
30
24
|
"""
|
31
25
|
Unified configuration type that can represent either parse or extract
|
@@ -64,8 +58,8 @@ class TaskResponse(BaseModel):
|
|
64
58
|
output: Optional[Output] = None
|
65
59
|
"""Unified output type that can represent either parse or extract results"""
|
66
60
|
|
67
|
-
|
68
|
-
"""The ID of the source
|
61
|
+
source_task_id: Optional[str] = None
|
62
|
+
"""The ID of the source task that was used for the task"""
|
69
63
|
|
70
64
|
started_at: Optional[datetime] = None
|
71
65
|
"""The date and time when the task was started."""
|
{chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/tasks/extract_create_response.py
RENAMED
@@ -14,12 +14,6 @@ __all__ = ["ExtractCreateResponse"]
|
|
14
14
|
|
15
15
|
|
16
16
|
class ExtractCreateResponse(BaseModel):
|
17
|
-
completed: bool
|
18
|
-
"""True when the task reaches a terminal state i.e.
|
19
|
-
|
20
|
-
`status` is `Succeeded` or `Failed` or `Cancelled`
|
21
|
-
"""
|
22
|
-
|
23
17
|
configuration: ExtractConfiguration
|
24
18
|
|
25
19
|
created_at: datetime
|
@@ -63,7 +57,7 @@ class ExtractCreateResponse(BaseModel):
|
|
63
57
|
for that field.
|
64
58
|
"""
|
65
59
|
|
66
|
-
|
60
|
+
source_task_id: Optional[str] = None
|
67
61
|
"""The ID of the source `parse` task that was used for extraction"""
|
68
62
|
|
69
63
|
started_at: Optional[datetime] = None
|
@@ -14,12 +14,6 @@ __all__ = ["ExtractGetResponse"]
|
|
14
14
|
|
15
15
|
|
16
16
|
class ExtractGetResponse(BaseModel):
|
17
|
-
completed: bool
|
18
|
-
"""True when the task reaches a terminal state i.e.
|
19
|
-
|
20
|
-
`status` is `Succeeded` or `Failed` or `Cancelled`
|
21
|
-
"""
|
22
|
-
|
23
17
|
configuration: ExtractConfiguration
|
24
18
|
|
25
19
|
created_at: datetime
|
@@ -63,7 +57,7 @@ class ExtractGetResponse(BaseModel):
|
|
63
57
|
for that field.
|
64
58
|
"""
|
65
59
|
|
66
|
-
|
60
|
+
source_task_id: Optional[str] = None
|
67
61
|
"""The ID of the source `parse` task that was used for extraction"""
|
68
62
|
|
69
63
|
started_at: Optional[datetime] = None
|
@@ -5,6 +5,7 @@ from __future__ import annotations
|
|
5
5
|
from typing import Optional
|
6
6
|
from typing_extensions import Literal, Required, TypedDict
|
7
7
|
|
8
|
+
from ..llm_processing_param import LlmProcessingParam
|
8
9
|
from ..chunk_processing_param import ChunkProcessingParam
|
9
10
|
from ..segment_processing_param import SegmentProcessingParam
|
10
11
|
|
@@ -41,6 +42,9 @@ class ParseCreateParams(TypedDict, total=False):
|
|
41
42
|
file_name: Optional[str]
|
42
43
|
"""The name of the file to be parsed. If not set a name will be generated."""
|
43
44
|
|
45
|
+
llm_processing: LlmProcessingParam
|
46
|
+
"""Controls the LLM used for the task."""
|
47
|
+
|
44
48
|
ocr_strategy: Literal["All", "Auto"]
|
45
49
|
"""Controls the Optical Character Recognition (OCR) strategy.
|
46
50
|
|
{chunkr_ai-0.1.0a10 → chunkr_ai-0.1.0a11}/src/chunkr_ai/types/tasks/parse_create_response.py
RENAMED
@@ -14,12 +14,6 @@ __all__ = ["ParseCreateResponse"]
|
|
14
14
|
|
15
15
|
|
16
16
|
class ParseCreateResponse(BaseModel):
|
17
|
-
completed: bool
|
18
|
-
"""True when the task reaches a terminal state i.e.
|
19
|
-
|
20
|
-
`status` is `Succeeded` or `Failed` or `Cancelled`
|
21
|
-
"""
|
22
|
-
|
23
17
|
configuration: ParseConfiguration
|
24
18
|
|
25
19
|
created_at: datetime
|
@@ -14,12 +14,6 @@ __all__ = ["ParseGetResponse"]
|
|
14
14
|
|
15
15
|
|
16
16
|
class ParseGetResponse(BaseModel):
|
17
|
-
completed: bool
|
18
|
-
"""True when the task reaches a terminal state i.e.
|
19
|
-
|
20
|
-
`status` is `Succeeded` or `Failed` or `Cancelled`
|
21
|
-
"""
|
22
|
-
|
23
17
|
configuration: ParseConfiguration
|
24
18
|
|
25
19
|
created_at: datetime
|